runtime error

model-00001-of-00002.safetensors: 83%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 4.17G/5.00G [00:19<00:06, 122MB/s] model-00001-of-00002.safetensors: 83%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 4.17G/5.00G [00:19<00:06, 122MB/s] model-00001-of-00002.safetensors: 84%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 4.21G/5.00G [00:20<00:07, 109MB/s] model-00001-of-00002.safetensors: 92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 4.60G/5.00G [00:24<00:03, 109MB/s] model-00001-of-00002.safetensors: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰| 5.00G/5.00G [00:26<00:00, 185MB/s] Downloading shards: 50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 1/2 [00:27<00:27, 27.34s/it] model-00002-of-00002.safetensors: 0%| | 0.00/564M [00:00<?, ?B/s] model-00002-of-00002.safetensors: 9%|β–‰ | 52.4M/564M [00:01<00:12, 40.9MB/s] model-00002-of-00002.safetensors: 70%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 396M/564M [00:02<00:00, 197MB/s]  model-00002-of-00002.safetensors: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰| 564M/564M [00:03<00:00, 159MB/s] Downloading shards: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 2/2 [00:31<00:00, 13.48s/it] Downloading shards: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 2/2 [00:31<00:00, 15.56s/it] Traceback (most recent call last): File "/home/user/app/app.py", line 41, in <module> model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float32, device_map="auto", trust_remote_code=True) File "/home/user/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 566, in from_pretrained return model_class.from_pretrained( File "/home/user/.local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3606, in from_pretrained no_split_modules = model._get_no_split_modules(device_map) File "/home/user/.local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 1690, in _get_no_split_modules raise ValueError( ValueError: PhiForCausalLM does not support `device_map='auto'`. To implement support, the model class needs to implement the `_no_split_modules` attribute.

Container logs:

Fetching error logs...