runtime error
model-00001-of-00002.safetensors: 83%|βββββββββ | 4.17G/5.00G [00:19<00:06, 122MB/s][A model-00001-of-00002.safetensors: 83%|βββββββββ | 4.17G/5.00G [00:19<00:06, 122MB/s][A model-00001-of-00002.safetensors: 84%|βββββββββ | 4.21G/5.00G [00:20<00:07, 109MB/s][A model-00001-of-00002.safetensors: 92%|ββββββββββ| 4.60G/5.00G [00:24<00:03, 109MB/s][A model-00001-of-00002.safetensors: 100%|ββββββββββ| 5.00G/5.00G [00:26<00:00, 185MB/s] Downloading shards: 50%|βββββ | 1/2 [00:27<00:27, 27.34s/it][A model-00002-of-00002.safetensors: 0%| | 0.00/564M [00:00<?, ?B/s][A model-00002-of-00002.safetensors: 9%|β | 52.4M/564M [00:01<00:12, 40.9MB/s][A model-00002-of-00002.safetensors: 70%|βββββββ | 396M/564M [00:02<00:00, 197MB/s] [A model-00002-of-00002.safetensors: 100%|ββββββββββ| 564M/564M [00:03<00:00, 159MB/s] Downloading shards: 100%|ββββββββββ| 2/2 [00:31<00:00, 13.48s/it][A Downloading shards: 100%|ββββββββββ| 2/2 [00:31<00:00, 15.56s/it] Traceback (most recent call last): File "/home/user/app/app.py", line 41, in <module> model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float32, device_map="auto", trust_remote_code=True) File "/home/user/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 566, in from_pretrained return model_class.from_pretrained( File "/home/user/.local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3606, in from_pretrained no_split_modules = model._get_no_split_modules(device_map) File "/home/user/.local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 1690, in _get_no_split_modules raise ValueError( ValueError: PhiForCausalLM does not support `device_map='auto'`. To implement support, the model class needs to implement the `_no_split_modules` attribute.
Container logs:
Fetching error logs...