runtime error
Exit code: 1. Reason: Downloading shards: 0%| | 0/4 [00:00<?, ?it/s][A Downloading shards: 25%|βββ | 1/4 [00:10<00:32, 10.89s/it][A Downloading shards: 50%|βββββ | 2/4 [00:20<00:20, 10.38s/it][A Downloading shards: 75%|ββββββββ | 3/4 [00:30<00:09, 9.86s/it][A Downloading shards: 100%|ββββββββββ| 4/4 [00:33<00:00, 7.30s/it][A Downloading shards: 100%|ββββββββββ| 4/4 [00:33<00:00, 8.38s/it] Loading checkpoint shards: 0%| | 0/4 [00:00<?, ?it/s][A Loading checkpoint shards: 100%|ββββββββββ| 4/4 [00:00<00:00, 79512.87it/s] Traceback (most recent call last): File "/home/user/app/app.py", line 34, in <module> model = AutoModelForCausalLM.from_pretrained(MODEL, torch_dtype=torch.bfloat16, trust_remote_code=True, device_map="auto") File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3977, in from_pretrained dispatch_model(model, **device_map_kwargs) File "/usr/local/lib/python3.10/site-packages/accelerate/big_modeling.py", line 496, in dispatch_model raise ValueError( ValueError: You are trying to offload the whole model to the disk. Please use the `disk_offload` function instead.
Container logs:
Fetching error logs...