runtime error

| 0.00/13.2k [00:00<?, ?B/s] train_script.py: 100%|██████████| 13.2k/13.2k [00:00<00:00, 79.1MB/s] vocab.txt: 0%| | 0.00/232k [00:00<?, ?B/s] vocab.txt: 100%|██████████| 232k/232k [00:00<00:00, 40.3MB/s] modules.json: 0%| | 0.00/349 [00:00<?, ?B/s] modules.json: 100%|██████████| 349/349 [00:00<00:00, 1.11MB/s] tokenizer_config.json: 0%| | 0.00/1.69k [00:00<?, ?B/s] tokenizer_config.json: 100%|██████████| 1.69k/1.69k [00:00<00:00, 7.01MB/s] tokenizer.model: 0%| | 0.00/493k [00:00<?, ?B/s] tokenizer.model: 100%|██████████| 493k/493k [00:00<00:00, 230MB/s] added_tokens.json: 0%| | 0.00/90.0 [00:00<?, ?B/s] added_tokens.json: 100%|██████████| 90.0/90.0 [00:00<00:00, 409kB/s] special_tokens_map.json: 0%| | 0.00/101 [00:00<?, ?B/s] special_tokens_map.json: 100%|██████████| 101/101 [00:00<00:00, 751kB/s] Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained. Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained. config.json: 0%| | 0.00/623 [00:00<?, ?B/s] config.json: 100%|██████████| 623/623 [00:00<00:00, 2.96MB/s] Traceback (most recent call last): File "/home/user/app/app.py", line 51, in <module> model = AutoModelForCausalLM.from_pretrained(model_path, quantization_config=bnb_config, device_map={"":0}) File "/home/user/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 566, in from_pretrained return model_class.from_pretrained( File "/home/user/.local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 2897, in from_pretrained raise RuntimeError("No GPU found. A GPU is needed for quantization.") RuntimeError: No GPU found. A GPU is needed for quantization.

Container logs:

Fetching error logs...