ariG23498 HF Staff commited on
Commit
2d3f800
·
verified ·
1 Parent(s): 309e7bb

Upload cerebras_GLM-4.6-REAP-218B-A32B-FP8_1.txt with huggingface_hub

Browse files
cerebras_GLM-4.6-REAP-218B-A32B-FP8_1.txt ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ```CODE:
2
+ # Load model directly
3
+ from transformers import AutoTokenizer, AutoModelForCausalLM
4
+
5
+ tokenizer = AutoTokenizer.from_pretrained("cerebras/GLM-4.6-REAP-218B-A32B-FP8")
6
+ model = AutoModelForCausalLM.from_pretrained("cerebras/GLM-4.6-REAP-218B-A32B-FP8")
7
+ messages = [
8
+ {"role": "user", "content": "Who are you?"},
9
+ ]
10
+ inputs = tokenizer.apply_chat_template(
11
+ messages,
12
+ add_generation_prompt=True,
13
+ tokenize=True,
14
+ return_dict=True,
15
+ return_tensors="pt",
16
+ ).to(model.device)
17
+
18
+ outputs = model.generate(**inputs, max_new_tokens=40)
19
+ print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))
20
+ ```
21
+
22
+ ERROR:
23
+ Traceback (most recent call last):
24
+ File "/tmp/cerebras_GLM-4.6-REAP-218B-A32B-FP8_1q6NVLk.py", line 18, in <module>
25
+ model = AutoModelForCausalLM.from_pretrained("cerebras/GLM-4.6-REAP-218B-A32B-FP8")
26
+ File "/tmp/.cache/uv/environments-v2/7eb248acb9b9547f/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 604, in from_pretrained
27
+ return model_class.from_pretrained(
28
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~^
29
+ pretrained_model_name_or_path, *model_args, config=config, **hub_kwargs, **kwargs
30
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
31
+ )
32
+ ^
33
+ File "/tmp/.cache/uv/environments-v2/7eb248acb9b9547f/lib/python3.13/site-packages/transformers/modeling_utils.py", line 277, in _wrapper
34
+ return func(*args, **kwargs)
35
+ File "/tmp/.cache/uv/environments-v2/7eb248acb9b9547f/lib/python3.13/site-packages/transformers/modeling_utils.py", line 4881, in from_pretrained
36
+ hf_quantizer, config, dtype, device_map = get_hf_quantizer(
37
+ ~~~~~~~~~~~~~~~~^
38
+ config, quantization_config, dtype, from_tf, from_flax, device_map, weights_only, user_agent
39
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
40
+ )
41
+ ^
42
+ File "/tmp/.cache/uv/environments-v2/7eb248acb9b9547f/lib/python3.13/site-packages/transformers/quantizers/auto.py", line 305, in get_hf_quantizer
43
+ config.quantization_config = AutoHfQuantizer.merge_quantization_configs(
44
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^
45
+ config.quantization_config, quantization_config
46
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
47
+ )
48
+ ^
49
+ File "/tmp/.cache/uv/environments-v2/7eb248acb9b9547f/lib/python3.13/site-packages/transformers/quantizers/auto.py", line 214, in merge_quantization_configs
50
+ quantization_config = AutoQuantizationConfig.from_dict(quantization_config)
51
+ File "/tmp/.cache/uv/environments-v2/7eb248acb9b9547f/lib/python3.13/site-packages/transformers/quantizers/auto.py", line 140, in from_dict
52
+ return target_cls.from_dict(quantization_config_dict)
53
+ ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^
54
+ File "/tmp/.cache/uv/environments-v2/7eb248acb9b9547f/lib/python3.13/site-packages/transformers/utils/quantization_config.py", line 1398, in from_dict
55
+ return super().from_dict(config_dict, return_unused_kwargs=return_unused_kwargs, **kwargs)
56
+ ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
57
+ File "/tmp/.cache/uv/environments-v2/7eb248acb9b9547f/lib/python3.13/site-packages/transformers/utils/quantization_config.py", line 122, in from_dict
58
+ config = cls(**config_dict)
59
+ File "/tmp/.cache/uv/environments-v2/7eb248acb9b9547f/lib/python3.13/site-packages/transformers/utils/quantization_config.py", line 1328, in __init__
60
+ raise ImportError(
61
+ "compressed_tensors is not installed and is required for compressed-tensors quantization. Please install it with `pip install compressed-tensors`."
62
+ )
63
+ ImportError: compressed_tensors is not installed and is required for compressed-tensors quantization. Please install it with `pip install compressed-tensors`.