ariG23498 HF Staff commited on
Commit
68ca2d8
·
verified ·
1 Parent(s): 13f37ef

Upload moonshotai_Kimi-K2-Thinking_1.txt with huggingface_hub

Browse files
Files changed (1) hide show
  1. moonshotai_Kimi-K2-Thinking_1.txt +48 -0
moonshotai_Kimi-K2-Thinking_1.txt ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ```CODE:
2
+ # Load model directly
3
+ from transformers import AutoModelForCausalLM
4
+ model = AutoModelForCausalLM.from_pretrained("moonshotai/Kimi-K2-Thinking", trust_remote_code=True, torch_dtype="auto")
5
+ ```
6
+
7
+ ERROR:
8
+ Traceback (most recent call last):
9
+ File "/tmp/moonshotai_Kimi-K2-Thinking_1rdNusL.py", line 24, in <module>
10
+ model = AutoModelForCausalLM.from_pretrained("moonshotai/Kimi-K2-Thinking", trust_remote_code=True, torch_dtype="auto")
11
+ File "/tmp/.cache/uv/environments-v2/2592c6825cb441bd/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 597, in from_pretrained
12
+ return model_class.from_pretrained(
13
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~^
14
+ pretrained_model_name_or_path, *model_args, config=config, **hub_kwargs, **kwargs
15
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
16
+ )
17
+ ^
18
+ File "/tmp/.cache/uv/environments-v2/2592c6825cb441bd/lib/python3.13/site-packages/transformers/modeling_utils.py", line 277, in _wrapper
19
+ return func(*args, **kwargs)
20
+ File "/tmp/.cache/uv/environments-v2/2592c6825cb441bd/lib/python3.13/site-packages/transformers/modeling_utils.py", line 4881, in from_pretrained
21
+ hf_quantizer, config, dtype, device_map = get_hf_quantizer(
22
+ ~~~~~~~~~~~~~~~~^
23
+ config, quantization_config, dtype, from_tf, from_flax, device_map, weights_only, user_agent
24
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
25
+ )
26
+ ^
27
+ File "/tmp/.cache/uv/environments-v2/2592c6825cb441bd/lib/python3.13/site-packages/transformers/quantizers/auto.py", line 305, in get_hf_quantizer
28
+ config.quantization_config = AutoHfQuantizer.merge_quantization_configs(
29
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^
30
+ config.quantization_config, quantization_config
31
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
32
+ )
33
+ ^
34
+ File "/tmp/.cache/uv/environments-v2/2592c6825cb441bd/lib/python3.13/site-packages/transformers/quantizers/auto.py", line 214, in merge_quantization_configs
35
+ quantization_config = AutoQuantizationConfig.from_dict(quantization_config)
36
+ File "/tmp/.cache/uv/environments-v2/2592c6825cb441bd/lib/python3.13/site-packages/transformers/quantizers/auto.py", line 140, in from_dict
37
+ return target_cls.from_dict(quantization_config_dict)
38
+ ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^
39
+ File "/tmp/.cache/uv/environments-v2/2592c6825cb441bd/lib/python3.13/site-packages/transformers/utils/quantization_config.py", line 1398, in from_dict
40
+ return super().from_dict(config_dict, return_unused_kwargs=return_unused_kwargs, **kwargs)
41
+ ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
42
+ File "/tmp/.cache/uv/environments-v2/2592c6825cb441bd/lib/python3.13/site-packages/transformers/utils/quantization_config.py", line 122, in from_dict
43
+ config = cls(**config_dict)
44
+ File "/tmp/.cache/uv/environments-v2/2592c6825cb441bd/lib/python3.13/site-packages/transformers/utils/quantization_config.py", line 1328, in __init__
45
+ raise ImportError(
46
+ "compressed_tensors is not installed and is required for compressed-tensors quantization. Please install it with `pip install compressed-tensors`."
47
+ )
48
+ ImportError: compressed_tensors is not installed and is required for compressed-tensors quantization. Please install it with `pip install compressed-tensors`.