ariG23498 HF Staff commited on
Commit
e6992d0
·
verified ·
1 Parent(s): e7c5487

Upload moonshotai_Kimi-K2-Instruct-0905_1.txt with huggingface_hub

Browse files
moonshotai_Kimi-K2-Instruct-0905_1.txt ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ```CODE:
2
+ # Load model directly
3
+ from transformers import AutoModelForCausalLM
4
+ model = AutoModelForCausalLM.from_pretrained("moonshotai/Kimi-K2-Instruct-0905", trust_remote_code=True, torch_dtype="auto")
5
+ ```
6
+
7
+ ERROR:
8
+ Traceback (most recent call last):
9
+ File "/tmp/moonshotai_Kimi-K2-Instruct-0905_10soRf9.py", line 16, in <module>
10
+ model = AutoModelForCausalLM.from_pretrained("moonshotai/Kimi-K2-Instruct-0905", trust_remote_code=True, torch_dtype="auto")
11
+ File "/tmp/.cache/uv/environments-v2/a026bb39f18df708/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 597, in from_pretrained
12
+ return model_class.from_pretrained(
13
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~^
14
+ pretrained_model_name_or_path, *model_args, config=config, **hub_kwargs, **kwargs
15
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
16
+ )
17
+ ^
18
+ File "/tmp/.cache/uv/environments-v2/a026bb39f18df708/lib/python3.13/site-packages/transformers/modeling_utils.py", line 277, in _wrapper
19
+ return func(*args, **kwargs)
20
+ File "/tmp/.cache/uv/environments-v2/a026bb39f18df708/lib/python3.13/site-packages/transformers/modeling_utils.py", line 4881, in from_pretrained
21
+ hf_quantizer, config, dtype, device_map = get_hf_quantizer(
22
+ ~~~~~~~~~~~~~~~~^
23
+ config, quantization_config, dtype, from_tf, from_flax, device_map, weights_only, user_agent
24
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
25
+ )
26
+ ^
27
+ File "/tmp/.cache/uv/environments-v2/a026bb39f18df708/lib/python3.13/site-packages/transformers/quantizers/auto.py", line 319, in get_hf_quantizer
28
+ hf_quantizer.validate_environment(
29
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^
30
+ dtype=dtype,
31
+ ^^^^^^^^^^^^
32
+ ...<3 lines>...
33
+ weights_only=weights_only,
34
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^
35
+ )
36
+ ^
37
+ File "/tmp/.cache/uv/environments-v2/a026bb39f18df708/lib/python3.13/site-packages/transformers/quantizers/quantizer_finegrained_fp8.py", line 48, in validate_environment
38
+ raise RuntimeError("No GPU or XPU found. A GPU or XPU is needed for FP8 quantization.")
39
+ RuntimeError: No GPU or XPU found. A GPU or XPU is needed for FP8 quantization.