Upload inclusionAI_Ring-flash-linear-2.0-128k_1.txt with huggingface_hub
Browse files
inclusionAI_Ring-flash-linear-2.0-128k_1.txt
CHANGED
|
@@ -6,7 +6,7 @@ model = AutoModelForCausalLM.from_pretrained("inclusionAI/Ring-flash-linear-2.0-
|
|
| 6 |
|
| 7 |
ERROR:
|
| 8 |
Traceback (most recent call last):
|
| 9 |
-
File "/tmp/inclusionAI_Ring-flash-linear-2.0-
|
| 10 |
model = AutoModelForCausalLM.from_pretrained("inclusionAI/Ring-flash-linear-2.0-128k", trust_remote_code=True, torch_dtype="auto")
|
| 11 |
File "/tmp/.cache/uv/environments-v2/7a001057f9cd0d8c/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 586, in from_pretrained
|
| 12 |
model_class = get_class_from_dynamic_module(
|
|
|
|
| 6 |
|
| 7 |
ERROR:
|
| 8 |
Traceback (most recent call last):
|
| 9 |
+
File "/tmp/inclusionAI_Ring-flash-linear-2.0-128k_1Jth61K.py", line 18, in <module>
|
| 10 |
model = AutoModelForCausalLM.from_pretrained("inclusionAI/Ring-flash-linear-2.0-128k", trust_remote_code=True, torch_dtype="auto")
|
| 11 |
File "/tmp/.cache/uv/environments-v2/7a001057f9cd0d8c/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 586, in from_pretrained
|
| 12 |
model_class = get_class_from_dynamic_module(
|