ariG23498 HF Staff commited on
Commit
f433f39
·
verified ·
1 Parent(s): ec47d5c

Upload utter-project_EuroLLM-1.7B_0.txt with huggingface_hub

Browse files
Files changed (1) hide show
  1. utter-project_EuroLLM-1.7B_0.txt +61 -0
utter-project_EuroLLM-1.7B_0.txt ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ```CODE:
2
+ # Use a pipeline as a high-level helper
3
+ from transformers import pipeline
4
+
5
+ pipe = pipeline("text-generation", model="utter-project/EuroLLM-1.7B")
6
+ ```
7
+
8
+ ERROR:
9
+ Traceback (most recent call last):
10
+ File "/tmp/.cache/uv/environments-v2/3c00d0927f67afe2/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2343, in _from_pretrained
11
+ tokenizer = cls(*init_inputs, **init_kwargs)
12
+ File "/tmp/.cache/uv/environments-v2/3c00d0927f67afe2/lib/python3.13/site-packages/transformers/models/llama/tokenization_llama_fast.py", line 154, in __init__
13
+ super().__init__(
14
+ ~~~~~~~~~~~~~~~~^
15
+ vocab_file=vocab_file,
16
+ ^^^^^^^^^^^^^^^^^^^^^^
17
+ ...<10 lines>...
18
+ **kwargs,
19
+ ^^^^^^^^^
20
+ )
21
+ ^
22
+ File "/tmp/.cache/uv/environments-v2/3c00d0927f67afe2/lib/python3.13/site-packages/transformers/tokenization_utils_fast.py", line 108, in __init__
23
+ raise ValueError(
24
+ ...<2 lines>...
25
+ )
26
+ ValueError: Cannot instantiate this tokenizer from a slow version. If it's based on sentencepiece, make sure you have sentencepiece installed.
27
+
28
+ During handling of the above exception, another exception occurred:
29
+
30
+ Traceback (most recent call last):
31
+ File "/tmp/utter-project_EuroLLM-1.7B_0opO2GF.py", line 19, in <module>
32
+ pipe = pipeline("text-generation", model="utter-project/EuroLLM-1.7B")
33
+ File "/tmp/.cache/uv/environments-v2/3c00d0927f67afe2/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1078, in pipeline
34
+ raise e
35
+ File "/tmp/.cache/uv/environments-v2/3c00d0927f67afe2/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1073, in pipeline
36
+ tokenizer = AutoTokenizer.from_pretrained(
37
+ tokenizer_identifier, use_fast=use_fast, _from_pipeline=task, **hub_kwargs, **tokenizer_kwargs
38
+ )
39
+ File "/tmp/.cache/uv/environments-v2/3c00d0927f67afe2/lib/python3.13/site-packages/transformers/models/auto/tokenization_auto.py", line 1140, in from_pretrained
40
+ return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs)
41
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
42
+ File "/tmp/.cache/uv/environments-v2/3c00d0927f67afe2/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2097, in from_pretrained
43
+ return cls._from_pretrained(
44
+ ~~~~~~~~~~~~~~~~~~~~^
45
+ resolved_vocab_files,
46
+ ^^^^^^^^^^^^^^^^^^^^^
47
+ ...<9 lines>...
48
+ **kwargs,
49
+ ^^^^^^^^^
50
+ )
51
+ ^
52
+ File "/tmp/.cache/uv/environments-v2/3c00d0927f67afe2/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2344, in _from_pretrained
53
+ except import_protobuf_decode_error():
54
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^
55
+ File "/tmp/.cache/uv/environments-v2/3c00d0927f67afe2/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 87, in import_protobuf_decode_error
56
+ raise ImportError(PROTOBUF_IMPORT_ERROR.format(error_message))
57
+ ImportError:
58
+ requires the protobuf library but it was not found in your environment. Check out the instructions on the
59
+ installation page of its repo: https://github.com/protocolbuffers/protobuf/tree/master/python#installation and follow the ones
60
+ that match your environment. Please note that you may need to restart your runtime after installation.
61
+