Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -27,7 +27,8 @@ AVAILABLE_MODELS = {
|
|
| 27 |
"opt-350m": "facebook/opt-350m",
|
| 28 |
"pythia-160m": "EleutherAI/pythia-160m",
|
| 29 |
"mistral-7b": "mistralai/Mistral-7B-Instruct-v0.3",
|
| 30 |
-
"deepseek-small": "deepseek-ai/DeepSeek-V2-Lite"
|
|
|
|
| 31 |
}
|
| 32 |
|
| 33 |
# Initialize model and tokenizer
|
|
@@ -45,7 +46,7 @@ def load_model(model_name):
|
|
| 45 |
current_model_name = model_name
|
| 46 |
|
| 47 |
# Load the default model at startup
|
| 48 |
-
load_model("
|
| 49 |
|
| 50 |
@spaces.GPU()
|
| 51 |
def get_next_token_predictions(text, model_name, top_k=10):
|
|
|
|
| 27 |
"opt-350m": "facebook/opt-350m",
|
| 28 |
"pythia-160m": "EleutherAI/pythia-160m",
|
| 29 |
"mistral-7b": "mistralai/Mistral-7B-Instruct-v0.3",
|
| 30 |
+
"deepseek-small": "deepseek-ai/DeepSeek-V2-Lite",
|
| 31 |
+
"llama": "meta-llama/Llama-3.2-1B"
|
| 32 |
}
|
| 33 |
|
| 34 |
# Initialize model and tokenizer
|
|
|
|
| 46 |
current_model_name = model_name
|
| 47 |
|
| 48 |
# Load the default model at startup
|
| 49 |
+
load_model("mistral-7b")
|
| 50 |
|
| 51 |
@spaces.GPU()
|
| 52 |
def get_next_token_predictions(text, model_name, top_k=10):
|