Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
else
Browse files
app.py
CHANGED
|
@@ -19,6 +19,7 @@ def translate_llama2(text):
|
|
| 19 |
"Translates llama-2 to its hf counterpart"
|
| 20 |
if not text.endswith("-hf"):
|
| 21 |
return text + "-hf"
|
|
|
|
| 22 |
|
| 23 |
def check_for_discussion(model_name:str):
|
| 24 |
"Checks if an automated discussion has been opened on the model by `model-sizer-bot`"
|
|
@@ -62,7 +63,7 @@ def convert_url_to_name(url:str):
|
|
| 62 |
|
| 63 |
def calculate_memory(model_name:str, library:str, options:list, access_token:str, raw=False):
|
| 64 |
"Calculates the memory usage for a model"
|
| 65 |
-
if
|
| 66 |
model_name = translate_llama2(model_name)
|
| 67 |
if library == "auto":
|
| 68 |
library = None
|
|
|
|
| 19 |
"Translates llama-2 to its hf counterpart"
|
| 20 |
if not text.endswith("-hf"):
|
| 21 |
return text + "-hf"
|
| 22 |
+
return text
|
| 23 |
|
| 24 |
def check_for_discussion(model_name:str):
|
| 25 |
"Checks if an automated discussion has been opened on the model by `model-sizer-bot`"
|
|
|
|
| 63 |
|
| 64 |
def calculate_memory(model_name:str, library:str, options:list, access_token:str, raw=False):
|
| 65 |
"Calculates the memory usage for a model"
|
| 66 |
+
if "meta-llama" in model_name:
|
| 67 |
model_name = translate_llama2(model_name)
|
| 68 |
if library == "auto":
|
| 69 |
library = None
|