Spaces:
Running
Running
Vokturz
commited on
Commit
·
6826b0f
1
Parent(s):
f6e2171
cached model list to memory (exclude falcon)
Browse files- src/app.py +18 -6
src/app.py
CHANGED
|
@@ -47,14 +47,24 @@ st.markdown(
|
|
| 47 |
""",
|
| 48 |
unsafe_allow_html=True,
|
| 49 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 50 |
@st.cache_resource
|
| 51 |
def get_gpu_specs():
|
| 52 |
return pd.read_csv("data/gpu_specs.csv")
|
| 53 |
|
| 54 |
-
@st.cache_resource
|
| 55 |
-
def get_mistralai_table():
|
| 56 |
-
|
| 57 |
-
|
| 58 |
|
| 59 |
def show_gpu_info(info, trainable_params=0, vendor=""):
|
| 60 |
for var in ['Inference', 'Full Training Adam', 'LoRa Fine-tuning']:
|
|
@@ -91,7 +101,9 @@ def get_name(index):
|
|
| 91 |
|
| 92 |
def custom_ceil(a, precision=0):
|
| 93 |
return np.round(a + 0.5 * 10**(-precision), precision)
|
|
|
|
| 94 |
gpu_specs = get_gpu_specs()
|
|
|
|
| 95 |
|
| 96 |
_, col, _ = st.columns([1,3,1])
|
| 97 |
with col.expander("Information", expanded=True):
|
|
@@ -123,8 +135,8 @@ if model_name not in st.session_state:
|
|
| 123 |
del st.session_state[st.session_state['actual_model']]
|
| 124 |
del st.session_state['actual_model']
|
| 125 |
gc.collect()
|
| 126 |
-
if model_name
|
| 127 |
-
st.session_state[model_name] =
|
| 128 |
else:
|
| 129 |
model = get_model(model_name, library="transformers", access_token=access_token)
|
| 130 |
st.session_state[model_name] = calculate_memory(model, ["float32", "float16/bfloat16", "int8", "int4"])
|
|
|
|
| 47 |
""",
|
| 48 |
unsafe_allow_html=True,
|
| 49 |
)
|
| 50 |
+
|
| 51 |
+
@st.cache_resource()
|
| 52 |
+
def cache_model_list():
|
| 53 |
+
model_list_info = {}
|
| 54 |
+
for model_name in model_list:
|
| 55 |
+
if not "tiiuae/falcon" in model_name: # Exclude Falcon models
|
| 56 |
+
model = get_model(model_name, library="transformers", access_token="")
|
| 57 |
+
model_list_info[model_name] = calculate_memory(model, ["float32", "float16/bfloat16", "int8", "int4"])
|
| 58 |
+
return model_list_info
|
| 59 |
+
|
| 60 |
@st.cache_resource
|
| 61 |
def get_gpu_specs():
|
| 62 |
return pd.read_csv("data/gpu_specs.csv")
|
| 63 |
|
| 64 |
+
# @st.cache_resource
|
| 65 |
+
# def get_mistralai_table():
|
| 66 |
+
# model = get_model("mistralai/Mistral-7B-v0.1", library="transformers", access_token="")
|
| 67 |
+
# return calculate_memory(model, ["float32", "float16/bfloat16", "int8", "int4"])
|
| 68 |
|
| 69 |
def show_gpu_info(info, trainable_params=0, vendor=""):
|
| 70 |
for var in ['Inference', 'Full Training Adam', 'LoRa Fine-tuning']:
|
|
|
|
| 101 |
|
| 102 |
def custom_ceil(a, precision=0):
|
| 103 |
return np.round(a + 0.5 * 10**(-precision), precision)
|
| 104 |
+
|
| 105 |
gpu_specs = get_gpu_specs()
|
| 106 |
+
model_list_info = cache_model_list()
|
| 107 |
|
| 108 |
_, col, _ = st.columns([1,3,1])
|
| 109 |
with col.expander("Information", expanded=True):
|
|
|
|
| 135 |
del st.session_state[st.session_state['actual_model']]
|
| 136 |
del st.session_state['actual_model']
|
| 137 |
gc.collect()
|
| 138 |
+
if model_name in model_list_info.keys():
|
| 139 |
+
st.session_state[model_name] = model_list_info[model_name]
|
| 140 |
else:
|
| 141 |
model = get_model(model_name, library="transformers", access_token=access_token)
|
| 142 |
st.session_state[model_name] = calculate_memory(model, ["float32", "float16/bfloat16", "int8", "int4"])
|