Spaces:
Running
on
Zero
Running
on
Zero
Commit
·
f35135e
1
Parent(s):
e24b19e
change where trust_remote is applied
Browse files- utils/models.py +2 -3
utils/models.py
CHANGED
|
@@ -101,9 +101,7 @@ def run_inference(model_name, context, question):
|
|
| 101 |
|
| 102 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 103 |
result = ""
|
| 104 |
-
model_kwargs = {
|
| 105 |
-
"trust_remote_code": True,
|
| 106 |
-
} # make sure qwen3 doesn't use thinking
|
| 107 |
if "qwen3" in model_name.lower():
|
| 108 |
print(f"Recognized {model_name} as a Qwen3 model. Setting enable_thinking=False.")
|
| 109 |
model_kwargs["enable_thinking"] = False
|
|
@@ -131,6 +129,7 @@ def run_inference(model_name, context, question):
|
|
| 131 |
temperature=0.6,
|
| 132 |
top_p=0.9,
|
| 133 |
model_kwargs=model_kwargs,
|
|
|
|
| 134 |
)
|
| 135 |
|
| 136 |
text_input = format_rag_prompt(question, context, accepts_sys)
|
|
|
|
| 101 |
|
| 102 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 103 |
result = ""
|
| 104 |
+
model_kwargs = {} # make sure qwen3 doesn't use thinking
|
|
|
|
|
|
|
| 105 |
if "qwen3" in model_name.lower():
|
| 106 |
print(f"Recognized {model_name} as a Qwen3 model. Setting enable_thinking=False.")
|
| 107 |
model_kwargs["enable_thinking"] = False
|
|
|
|
| 129 |
temperature=0.6,
|
| 130 |
top_p=0.9,
|
| 131 |
model_kwargs=model_kwargs,
|
| 132 |
+
trust_remote_code=True,
|
| 133 |
)
|
| 134 |
|
| 135 |
text_input = format_rag_prompt(question, context, accepts_sys)
|