Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -169,10 +169,10 @@ class LlamaCppModel(Model):
|
|
| 169 |
model_path=model_path,
|
| 170 |
flash_attn=False,
|
| 171 |
n_gpu_layers=0,
|
| 172 |
-
n_batch=
|
| 173 |
n_ctx=n_ctx,
|
| 174 |
n_threads=2,
|
| 175 |
-
n_threads_batch=
|
| 176 |
)
|
| 177 |
|
| 178 |
elif repo_id and filename:
|
|
@@ -263,7 +263,7 @@ class LlamaCppModel(Model):
|
|
| 263 |
llm_sampling_settings=settings,
|
| 264 |
chat_history=messages,
|
| 265 |
returns_streaming_generator=False,
|
| 266 |
-
print_output=
|
| 267 |
|
| 268 |
)
|
| 269 |
|
|
@@ -298,7 +298,6 @@ Reference
|
|
| 298 |
- [Dataset(m-ric/huggingface_doc)](https://huggingface.co/datasets/m-ric/huggingface_doc)
|
| 299 |
|
| 300 |
"""
|
| 301 |
-
#Tool not support
|
| 302 |
agent = CodeAgent(prompt_templates =yaml.safe_load(prompt),model=model, tools=[retriever_tool],max_steps=1,verbosity_level=0,name="AGENT",description=description)
|
| 303 |
|
| 304 |
demo = GradioUI(agent)
|
|
|
|
| 169 |
model_path=model_path,
|
| 170 |
flash_attn=False,
|
| 171 |
n_gpu_layers=0,
|
| 172 |
+
#n_batch=1024,
|
| 173 |
n_ctx=n_ctx,
|
| 174 |
n_threads=2,
|
| 175 |
+
n_threads_batch=2,verbose=False
|
| 176 |
)
|
| 177 |
|
| 178 |
elif repo_id and filename:
|
|
|
|
| 263 |
llm_sampling_settings=settings,
|
| 264 |
chat_history=messages,
|
| 265 |
returns_streaming_generator=False,
|
| 266 |
+
print_output=False,
|
| 267 |
|
| 268 |
)
|
| 269 |
|
|
|
|
| 298 |
- [Dataset(m-ric/huggingface_doc)](https://huggingface.co/datasets/m-ric/huggingface_doc)
|
| 299 |
|
| 300 |
"""
|
|
|
|
| 301 |
agent = CodeAgent(prompt_templates =yaml.safe_load(prompt),model=model, tools=[retriever_tool],max_steps=1,verbosity_level=0,name="AGENT",description=description)
|
| 302 |
|
| 303 |
demo = GradioUI(agent)
|