Spaces:
Running
Running
Update
Browse files
models.py
CHANGED
|
@@ -84,6 +84,12 @@ def generate(
|
|
| 84 |
print(gradio_helpers.get_paths())
|
| 85 |
model = llama_cpp.Llama(
|
| 86 |
model_path,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 87 |
)
|
| 88 |
# return model.tokenizer.to_str(tokens[0])
|
| 89 |
return "output"
|
|
|
|
| 84 |
print(gradio_helpers.get_paths())
|
| 85 |
model = llama_cpp.Llama(
|
| 86 |
model_path,
|
| 87 |
+
chat_handler=llama_cpp.llama_chat_format.PaligemmaChatHandler(
|
| 88 |
+
clip_path
|
| 89 |
+
),
|
| 90 |
+
n_ctx=1024,
|
| 91 |
+
n_ubatch=512,
|
| 92 |
+
n_batch=512,
|
| 93 |
)
|
| 94 |
# return model.tokenizer.to_str(tokens[0])
|
| 95 |
return "output"
|