Spaces:
Build error
Build error
fix chat history, update settings to use GPU
Browse files- chat.py +1 -1
- config.yml +1 -0
chat.py
CHANGED
|
@@ -38,7 +38,7 @@ def chat(history, system_message):
|
|
| 38 |
history[-1][1] = ""
|
| 39 |
for output in llm(messages, max_tokens=512, stop=["</s>", "<unk>", "### User:"], echo=False, stream=True):
|
| 40 |
answer = output['choices'][0]['text']
|
| 41 |
-
history[-1][1]
|
| 42 |
|
| 43 |
yield history, history
|
| 44 |
|
|
|
|
| 38 |
history[-1][1] = ""
|
| 39 |
for output in llm(messages, max_tokens=512, stop=["</s>", "<unk>", "### User:"], echo=False, stream=True):
|
| 40 |
answer = output['choices'][0]['text']
|
| 41 |
+
history[-1][1] += answer
|
| 42 |
|
| 43 |
yield history, history
|
| 44 |
|
config.yml
CHANGED
|
@@ -5,3 +5,4 @@ file: wizard-vicuna-13B.ggml.q5_1.bin
|
|
| 5 |
base_model: junelee/wizard-vicuna-13b
|
| 6 |
llama_cpp:
|
| 7 |
n_ctx: 1024
|
|
|
|
|
|
| 5 |
base_model: junelee/wizard-vicuna-13b
|
| 6 |
llama_cpp:
|
| 7 |
n_ctx: 1024
|
| 8 |
+
n_gpu_layers: 40 # llama 13b has 40 layers
|