Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -106,6 +106,29 @@ hf_hub_download(
|
|
| 106 |
|
| 107 |
retriever_tool = RetrieverTool(docs_processed)
|
| 108 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 109 |
# based https://github.com/huggingface/smolagents/pull/450
|
| 110 |
# almost overwrite with https://huggingface.co/spaces/sitammeur/Gemma-llamacpp
|
| 111 |
class LlamaCppModel(Model):
|
|
@@ -207,13 +230,14 @@ class LlamaCppModel(Model):
|
|
| 207 |
agent = LlamaCppAgent(
|
| 208 |
provider,
|
| 209 |
system_prompt=f"{system_message}",
|
| 210 |
-
|
|
|
|
| 211 |
debug_output=True,
|
| 212 |
)
|
| 213 |
-
temperature = 0.
|
| 214 |
top_k=40
|
| 215 |
top_p=0.95
|
| 216 |
-
max_tokens=
|
| 217 |
repeat_penalty=1.1
|
| 218 |
settings = provider.get_provider_default_settings()
|
| 219 |
settings.temperature = temperature
|
|
|
|
| 106 |
|
| 107 |
retriever_tool = RetrieverTool(docs_processed)
|
| 108 |
|
| 109 |
+
# Define the prompt markers for Gemma 3
|
| 110 |
+
gemma_3_prompt_markers = {
|
| 111 |
+
Roles.system: PromptMarkers("", "\n"), # System prompt should be included within user message
|
| 112 |
+
Roles.user: PromptMarkers("<start_of_turn>user\n", "<end_of_turn>\n"),
|
| 113 |
+
Roles.assistant: PromptMarkers("<start_of_turn>model\n", "<end_of_turn>\n"),
|
| 114 |
+
Roles.tool: PromptMarkers("", ""), # If you need tool support
|
| 115 |
+
}
|
| 116 |
+
|
| 117 |
+
|
| 118 |
+
|
| 119 |
+
|
| 120 |
+
|
| 121 |
+
# Create the formatter
|
| 122 |
+
gemma_3_formatter = MessagesFormatter(
|
| 123 |
+
pre_prompt="", # No pre-prompt
|
| 124 |
+
prompt_markers=gemma_3_prompt_markers,
|
| 125 |
+
include_sys_prompt_in_first_user_message=True, # Include system prompt in first user message
|
| 126 |
+
default_stop_sequences=["<end_of_turn>", "<start_of_turn>"],
|
| 127 |
+
strip_prompt=False, # Don't strip whitespace from the prompt
|
| 128 |
+
bos_token="<bos>", # Beginning of sequence token for Gemma 3
|
| 129 |
+
eos_token="<eos>", # End of sequence token for Gemma 3
|
| 130 |
+
)
|
| 131 |
+
|
| 132 |
# based https://github.com/huggingface/smolagents/pull/450
|
| 133 |
# almost overwrite with https://huggingface.co/spaces/sitammeur/Gemma-llamacpp
|
| 134 |
class LlamaCppModel(Model):
|
|
|
|
| 230 |
agent = LlamaCppAgent(
|
| 231 |
provider,
|
| 232 |
system_prompt=f"{system_message}",
|
| 233 |
+
custom_messages_formatter=gemma_3_formatter,
|
| 234 |
+
#predefined_messages_formatter_type=MessagesFormatterType.GEMMA_2,
|
| 235 |
debug_output=True,
|
| 236 |
)
|
| 237 |
+
temperature = 0.5
|
| 238 |
top_k=40
|
| 239 |
top_p=0.95
|
| 240 |
+
max_tokens=2048
|
| 241 |
repeat_penalty=1.1
|
| 242 |
settings = provider.get_provider_default_settings()
|
| 243 |
settings.temperature = temperature
|