Spaces:
Sleeping
Sleeping
Update main.py
Browse files
main.py
CHANGED
|
@@ -55,11 +55,15 @@ class QueryModel(BaseModel):
|
|
| 55 |
|
| 56 |
class NewsQueryModel(BaseModel):
|
| 57 |
query: str = Field(..., description="News topic to search for")
|
| 58 |
-
|
|
|
|
|
|
|
|
|
|
| 59 |
class Config:
|
| 60 |
schema_extra = {
|
| 61 |
"example": {
|
| 62 |
-
"query": "Latest developments in AI"
|
|
|
|
| 63 |
}
|
| 64 |
}
|
| 65 |
|
|
@@ -260,7 +264,7 @@ async def news_assistant(query: NewsQueryModel, api_key: str = Depends(verify_ap
|
|
| 260 |
raise HTTPException(status_code=500, detail="Failed to fetch news data")
|
| 261 |
|
| 262 |
def process_response():
|
| 263 |
-
for content in chat_with_llama_stream(messages, model=
|
| 264 |
yield content
|
| 265 |
#meta-llama/llama-3-70b-instruct google/gemini-pro-1.5
|
| 266 |
return StreamingResponse(process_response(), media_type="text/event-stream")
|
|
|
|
| 55 |
|
| 56 |
class NewsQueryModel(BaseModel):
|
| 57 |
query: str = Field(..., description="News topic to search for")
|
| 58 |
+
model_id: ModelID = Field(
|
| 59 |
+
default="meta-llama/llama-3-70b-instruct",
|
| 60 |
+
description="ID of the model to use for response generation"
|
| 61 |
+
)
|
| 62 |
class Config:
|
| 63 |
schema_extra = {
|
| 64 |
"example": {
|
| 65 |
+
"query": "Latest developments in AI",
|
| 66 |
+
"model_id": "meta-llama/llama-3-70b-instruct"
|
| 67 |
}
|
| 68 |
}
|
| 69 |
|
|
|
|
| 264 |
raise HTTPException(status_code=500, detail="Failed to fetch news data")
|
| 265 |
|
| 266 |
def process_response():
|
| 267 |
+
for content in chat_with_llama_stream(messages, model=query.model_id):
|
| 268 |
yield content
|
| 269 |
#meta-llama/llama-3-70b-instruct google/gemini-pro-1.5
|
| 270 |
return StreamingResponse(process_response(), media_type="text/event-stream")
|