Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -64,7 +64,7 @@ class MultiAgentConversationalSystem:
|
|
| 64 |
|
| 65 |
try:
|
| 66 |
response = await self.client.chat.completions.create(
|
| 67 |
-
model="grok-
|
| 68 |
messages=[{"role": "user", "content": prompt}]
|
| 69 |
)
|
| 70 |
agent_response = {
|
|
@@ -115,7 +115,7 @@ class MultiAgentConversationalSystem:
|
|
| 115 |
async def manager_summary(self, policy):
|
| 116 |
try:
|
| 117 |
response = await self.client.chat.completions.create(
|
| 118 |
-
model="grok-
|
| 119 |
messages=[{"role": "user", "content": f"""Summarized this.\n\n{policy}"""}],
|
| 120 |
stream=False
|
| 121 |
)
|
|
@@ -149,7 +149,7 @@ class MultiAgentConversationalSystem:
|
|
| 149 |
|
| 150 |
try:
|
| 151 |
response = await self.client.chat.completions.create(
|
| 152 |
-
model="grok-
|
| 153 |
messages=conversation_history,
|
| 154 |
stream=True
|
| 155 |
)
|
|
@@ -194,7 +194,7 @@ def predict(message, history, policy_summary):
|
|
| 194 |
print("history_openai_format:", history_openai_format)
|
| 195 |
|
| 196 |
response = simple_client.chat.completions.create(
|
| 197 |
-
model='grok-
|
| 198 |
messages=history_openai_format,
|
| 199 |
temperature=0.6,
|
| 200 |
stream=True
|
|
|
|
| 64 |
|
| 65 |
try:
|
| 66 |
response = await self.client.chat.completions.create(
|
| 67 |
+
model="grok-2-1212",
|
| 68 |
messages=[{"role": "user", "content": prompt}]
|
| 69 |
)
|
| 70 |
agent_response = {
|
|
|
|
| 115 |
async def manager_summary(self, policy):
|
| 116 |
try:
|
| 117 |
response = await self.client.chat.completions.create(
|
| 118 |
+
model="grok-2-1212",
|
| 119 |
messages=[{"role": "user", "content": f"""Summarized this.\n\n{policy}"""}],
|
| 120 |
stream=False
|
| 121 |
)
|
|
|
|
| 149 |
|
| 150 |
try:
|
| 151 |
response = await self.client.chat.completions.create(
|
| 152 |
+
model="grok-2-1212",
|
| 153 |
messages=conversation_history,
|
| 154 |
stream=True
|
| 155 |
)
|
|
|
|
| 194 |
print("history_openai_format:", history_openai_format)
|
| 195 |
|
| 196 |
response = simple_client.chat.completions.create(
|
| 197 |
+
model='grok-2-1212',
|
| 198 |
messages=history_openai_format,
|
| 199 |
temperature=0.6,
|
| 200 |
stream=True
|