Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -490,9 +490,9 @@ class AIAnalyzer:
|
|
| 490 |
|
| 491 |
try:
|
| 492 |
response = self.client.chat.completions.create(
|
| 493 |
-
model=
|
| 494 |
-
max_tokens=min(
|
| 495 |
-
temperature=0.
|
| 496 |
messages=[{"role": "user", "content": f"{prompt}\n\n{content[:10000]}"}]
|
| 497 |
)
|
| 498 |
|
|
@@ -679,7 +679,7 @@ class AIAnalyzer:
|
|
| 679 |
response = self.client.chat.completions.create(
|
| 680 |
model=qwen_model,
|
| 681 |
max_tokens=min(max_output_tokens, 100000),
|
| 682 |
-
temperature=0.
|
| 683 |
messages=[{
|
| 684 |
"role": "user",
|
| 685 |
"content": f"{prompt}\n\n{data_summary}"
|
|
|
|
| 490 |
|
| 491 |
try:
|
| 492 |
response = self.client.chat.completions.create(
|
| 493 |
+
model=qwen_model,
|
| 494 |
+
max_tokens=min(max_output_tokens, 100000),
|
| 495 |
+
temperature=0.1,
|
| 496 |
messages=[{"role": "user", "content": f"{prompt}\n\n{content[:10000]}"}]
|
| 497 |
)
|
| 498 |
|
|
|
|
| 679 |
response = self.client.chat.completions.create(
|
| 680 |
model=qwen_model,
|
| 681 |
max_tokens=min(max_output_tokens, 100000),
|
| 682 |
+
temperature=0.1,
|
| 683 |
messages=[{
|
| 684 |
"role": "user",
|
| 685 |
"content": f"{prompt}\n\n{data_summary}"
|