Spaces:
Sleeping
Sleeping
fixed output?
Browse files
app.py
CHANGED
|
@@ -6,6 +6,7 @@ import os
|
|
| 6 |
import re
|
| 7 |
import threading
|
| 8 |
import queue
|
|
|
|
| 9 |
|
| 10 |
zero = torch.Tensor([0]).cuda()
|
| 11 |
print(zero.device) # <-- 'cpu' 🤔
|
|
@@ -22,17 +23,15 @@ def run_evaluation(model_name):
|
|
| 22 |
results = []
|
| 23 |
manifest_logs = []
|
| 24 |
|
| 25 |
-
# Use the secret HF token from the Hugging Face space
|
| 26 |
if "HF_TOKEN" not in os.environ:
|
| 27 |
return "Error: HF_TOKEN not found in environment variables.", "Error: Cannot start manifest server without HF_TOKEN."
|
| 28 |
|
| 29 |
manifest_process = None
|
| 30 |
log_queue = queue.Queue()
|
| 31 |
try:
|
| 32 |
-
# Start manifest server in background with explicit CUDA_VISIBLE_DEVICES
|
| 33 |
manifest_cmd = f"""
|
| 34 |
cd duckdb-nsql/ &&
|
| 35 |
-
CUDA_VISIBLE_DEVICES=0 HF_TOKEN={os.environ['HF_TOKEN']}
|
| 36 |
--model_type huggingface \
|
| 37 |
--model_generation_type text-generation \
|
| 38 |
--model_name_or_path {model_name} \
|
|
@@ -46,15 +45,20 @@ def run_evaluation(model_name):
|
|
| 46 |
results.append("Started manifest server in background.")
|
| 47 |
|
| 48 |
# Wait for the server to initialize (adjust time as needed)
|
| 49 |
-
|
|
|
|
| 50 |
try:
|
| 51 |
line = log_queue.get(timeout=1)
|
| 52 |
manifest_logs.append(line)
|
| 53 |
if "Running on" in line: # Server is ready
|
|
|
|
| 54 |
break
|
| 55 |
except queue.Empty:
|
| 56 |
pass
|
| 57 |
|
|
|
|
|
|
|
|
|
|
| 58 |
# Run inference
|
| 59 |
inference_cmd = f"""
|
| 60 |
cd duckdb-nsql/ &&
|
|
|
|
| 6 |
import re
|
| 7 |
import threading
|
| 8 |
import queue
|
| 9 |
+
import time
|
| 10 |
|
| 11 |
zero = torch.Tensor([0]).cuda()
|
| 12 |
print(zero.device) # <-- 'cpu' 🤔
|
|
|
|
| 23 |
results = []
|
| 24 |
manifest_logs = []
|
| 25 |
|
|
|
|
| 26 |
if "HF_TOKEN" not in os.environ:
|
| 27 |
return "Error: HF_TOKEN not found in environment variables.", "Error: Cannot start manifest server without HF_TOKEN."
|
| 28 |
|
| 29 |
manifest_process = None
|
| 30 |
log_queue = queue.Queue()
|
| 31 |
try:
|
|
|
|
| 32 |
manifest_cmd = f"""
|
| 33 |
cd duckdb-nsql/ &&
|
| 34 |
+
CUDA_VISIBLE_DEVICES=0 HF_TOKEN={os.environ['HF_TOKEN']} python -m manifest.api.app \
|
| 35 |
--model_type huggingface \
|
| 36 |
--model_generation_type text-generation \
|
| 37 |
--model_name_or_path {model_name} \
|
|
|
|
| 45 |
results.append("Started manifest server in background.")
|
| 46 |
|
| 47 |
# Wait for the server to initialize (adjust time as needed)
|
| 48 |
+
server_ready = False
|
| 49 |
+
for _ in range(60): # Increased timeout to 60 seconds
|
| 50 |
try:
|
| 51 |
line = log_queue.get(timeout=1)
|
| 52 |
manifest_logs.append(line)
|
| 53 |
if "Running on" in line: # Server is ready
|
| 54 |
+
server_ready = True
|
| 55 |
break
|
| 56 |
except queue.Empty:
|
| 57 |
pass
|
| 58 |
|
| 59 |
+
if not server_ready:
|
| 60 |
+
raise Exception("Manifest server failed to start within the expected time.")
|
| 61 |
+
|
| 62 |
# Run inference
|
| 63 |
inference_cmd = f"""
|
| 64 |
cd duckdb-nsql/ &&
|