Spaces:
Running
on
L40S
Running
on
L40S
Add sleep time to help server recovery of the fastness
Browse files
app.py
CHANGED
|
@@ -43,12 +43,14 @@ def predict(request: gr.Request, *args, progress=gr.Progress(track_tqdm=True)):
|
|
| 43 |
json_response = response.json()
|
| 44 |
#If the output component is JSON return the entire output response
|
| 45 |
if(outputs[0].get_config()["name"] == "json"):
|
|
|
|
| 46 |
return json_response["output"]
|
| 47 |
predict_outputs = parse_outputs(json_response["output"])
|
| 48 |
processed_outputs = process_outputs(predict_outputs)
|
| 49 |
-
|
| 50 |
return tuple(processed_outputs) if len(processed_outputs) > 1 else processed_outputs[0]
|
| 51 |
else:
|
|
|
|
| 52 |
if(response.status_code == 409):
|
| 53 |
raise gr.Error(f"Sorry, the Cog image is still processing. Try again in a bit.")
|
| 54 |
raise gr.Error(f"The submission failed! Error: {response.status_code}")
|
|
|
|
| 43 |
json_response = response.json()
|
| 44 |
#If the output component is JSON return the entire output response
|
| 45 |
if(outputs[0].get_config()["name"] == "json"):
|
| 46 |
+
time.sleep(1)
|
| 47 |
return json_response["output"]
|
| 48 |
predict_outputs = parse_outputs(json_response["output"])
|
| 49 |
processed_outputs = process_outputs(predict_outputs)
|
| 50 |
+
time.sleep(1)
|
| 51 |
return tuple(processed_outputs) if len(processed_outputs) > 1 else processed_outputs[0]
|
| 52 |
else:
|
| 53 |
+
time.sleep(1)
|
| 54 |
if(response.status_code == 409):
|
| 55 |
raise gr.Error(f"Sorry, the Cog image is still processing. Try again in a bit.")
|
| 56 |
raise gr.Error(f"The submission failed! Error: {response.status_code}")
|