Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -176,9 +176,9 @@ CPUs will be slower but there is no time limit.
|
|
| 176 |
|
| 177 |
def model_fn(prefix, history, server):
|
| 178 |
if server == 'Huggingface ZeroGPU':
|
| 179 |
-
generator = zerogpu_model_fn(prefix, history,
|
| 180 |
elif server == 'CPU':
|
| 181 |
-
generator = cpu_model_fn(prefix, history,
|
| 182 |
# elif server == 'RunPod':
|
| 183 |
# generator = runpod_model_fn(prefix, history)
|
| 184 |
else:
|
|
@@ -186,8 +186,13 @@ CPUs will be slower but there is no time limit.
|
|
| 186 |
for history in generator:
|
| 187 |
yield history
|
| 188 |
|
| 189 |
-
def cpu_model_fn(prefix, history,
|
| 190 |
queue = Queue(maxsize=10)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 191 |
class MyStreamer:
|
| 192 |
def put(self, tokens):
|
| 193 |
for token in tokens.flatten():
|
|
|
|
| 176 |
|
| 177 |
def model_fn(prefix, history, server):
|
| 178 |
if server == 'Huggingface ZeroGPU':
|
| 179 |
+
generator = zerogpu_model_fn(prefix, history, server)
|
| 180 |
elif server == 'CPU':
|
| 181 |
+
generator = cpu_model_fn(prefix, history, server)
|
| 182 |
# elif server == 'RunPod':
|
| 183 |
# generator = runpod_model_fn(prefix, history)
|
| 184 |
else:
|
|
|
|
| 186 |
for history in generator:
|
| 187 |
yield history
|
| 188 |
|
| 189 |
+
def cpu_model_fn(prefix, history, server):
|
| 190 |
queue = Queue(maxsize=10)
|
| 191 |
+
if server == 'CPU':
|
| 192 |
+
pipe = cpu_pipe
|
| 193 |
+
else:
|
| 194 |
+
pipe = gpu_pipe
|
| 195 |
+
|
| 196 |
class MyStreamer:
|
| 197 |
def put(self, tokens):
|
| 198 |
for token in tokens.flatten():
|