Spaces:
Running
Running
Update
Browse files
models.py
CHANGED
|
@@ -9,6 +9,8 @@ import PIL.Image
|
|
| 9 |
# pylint: disable=g-bad-import-order
|
| 10 |
import gradio_helpers
|
| 11 |
|
|
|
|
|
|
|
| 12 |
|
| 13 |
ORGANIZATION = 'abetlen'
|
| 14 |
BASE_MODELS = [
|
|
@@ -69,10 +71,16 @@ def generate(
|
|
| 69 |
model_name: str, sampler: str, image: PIL.Image.Image, prompt: str
|
| 70 |
) -> str:
|
| 71 |
"""Generates output with specified `model_name`, `sampler`."""
|
| 72 |
-
model, params_cpu = get_cached_model(model_name)
|
| 73 |
-
batch = model.shard_batch(model.prepare_batch([image], [prompt]))
|
| 74 |
-
with gradio_helpers.timed('sharding'):
|
| 75 |
-
|
| 76 |
-
with gradio_helpers.timed('computation', start_message=True):
|
| 77 |
-
|
| 78 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
# pylint: disable=g-bad-import-order
|
| 10 |
import gradio_helpers
|
| 11 |
|
| 12 |
+
import llama_cpp
|
| 13 |
+
|
| 14 |
|
| 15 |
ORGANIZATION = 'abetlen'
|
| 16 |
BASE_MODELS = [
|
|
|
|
| 71 |
model_name: str, sampler: str, image: PIL.Image.Image, prompt: str
|
| 72 |
) -> str:
|
| 73 |
"""Generates output with specified `model_name`, `sampler`."""
|
| 74 |
+
# model, params_cpu = get_cached_model(model_name)
|
| 75 |
+
# batch = model.shard_batch(model.prepare_batch([image], [prompt]))
|
| 76 |
+
# with gradio_helpers.timed('sharding'):
|
| 77 |
+
# params = model.shard_params(params_cpu)
|
| 78 |
+
# with gradio_helpers.timed('computation', start_message=True):
|
| 79 |
+
# tokens = model.predict(params, batch, sampler=sampler)
|
| 80 |
+
|
| 81 |
+
model_path = gradio_helpers.get_paths()[model_name]
|
| 82 |
+
model = llama_cpp.Llama(
|
| 83 |
+
model_path,
|
| 84 |
+
)
|
| 85 |
+
# return model.tokenizer.to_str(tokens[0])
|
| 86 |
+
return "output"
|