Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
revert changes
#1917
by
pngwn
HF Staff
- opened
README.md
CHANGED
|
@@ -4,7 +4,7 @@ emoji: ποΈ #ππ
|
|
| 4 |
colorFrom: red
|
| 5 |
colorTo: blue
|
| 6 |
sdk: gradio
|
| 7 |
-
sdk_version:
|
| 8 |
app_file: app.py
|
| 9 |
license: mit
|
| 10 |
disable_embedding: true
|
|
|
|
| 4 |
colorFrom: red
|
| 5 |
colorTo: blue
|
| 6 |
sdk: gradio
|
| 7 |
+
sdk_version: 3.45.1
|
| 8 |
app_file: app.py
|
| 9 |
license: mit
|
| 10 |
disable_embedding: true
|
app.py
CHANGED
|
@@ -84,14 +84,7 @@ word_list = word_list_dataset["train"]['text']
|
|
| 84 |
|
| 85 |
#gradio.helpers.CACHED_FOLDER="/data/cache"
|
| 86 |
|
| 87 |
-
|
| 88 |
-
def infer(
|
| 89 |
-
prompt,
|
| 90 |
-
negative="low_quality",
|
| 91 |
-
scale=7,
|
| 92 |
-
style_name=None,
|
| 93 |
-
profile: gr.OAuthProfile | None = None,
|
| 94 |
-
):
|
| 95 |
for filter in word_list:
|
| 96 |
if re.search(rf"\b{filter}\b", prompt):
|
| 97 |
raise gr.Error("Please try again with a different prompt")
|
|
@@ -111,10 +104,11 @@ def infer(
|
|
| 111 |
|
| 112 |
for prediction in json_data["predictions"]:
|
| 113 |
for image in prediction["images"]:
|
| 114 |
-
|
| 115 |
-
images.append(
|
| 116 |
|
| 117 |
if profile is not None: # avoid conversion on non-logged-in users
|
|
|
|
| 118 |
user_history.save_image( # save images + metadata to user history
|
| 119 |
label=prompt,
|
| 120 |
image=pil_image,
|
|
@@ -255,11 +249,32 @@ css = """
|
|
| 255 |
block = gr.Blocks()
|
| 256 |
|
| 257 |
examples = [
|
| 258 |
-
[
|
| 259 |
-
|
| 260 |
-
|
| 261 |
-
|
| 262 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 263 |
]
|
| 264 |
|
| 265 |
|
|
@@ -318,19 +333,19 @@ with block:
|
|
| 318 |
</div>
|
| 319 |
"""
|
| 320 |
)
|
| 321 |
-
|
| 322 |
-
with gr.Row(elem_id="prompt-container", equal_height=True):
|
| 323 |
-
|
| 324 |
-
|
| 325 |
-
|
| 326 |
-
|
| 327 |
-
|
| 328 |
-
|
| 329 |
-
|
| 330 |
-
|
| 331 |
|
| 332 |
gallery = gr.Gallery(
|
| 333 |
-
|
| 334 |
)
|
| 335 |
|
| 336 |
|
|
@@ -357,40 +372,16 @@ with block:
|
|
| 357 |
label="Guidance Scale", minimum=0, maximum=50, value=7.5, step=0.1
|
| 358 |
)
|
| 359 |
|
| 360 |
-
ex = gr.Examples(
|
| 361 |
-
|
| 362 |
-
|
| 363 |
-
|
| 364 |
-
|
| 365 |
-
cache_examples=True,
|
| 366 |
-
)
|
| 367 |
-
negative.submit(
|
| 368 |
-
infer,
|
| 369 |
-
inputs=[text, negative, guidance_scale, style_selection],
|
| 370 |
-
outputs=[gallery, community_group],
|
| 371 |
-
concurrency_id="infer",
|
| 372 |
-
concurrency_limit=8,
|
| 373 |
-
)
|
| 374 |
-
text.submit(
|
| 375 |
-
infer,
|
| 376 |
-
inputs=[text, negative, guidance_scale, style_selection],
|
| 377 |
-
outputs=[gallery, community_group],
|
| 378 |
-
concurrency_id="infer",
|
| 379 |
-
concurrency_limit=8,
|
| 380 |
-
)
|
| 381 |
-
btn.click(
|
| 382 |
-
infer,
|
| 383 |
-
inputs=[text, negative, guidance_scale, style_selection],
|
| 384 |
-
outputs=[gallery, community_group],
|
| 385 |
-
concurrency_id="infer",
|
| 386 |
-
concurrency_limit=8,
|
| 387 |
-
)
|
| 388 |
-
|
| 389 |
share_button.click(
|
| 390 |
None,
|
| 391 |
[],
|
| 392 |
[],
|
| 393 |
-
|
| 394 |
)
|
| 395 |
gr.HTML(
|
| 396 |
"""
|
|
@@ -417,5 +408,5 @@ with gr.Blocks(css=css) as block_with_history:
|
|
| 417 |
with gr.Tab("Past generations"):
|
| 418 |
user_history.render()
|
| 419 |
|
| 420 |
-
block_with_history.queue(max_size=10, api_open=False).launch(show_api=False)
|
| 421 |
-
#block_with_history.launch(server_name="0.0.0.0")
|
|
|
|
| 84 |
|
| 85 |
#gradio.helpers.CACHED_FOLDER="/data/cache"
|
| 86 |
|
| 87 |
+
def infer(prompt, negative="low_quality", scale=7, style_name=None, profile: gr.OAuthProfile | None = None):
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 88 |
for filter in word_list:
|
| 89 |
if re.search(rf"\b{filter}\b", prompt):
|
| 90 |
raise gr.Error("Please try again with a different prompt")
|
|
|
|
| 104 |
|
| 105 |
for prediction in json_data["predictions"]:
|
| 106 |
for image in prediction["images"]:
|
| 107 |
+
image_b64 = (f"data:image/jpeg;base64,{image}")
|
| 108 |
+
images.append(image_b64)
|
| 109 |
|
| 110 |
if profile is not None: # avoid conversion on non-logged-in users
|
| 111 |
+
pil_image = Image.open(BytesIO(base64.b64decode(image)))
|
| 112 |
user_history.save_image( # save images + metadata to user history
|
| 113 |
label=prompt,
|
| 114 |
image=pil_image,
|
|
|
|
| 249 |
block = gr.Blocks()
|
| 250 |
|
| 251 |
examples = [
|
| 252 |
+
[
|
| 253 |
+
"A serious capybara at work, wearing a suit",
|
| 254 |
+
None,
|
| 255 |
+
None
|
| 256 |
+
],
|
| 257 |
+
[
|
| 258 |
+
'A Squirtle fine dining with a view to the London Eye',
|
| 259 |
+
None,
|
| 260 |
+
None
|
| 261 |
+
],
|
| 262 |
+
[
|
| 263 |
+
'A tamale food cart in front of a Japanese Castle',
|
| 264 |
+
None,
|
| 265 |
+
None
|
| 266 |
+
],
|
| 267 |
+
[
|
| 268 |
+
'a graffiti of a robot serving meals to people',
|
| 269 |
+
None,
|
| 270 |
+
None
|
| 271 |
+
],
|
| 272 |
+
[
|
| 273 |
+
'a beautiful cabin in Attersee, Austria, 3d animation style',
|
| 274 |
+
None,
|
| 275 |
+
None
|
| 276 |
+
],
|
| 277 |
+
|
| 278 |
]
|
| 279 |
|
| 280 |
|
|
|
|
| 333 |
</div>
|
| 334 |
"""
|
| 335 |
)
|
| 336 |
+
|
| 337 |
+
with gr.Row(elem_id="prompt-container").style(mobile_collapse=False, equal_height=True):
|
| 338 |
+
text = gr.Textbox(
|
| 339 |
+
label="Enter your prompt",
|
| 340 |
+
show_label=False,
|
| 341 |
+
max_lines=1,
|
| 342 |
+
placeholder="Enter your prompt",
|
| 343 |
+
elem_id="prompt-text-input",
|
| 344 |
+
)
|
| 345 |
+
btn = gr.Button("Generate", scale=0, elem_id="gen-button")
|
| 346 |
|
| 347 |
gallery = gr.Gallery(
|
| 348 |
+
label="Generated images", show_label=False, elem_id="gallery", grid=[2]
|
| 349 |
)
|
| 350 |
|
| 351 |
|
|
|
|
| 372 |
label="Guidance Scale", minimum=0, maximum=50, value=7.5, step=0.1
|
| 373 |
)
|
| 374 |
|
| 375 |
+
ex = gr.Examples(examples=examples, fn=infer, inputs=[text, negative, guidance_scale], outputs=[gallery, community_group], cache_examples=True, postprocess=False)
|
| 376 |
+
negative.submit(infer, inputs=[text, negative, guidance_scale, style_selection], outputs=[gallery, community_group], postprocess=False)
|
| 377 |
+
text.submit(infer, inputs=[text, negative, guidance_scale, style_selection], outputs=[gallery, community_group], postprocess=False)
|
| 378 |
+
btn.click(infer, inputs=[text, negative, guidance_scale, style_selection], outputs=[gallery, community_group], postprocess=False)
|
| 379 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 380 |
share_button.click(
|
| 381 |
None,
|
| 382 |
[],
|
| 383 |
[],
|
| 384 |
+
_js=share_js,
|
| 385 |
)
|
| 386 |
gr.HTML(
|
| 387 |
"""
|
|
|
|
| 408 |
with gr.Tab("Past generations"):
|
| 409 |
user_history.render()
|
| 410 |
|
| 411 |
+
block_with_history.queue(concurrency_count=8, max_size=10, api_open=False).launch(show_api=False)
|
| 412 |
+
#block_with_history.launch(server_name="0.0.0.0")
|