Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -88,7 +88,7 @@ def infer(edit_images,
|
|
| 88 |
strength=1.0,
|
| 89 |
num_inference_steps=30,
|
| 90 |
true_cfg_scale=4.0,
|
| 91 |
-
|
| 92 |
progress=gr.Progress(track_tqdm=True)):
|
| 93 |
|
| 94 |
image = edit_images["background"]
|
|
@@ -97,7 +97,7 @@ def infer(edit_images,
|
|
| 97 |
if randomize_seed:
|
| 98 |
seed = random.randint(0, MAX_SEED)
|
| 99 |
|
| 100 |
-
if
|
| 101 |
enhanced_prompt = rewrite_prompt(prompt)
|
| 102 |
print(f"Original prompt: {prompt}\nEnhanced prompt: {enhanced_prompt}")
|
| 103 |
prompt = enhanced_prompt
|
|
@@ -217,7 +217,7 @@ with gr.Blocks(css=css, theme=gr.themes.Citrus()) as demo:
|
|
| 217 |
value=30,
|
| 218 |
)
|
| 219 |
|
| 220 |
-
|
| 221 |
label="Enhance prompt (using HF Inference)",
|
| 222 |
value=True
|
| 223 |
)
|
|
@@ -239,7 +239,7 @@ with gr.Blocks(css=css, theme=gr.themes.Citrus()) as demo:
|
|
| 239 |
show_api=False
|
| 240 |
).then(
|
| 241 |
fn=infer,
|
| 242 |
-
inputs=[edit_image, prompt, negative_prompt, seed, randomize_seed, strength, num_inference_steps, true_cfg_scale,
|
| 243 |
outputs=[result, seed]
|
| 244 |
).then(
|
| 245 |
fn=lambda: gr.update(visible=True),
|
|
|
|
| 88 |
strength=1.0,
|
| 89 |
num_inference_steps=30,
|
| 90 |
true_cfg_scale=4.0,
|
| 91 |
+
prompt_enhance=True,
|
| 92 |
progress=gr.Progress(track_tqdm=True)):
|
| 93 |
|
| 94 |
image = edit_images["background"]
|
|
|
|
| 97 |
if randomize_seed:
|
| 98 |
seed = random.randint(0, MAX_SEED)
|
| 99 |
|
| 100 |
+
if prompt_enhance:
|
| 101 |
enhanced_prompt = rewrite_prompt(prompt)
|
| 102 |
print(f"Original prompt: {prompt}\nEnhanced prompt: {enhanced_prompt}")
|
| 103 |
prompt = enhanced_prompt
|
|
|
|
| 217 |
value=30,
|
| 218 |
)
|
| 219 |
|
| 220 |
+
prompt_enhance = gr.Checkbox(
|
| 221 |
label="Enhance prompt (using HF Inference)",
|
| 222 |
value=True
|
| 223 |
)
|
|
|
|
| 239 |
show_api=False
|
| 240 |
).then(
|
| 241 |
fn=infer,
|
| 242 |
+
inputs=[edit_image, prompt, negative_prompt, seed, randomize_seed, strength, num_inference_steps, true_cfg_scale, prompt_enhance],
|
| 243 |
outputs=[result, seed]
|
| 244 |
).then(
|
| 245 |
fn=lambda: gr.update(visible=True),
|