from os import abort, getenv
from sys import exception
from textwrap import dedent
import traceback
import gradio as gr
from torch import cuda
from detikzify.webui import BANNER, build_ui, make_light
def is_official_demo():
    return getenv("SPACE_AUTHOR_NAME") == "nllg"
def add_abort_hook(func, *errors):
    def wrapper(*args, **kwargs):
        if isinstance(exception(), errors):
            abort()
        return func(*args, **kwargs)
    return wrapper
if is_official_demo() and not cuda.is_available():
    center = ".gradio-container {text-align: center}"
    with gr.Blocks(css=center, theme=make_light(gr.themes.Soft()), title="DeTikZify") as demo:
        badge = "https://huggingface.co/datasets/huggingface/badges/resolve/main/duplicate-this-space-xl.svg"
        link = "https://huggingface.co/spaces/nllg/DeTikZify?duplicate=true"
        html = f'  '
        message = dedent("""\
        The resources required by our models surpass those provided by Hugging
        Face Spaces' free CPU tier. For full functionality, we suggest
        duplicating this space using a paid private GPU runtime.
        """)
        gr.HTML(f'{BANNER}\n
{message}
\n{html}') else: demo = build_ui(lock=is_official_demo(), light=True).queue() # Hack to temporarily work around memory leak, see: # * https://huggingface.co/spaces/nllg/DeTikZify/discussions/2 # * https://github.com/gradio-app/gradio/issues/8503 traceback.print_exc = add_abort_hook(traceback.print_exc, MemoryError, cuda.OutOfMemoryError) if __name__ == "__main__": demo.launch(server_name="0.0.0.0", server_port=7860)