Spaces:
				
			
			
	
			
			
		Running
		
			on 
			
			A10G
	
	
	
			
			
	
	
	
	
		
		
		Running
		
			on 
			
			A10G
	Update app.py
Browse files
    	
        app.py
    CHANGED
    
    | @@ -9,6 +9,8 @@ from huggingface_hub import snapshot_download | |
| 9 | 
             
            from huggingface_hub import whoami
         | 
| 10 | 
             
            from huggingface_hub import ModelCard
         | 
| 11 |  | 
|  | |
|  | |
| 12 | 
             
            from textwrap import dedent
         | 
| 13 |  | 
| 14 | 
             
            LLAMA_LIKE_ARCHS = ["MistralForCausalLM", "LlamaForCausalLM"]
         | 
| @@ -141,10 +143,10 @@ def process_model(model_id, q_method, hf_token, private_repo): | |
| 141 | 
             
            iface = gr.Interface(
         | 
| 142 | 
             
                fn=process_model,
         | 
| 143 | 
             
                inputs=[
         | 
| 144 | 
            -
                     | 
| 145 | 
            -
                        lines=1,
         | 
| 146 | 
             
                        label="Hub Model ID",
         | 
| 147 | 
            -
                         | 
|  | |
| 148 | 
             
                    ),
         | 
| 149 | 
             
                    gr.Dropdown(
         | 
| 150 | 
             
                        ["Q2_K", "Q3_K_S", "Q3_K_M", "Q3_K_L", "Q4_0", "Q4_K_S", "Q4_K_M", "Q5_0", "Q5_K_S", "Q5_K_M", "Q6_K", "Q8_0"],
         | 
| @@ -175,4 +177,4 @@ iface = gr.Interface( | |
| 175 | 
             
            )
         | 
| 176 |  | 
| 177 | 
             
            # Launch the interface
         | 
| 178 | 
            -
            iface.queue(default_concurrency_limit=1, max_size=5).launch(debug=True)
         | 
|  | |
| 9 | 
             
            from huggingface_hub import whoami
         | 
| 10 | 
             
            from huggingface_hub import ModelCard
         | 
| 11 |  | 
| 12 | 
            +
            from gradio_huggingfacehub_search import HuggingfaceHubSearch
         | 
| 13 | 
            +
             | 
| 14 | 
             
            from textwrap import dedent
         | 
| 15 |  | 
| 16 | 
             
            LLAMA_LIKE_ARCHS = ["MistralForCausalLM", "LlamaForCausalLM"]
         | 
|  | |
| 143 | 
             
            iface = gr.Interface(
         | 
| 144 | 
             
                fn=process_model,
         | 
| 145 | 
             
                inputs=[
         | 
| 146 | 
            +
                    HuggingfaceHubSearch(
         | 
|  | |
| 147 | 
             
                        label="Hub Model ID",
         | 
| 148 | 
            +
                        placeholder="Search for model id on Huggingface",
         | 
| 149 | 
            +
                        search_type="model",
         | 
| 150 | 
             
                    ),
         | 
| 151 | 
             
                    gr.Dropdown(
         | 
| 152 | 
             
                        ["Q2_K", "Q3_K_S", "Q3_K_M", "Q3_K_L", "Q4_0", "Q4_K_S", "Q4_K_M", "Q5_0", "Q5_K_S", "Q5_K_M", "Q6_K", "Q8_0"],
         | 
|  | |
| 177 | 
             
            )
         | 
| 178 |  | 
| 179 | 
             
            # Launch the interface
         | 
| 180 | 
            +
            iface.queue(default_concurrency_limit=1, max_size=5).launch(debug=True)
         | 

