Spaces:
				
			
			
	
			
			
		Running
		
			on 
			
			A10G
	
	
	
			
			
	
	
	
	
		
		
		Running
		
			on 
			
			A10G
	fix style
Browse files
    	
        app.py
    CHANGED
    
    | @@ -120,7 +120,7 @@ def process_model(model_id, q_method, use_imatrix, imatrix_q_method, private_rep | |
| 120 | 
             
                        config_dir = local_dir/"config.json"
         | 
| 121 | 
             
                        adapter_config_dir = local_dir/"adapter_config.json"
         | 
| 122 | 
             
                        if os.path.exists(adapter_config_dir) and not os.path.exists(config_dir):
         | 
| 123 | 
            -
                            raise Exception( | 
| 124 |  | 
| 125 | 
             
                        conversion_script = "convert_hf_to_gguf.py"
         | 
| 126 | 
             
                        fp16_conversion = f"python llama.cpp/{conversion_script} {local_dir} --outtype f16 --outfile {fp16}"
         | 
|  | |
| 120 | 
             
                        config_dir = local_dir/"config.json"
         | 
| 121 | 
             
                        adapter_config_dir = local_dir/"adapter_config.json"
         | 
| 122 | 
             
                        if os.path.exists(adapter_config_dir) and not os.path.exists(config_dir):
         | 
| 123 | 
            +
                            raise Exception('adapter_config.json is present.<br/><br/>If you are converting a LoRA adapter to GGUF, please use <a href="https://huggingface.co/spaces/ggml-org/gguf-my-lora" target="_blank" style="text-decoration:underline">GGUF-my-lora</a>.')
         | 
| 124 |  | 
| 125 | 
             
                        conversion_script = "convert_hf_to_gguf.py"
         | 
| 126 | 
             
                        fp16_conversion = f"python llama.cpp/{conversion_script} {local_dir} --outtype f16 --outfile {fp16}"
         | 

