Gustavo de Rosa
		
	commited on
		
		
					Commit 
							
							·
						
						57faa53
	
1
								Parent(s):
							
							34d6e8c
								
chore(root): Adds top_k information even if 50 is already the default.
Browse files- README.md +2 -1
- generation_config.json +1 -0
    	
        README.md
    CHANGED
    
    | @@ -56,7 +56,7 @@ library_name: transformers | |
| 56 | 
             
            ## Usage
         | 
| 57 |  | 
| 58 | 
             
            > [!IMPORTANT]  
         | 
| 59 | 
            -
            > To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
         | 
| 60 |  | 
| 61 | 
             
            ### Input Formats
         | 
| 62 |  | 
| @@ -88,6 +88,7 @@ outputs = model.generate( | |
| 88 | 
             
                inputs.to(model.device),
         | 
| 89 | 
             
                max_new_tokens=4096,
         | 
| 90 | 
             
                temperature=0.8,
         | 
|  | |
| 91 | 
             
                top_p=0.95,
         | 
| 92 | 
             
                do_sample=True,
         | 
| 93 | 
             
            )
         | 
|  | |
| 56 | 
             
            ## Usage
         | 
| 57 |  | 
| 58 | 
             
            > [!IMPORTANT]  
         | 
| 59 | 
            +
            > To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_k=50`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
         | 
| 60 |  | 
| 61 | 
             
            ### Input Formats
         | 
| 62 |  | 
|  | |
| 88 | 
             
                inputs.to(model.device),
         | 
| 89 | 
             
                max_new_tokens=4096,
         | 
| 90 | 
             
                temperature=0.8,
         | 
| 91 | 
            +
                top_k=50,
         | 
| 92 | 
             
                top_p=0.95,
         | 
| 93 | 
             
                do_sample=True,
         | 
| 94 | 
             
            )
         | 
    	
        generation_config.json
    CHANGED
    
    | @@ -5,6 +5,7 @@ | |
| 5 | 
             
              "eos_token_id": 100265,
         | 
| 6 | 
             
              "pad_token_id": 100349,
         | 
| 7 | 
             
              "temperature": 0.8,
         | 
|  | |
| 8 | 
             
              "top_p": 0.95,
         | 
| 9 | 
             
              "transformers_version": "4.51.1"
         | 
| 10 | 
             
            }
         | 
|  | |
| 5 | 
             
              "eos_token_id": 100265,
         | 
| 6 | 
             
              "pad_token_id": 100349,
         | 
| 7 | 
             
              "temperature": 0.8,
         | 
| 8 | 
            +
              "top_k": 50,
         | 
| 9 | 
             
              "top_p": 0.95,
         | 
| 10 | 
             
              "transformers_version": "4.51.1"
         | 
| 11 | 
             
            }
         | 
