Abinaya Mahendiran
		
	commited on
		
		
					Commit 
							
							·
						
						752d5c1
	
1
								Parent(s):
							
							0388313
								
Updated README
Browse files
    	
        README.md
    CHANGED
    
    | @@ -25,7 +25,7 @@ Pretrained model on Tamil language using a causal language modeling (CLM) object | |
| 25 | 
             
            The GTP-2 model is trained on [oscar dataset - ta](https://huggingface.co/datasets/oscar) and [IndicNLP dataset - ta](https://indicnlp.ai4bharat.org/corpora/)
         | 
| 26 |  | 
| 27 | 
             
            ## Intended uses & limitations:
         | 
| 28 | 
            -
            You can use the raw model for next sentence prediction, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter= | 
| 29 |  | 
| 30 | 
             
            ## How to pretrain the model:
         | 
| 31 | 
             
            To perform training, do the following steps,
         | 
|  | |
| 25 | 
             
            The GTP-2 model is trained on [oscar dataset - ta](https://huggingface.co/datasets/oscar) and [IndicNLP dataset - ta](https://indicnlp.ai4bharat.org/corpora/)
         | 
| 26 |  | 
| 27 | 
             
            ## Intended uses & limitations:
         | 
| 28 | 
            +
            You can use the raw model for next sentence prediction, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=gpt2) to look for fine-tuned versions on a task that interests you.
         | 
| 29 |  | 
| 30 | 
             
            ## How to pretrain the model:
         | 
| 31 | 
             
            To perform training, do the following steps,
         | 
