Update README.md
Browse files
    	
        README.md
    CHANGED
    
    | @@ -2608,7 +2608,7 @@ model-index: | |
| 2608 |  | 
| 2609 | 
             
            # gte-base-en-v1.5
         | 
| 2610 |  | 
| 2611 | 
            -
            We introduce `gte-v1.5` series, upgraded `gte` embeddings that support the context length of up to **8192 | 
| 2612 | 
             
            The models are built upon the `transformer++` encoder [backbone](https://huggingface.co/Alibaba-NLP/new-impl) (BERT + RoPE + GLU).
         | 
| 2613 |  | 
| 2614 | 
             
            The `gte-v1.5` series achieve state-of-the-art scores on the MTEB benchmark within the same model size category and prodvide competitive on the LoCo long-context retrieval tests (refer to [Evaluation](#evaluation)).
         | 
|  | |
| 2608 |  | 
| 2609 | 
             
            # gte-base-en-v1.5
         | 
| 2610 |  | 
| 2611 | 
            +
            We introduce `gte-v1.5` series, upgraded `gte` embeddings that support the context length of up to **8192**, while further enhancing model performance.
         | 
| 2612 | 
             
            The models are built upon the `transformer++` encoder [backbone](https://huggingface.co/Alibaba-NLP/new-impl) (BERT + RoPE + GLU).
         | 
| 2613 |  | 
| 2614 | 
             
            The `gte-v1.5` series achieve state-of-the-art scores on the MTEB benchmark within the same model size category and prodvide competitive on the LoCo long-context retrieval tests (refer to [Evaluation](#evaluation)).
         | 

