Update README.md
Browse files
README.md
CHANGED
|
@@ -159,7 +159,7 @@ Below are DeciCoder's pass@1 on MultiPL HumanEval scores
|
|
| 159 |
| Infery LLM | 3,889.3 | 11,676.8 |
|
| 160 |
|
| 161 |
- Throughput (tokens/sec) - Measured with optimal batch size per hardware - A10 on BS 128, A100 on BS 512
|
| 162 |
-
- Infery-LLM, Deci's optimization and inference SDK's features a suite of optimization techniques, including selective quantization, optimized beam search, continuous batching, and custom CUDA kernels. To explore the full capabilities of Infery-LLM, we invite you to book a demo with our experts.
|
| 163 |
|
| 164 |
## Documentation
|
| 165 |
|
|
|
|
| 159 |
| Infery LLM | 3,889.3 | 11,676.8 |
|
| 160 |
|
| 161 |
- Throughput (tokens/sec) - Measured with optimal batch size per hardware - A10 on BS 128, A100 on BS 512
|
| 162 |
+
- Infery-LLM, Deci's optimization and inference SDK's features a suite of optimization techniques, including selective quantization, optimized beam search, continuous batching, and custom CUDA kernels. To explore the full capabilities of Infery-LLM, we invite you to [book a demo](https://deci.ai/infery-llm-book-a-demo/) with our experts.
|
| 163 |
|
| 164 |
## Documentation
|
| 165 |
|