update README
#1
by
tianbin
- opened
README.md
CHANGED
|
@@ -69,6 +69,9 @@ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
|
| 69 |
|
| 70 |
### vLLM (OpenAI-compatible serving):
|
| 71 |
|
|
|
|
|
|
|
|
|
|
| 72 |
```bash
|
| 73 |
MODEL_PATH="OpenMedZoo/SafeMed-R1" # or a local path
|
| 74 |
PORT=50050
|
|
|
|
| 69 |
|
| 70 |
### vLLM (OpenAI-compatible serving):
|
| 71 |
|
| 72 |
+
[vLLM](https://github.com/vllm-project/vllm) is a high-throughput and memory-efficient inference and serving engine for LLMs.
|
| 73 |
+
`vllm>=0.9.0` is recommended.
|
| 74 |
+
|
| 75 |
```bash
|
| 76 |
MODEL_PATH="OpenMedZoo/SafeMed-R1" # or a local path
|
| 77 |
PORT=50050
|