Upload folder using huggingface_hub
Browse files
docs/vllm_deploy_guide.md
CHANGED
|
@@ -1,12 +1,6 @@
|
|
| 1 |
# MiniMax M2 Model vLLM Deployment Guide
|
| 2 |
|
| 3 |
-
We recommend using [vLLM](https://docs.vllm.ai/en/
|
| 4 |
-
|
| 5 |
-
## Applicable Models
|
| 6 |
-
|
| 7 |
-
This guide applies to the following models. Simply modify the model name during deployment. The deployment process is illustrated using MiniMax-M2 as an example.
|
| 8 |
-
|
| 9 |
-
- MiniMaxAI/MiniMax-M2
|
| 10 |
|
| 11 |
## System Requirements
|
| 12 |
|
|
|
|
| 1 |
# MiniMax M2 Model vLLM Deployment Guide
|
| 2 |
|
| 3 |
+
We recommend using [vLLM](https://docs.vllm.ai/en/stable/) to deploy the [MiniMax-M2](https://huggingface.co/MiniMaxAI/MiniMax-M2) model. vLLM is a high-performance inference engine with excellent serving throughput, efficient and intelligent memory management, powerful batch request processing capabilities, and deeply optimized underlying performance. We recommend reviewing vLLM's official documentation to check hardware compatibility before deployment.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
## System Requirements
|
| 6 |
|
docs/vllm_deploy_guide_cn.md
CHANGED
|
@@ -1,12 +1,6 @@
|
|
| 1 |
# MiniMax M2 模型 vLLM 部署指南
|
| 2 |
|
| 3 |
-
我们推荐使用 [vLLM](https://docs.vllm.ai/en/
|
| 4 |
-
|
| 5 |
-
## 本文档适用模型
|
| 6 |
-
|
| 7 |
-
本文档适用以下模型,只需在部署时修改模型名称即可。以下以 MiniMax-M2 为例说明部署流程。
|
| 8 |
-
|
| 9 |
-
- MiniMaxAI/MiniMax-M2
|
| 10 |
|
| 11 |
## 环境要求
|
| 12 |
|
|
|
|
| 1 |
# MiniMax M2 模型 vLLM 部署指南
|
| 2 |
|
| 3 |
+
我们推荐使用 [vLLM](https://docs.vllm.ai/en/stable/) 来部署 [MiniMax-M2](https://huggingface.co/MiniMaxAI/MiniMax-M2) 模型。vLLM 是一个高性能的推理引擎,其具有卓越的服务吞吐、高效智能的内存管理机制、强大的批量请求处理能力、深度优化的底层性能等特性。我们建议在部署之前查看 vLLM 的官方文档以检查硬件兼容性。
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
## 环境要求
|
| 6 |
|