Add KTransformers Support in README.md
#51
by
oql
- opened
README.md
CHANGED
|
@@ -184,6 +184,10 @@ We recommend using [SGLang](https://docs.sglang.ai/) to serve MiniMax-M2. SGLang
|
|
| 184 |
|
| 185 |
We recommend using [vLLM](https://docs.vllm.ai/en/stable/) to serve MiniMax-M2. vLLM provides efficient day-0 support of MiniMax-M2 model, check https://docs.vllm.ai/projects/recipes/en/latest/MiniMax/MiniMax-M2.html for latest deployment guide. We also provide our [vLLM Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/vllm_deploy_guide.md).
|
| 186 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 187 |
### MLX
|
| 188 |
|
| 189 |
We recommend using [MLX-LM](https://github.com/ml-explore/mlx-lm) to serve MiniMax-M2. Please refer to our [MLX Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/mlx_deploy_guide.md) for more details.
|
|
@@ -213,4 +217,4 @@ Please refer to our [Tool Calling Guide](https://huggingface.co/MiniMaxAI/MiniMa
|
|
| 213 |
|
| 214 |
# Contact Us
|
| 215 |
|
| 216 |
-
Contact us at [[email protected]](mailto:[email protected]) | [WeChat](https://github.com/MiniMax-AI/MiniMax-AI.github.io/blob/main/images/wechat-qrcode.jpeg).
|
|
|
|
| 184 |
|
| 185 |
We recommend using [vLLM](https://docs.vllm.ai/en/stable/) to serve MiniMax-M2. vLLM provides efficient day-0 support of MiniMax-M2 model, check https://docs.vllm.ai/projects/recipes/en/latest/MiniMax/MiniMax-M2.html for latest deployment guide. We also provide our [vLLM Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/vllm_deploy_guide.md).
|
| 186 |
|
| 187 |
+
### KTransformers
|
| 188 |
+
|
| 189 |
+
We recommend using [KTransformers](https://github.com/kvcache-ai/ktransformers) to serve MiniMax-M2. KTransformers can run the native weights with **≥32GB VRAM** and **≥256GB DRAM**. For installation and usage, see [KT-Kernel Deployment Guide](https://github.com/kvcache-ai/ktransformers/blob/main/kt-kernel/README.md).
|
| 190 |
+
|
| 191 |
### MLX
|
| 192 |
|
| 193 |
We recommend using [MLX-LM](https://github.com/ml-explore/mlx-lm) to serve MiniMax-M2. Please refer to our [MLX Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/mlx_deploy_guide.md) for more details.
|
|
|
|
| 217 |
|
| 218 |
# Contact Us
|
| 219 |
|
| 220 |
+
Contact us at [[email protected]](mailto:[email protected]) | [WeChat](https://github.com/MiniMax-AI/MiniMax-AI.github.io/blob/main/images/wechat-qrcode.jpeg).
|