Text Generation
Transformers
Safetensors
minimax_m2
conversational
custom_code
fp8

Add KTransformers Support in README.md

#51
by oql - opened
Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -184,6 +184,10 @@ We recommend using [SGLang](https://docs.sglang.ai/) to serve MiniMax-M2. SGLang
184
 
185
  We recommend using [vLLM](https://docs.vllm.ai/en/stable/) to serve MiniMax-M2. vLLM provides efficient day-0 support of MiniMax-M2 model, check https://docs.vllm.ai/projects/recipes/en/latest/MiniMax/MiniMax-M2.html for latest deployment guide. We also provide our [vLLM Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/vllm_deploy_guide.md).
186
 
 
 
 
 
187
  ### MLX
188
 
189
  We recommend using [MLX-LM](https://github.com/ml-explore/mlx-lm) to serve MiniMax-M2. Please refer to our [MLX Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/mlx_deploy_guide.md) for more details.
@@ -213,4 +217,4 @@ Please refer to our [Tool Calling Guide](https://huggingface.co/MiniMaxAI/MiniMa
213
 
214
  # Contact Us
215
 
216
- Contact us at [[email protected]](mailto:[email protected]) | [WeChat](https://github.com/MiniMax-AI/MiniMax-AI.github.io/blob/main/images/wechat-qrcode.jpeg).
 
184
 
185
  We recommend using [vLLM](https://docs.vllm.ai/en/stable/) to serve MiniMax-M2. vLLM provides efficient day-0 support of MiniMax-M2 model, check https://docs.vllm.ai/projects/recipes/en/latest/MiniMax/MiniMax-M2.html for latest deployment guide. We also provide our [vLLM Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/vllm_deploy_guide.md).
186
 
187
+ ### KTransformers
188
+
189
+ We recommend using [KTransformers](https://github.com/kvcache-ai/ktransformers) to serve MiniMax-M2. KTransformers can run the native weights with **≥32GB VRAM** and **≥256GB DRAM**. For installation and usage, see [KT-Kernel Deployment Guide](https://github.com/kvcache-ai/ktransformers/blob/main/kt-kernel/README.md).
190
+
191
  ### MLX
192
 
193
  We recommend using [MLX-LM](https://github.com/ml-explore/mlx-lm) to serve MiniMax-M2. Please refer to our [MLX Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/mlx_deploy_guide.md) for more details.
 
217
 
218
  # Contact Us
219
 
220
+ Contact us at [[email protected]](mailto:[email protected]) | [WeChat](https://github.com/MiniMax-AI/MiniMax-AI.github.io/blob/main/images/wechat-qrcode.jpeg).