This is a MXFP4_MOE quantization of the model MiniMax-M2
Original model: https://huggingface.co/unsloth/MiniMax-M2
Download the latest llama.cpp to use it.
It seems that the original model I quantized had chat template problems, so I re-quantized the unsloth version of it that has template fixes. Please delete the old one and download the new quant.
Also keep in mind that this a coding model.
- Downloads last month
- 2,072
Hardware compatibility
Log In
to view the estimation
4-bit
Model tree for noctrex/MiniMax-M2-MXFP4_MOE-GGUF
Base model
MiniMaxAI/MiniMax-M2