Feature Description
Minimax M2 is a new competitor to GLM 4.6, with a strong focus on agent-related capabilities, and It seems a MOE-based model.
Motivation
Maybe the MoE optimization implemented in ik_llama.cpp could provide better performance than the upstream llama.cpp?
Possible Implementation
There’s already a pending upstream PR in llama.cpp: #16831
Would it be feasible to add support by merging that PR and adjusting any conflicts?
Feature Description
Minimax M2 is a new competitor to GLM 4.6, with a strong focus on agent-related capabilities, and It seems a MOE-based model.
Motivation
Maybe the MoE optimization implemented in ik_llama.cpp could provide better performance than the upstream llama.cpp?
Possible Implementation
There’s already a pending upstream PR in llama.cpp: #16831
Would it be feasible to add support by merging that PR and adjusting any conflicts?