MiniMaxAI/MiniMax-M2.1 Text Generation • 229B • Updated about 1 month ago • 53.2k • • 1.27k
view reply llama-server by default in most implementation keeps the reasoning content in reasoning_content variable in response attribute. You can get it from there. Otherwise use reasoning-format flag and pass DeepSeek value to get pure tokens