MiniMax M2.1: Built for Real-World Complex Tasks, Multi-Language Programming
- stpedgwdgfhgdd - 139 sekunder sedanInternal Server Error
- Tepix - 598 sekunder sedanThe weights got released on huggingface now.
- kachapopopow - 3561 sekunder sedanI think people should stop comparing to sonnet, but to opus instead since it's so far ahead on producing code I would actually want to use (gemini 3 pro tends to be lacking in generalization and wants things to be using it's own style rather than adapting).
Whatever benchmark opus is ahead in should be treated as a very important metric of proper generalization in models.
- viraptor - 24397 sekunder sedanI've played with this a bit and it's ok. I'd place it somewhere around sonnet 4.5 level, probably below. But with this aggressive pricing you can just run 3 copies to do the same thing, choose the one that succeeded and still come out way ahead with the cost. Not as great as following instructions as Claude models and can get lost, but still "good enough".
I'm very happy with using it to just "do things". When doing in depth debugging or a massive plan is needed, I'd go with something better, but later going through the motions? It works.
- gcanyon - 18105 sekunder sedanWould it kill them to use the words "AI coding agent" somewhere prominent?
"MiniMax M2.1: Significantly Enhanced Multi-Language Programming, Built for Real-World Complex Tasks" could be an IDE, a UI framework, a performance library, or, or...
- jondwillis - 26456 sekunder sedan> MiniMax has been continuously transforming itself in a more AI-native way. The core driving forces of this process are models, Agent scaffolding, and organization. Throughout the exploration process, we have gained increasingly deeper understanding of these three aspects. Today we are releasing updates to the model component, namely MiniMax M2.1, hoping to help more enterprises and individuals find more AI-native ways of working (and living) sooner.
This compresses to: “We are updating our model, MiniMax, to 2.1. Agent harnesses exist and Agents are getting more capable.”
A good model and agent harness, pointed at the task of writing this post, might suggest less verbosity and complexity— it comes off as fake and hype-chasing to me, even if your model is actually good. I disengage there.
I saw yall give a lightning talk recently and it was similarly hype-y. Perhaps this is a translation or cultural thing.
- tomcam - 25546 sekunder sedanI still can’t figure out what it does
- integricho - 5622 sekunder sedanTheir site crashes my phone browser while scrolling. Is that the expected quality of output of their product?
- esafak - 22415 sekunder sedan> It exhibits consistent and stable results in tools such as Claude Code, Droid (Factory AI), Cline, Kilo Code, Roo Code, and BlackBox, while providing reliable support for Context Management mechanisms including Skill.md, Claude.md/agent.md/cursorrule, and Slash Commands.
One of the demos shows them using Claude Code, which is interesting. And the next sections are titled 'Digital Employee' and 'End-to-End Office Automation'. Their ambitions obviously go beyond coding. A sign of things to come...
- sosodev - 17162 sekunder sedanI’ve spent a little bit of time testing Minimax M2. It’s quite good given the small size but it did make some odd mistakes and struggle with precise instructions.
- mr_o47 - 23548 sekunder sedanI won't say it's same on the level of claude models but it's definitely good at coming up with frontend designs
- jdright - 26554 sekunder sedan
- Invictus0 - 19916 sekunder sedanHow is everyone monitoring the skill/utility of all these different models? I am overwhelmed by how many they are, and the challenge of monitoring their capability across so many different modalities.
- boredemployee - 17286 sekunder sedanInternal Server Error
- p-e-w - 26622 sekunder sedanOne of the cited reviews goes:
“We're excited for powerful open-source models like M2.1 […]”
Yet as far as I can tell, this model isn’t open at all. Not even open weights, nevermind open source.
- maximgeorge - 5091 sekunder sedan[dead]
- Yash16 - 17222 sekunder sedan[dead]
- monster_truck - 26174 sekunder sedanThat they are still training models against Objective-C is all the proof you need that it will outlive Swift.
When is someone going to vibe code Objective-C 3.0? Borrowing all of the actual good things that have happened since 2.0 is closer than you'd think thanks to LLVM and friends.
Nördnytt! 🤓