Measured by the DCI the Chinese AI models are about 1.5 years ahead of US models.
DCI = Dust42 Capability Index: MBP Max 64GB, Qwen3-80B MLX 4bit quant, 40 tokens per second. It is not on Claude Opus level but very, very useful if you have no internet, i.e. on a flight. And occasionally it surpasses even Opus by far and large. Opus is a pain in the neck once the coding task at hand surpasses its capabilities. Qwen3 is much better to guide to get step by step to a solution.
Almost all Chinese models are open weight research models.
My theory is that these models serve the purpose of being relatively easy to run/tweak for researchers, and mainly serve to demonstrate the effectiveness of new techniques in training and inference, as well as the strength of AI labs that created them.
They are not designed to be state of the art commercial models.
By choosing bigger model sizes, running more training epochs, and drilling the models a bit more on benchmarking questions, I'm sure the Chinese could close the gap, but that would delay these models, make them more expensive and harder to run without showing any tangible research benefit.
Also my 2c: I was perfectly happy with Sonnet 3.7 as of a year ago, if the Chinese have a model really as good as that (not only one that benchmarks as well), I'd definitely like to try it.
Minimax has been great for super high speed web/js/ts related work. It compares in my experience to Claude Sonnet, and at times gets stuff similar to Opus.
Design wise it produces some of the most beautiful AI generated page I've seen.
GLM-4.7 like a mix of Sonnet 4.5 and GPT-5 (the first version not the later ones). It has deep deep knowledge, but it's often just not as good in execution.
They're very cheap to try out, so you should see how your mileage varies.
Ofcourse for the hardest possible tasks that GPT 5.2 only approaches, they're not up to scratch. And for the hard-ish tasks in C++ for example that Opus 4.5 tackles Minimax feels closer, but just doesn't "grok" the problem space good enough.
No, for example Alibaba has huge proprietary Qwen models, like Qwen 3 Max. You just never hear about them because that space in western LLM discussions is occupied with the US labs.
Well, my answer was partially tongue in cheek. However, under the circumstances I mentioned I had better results with a local model. But then up to today I really like Deepseek for dev - for me and my use cases it still outperforms Opus. The problem is we don't have good and universal benchmarks, inststead we have benchmaxxing. For my use cases Chinese models may be perfect but maybe for yours American one's may outperform them by a huge margin? So I urge you to use YCI - Your Capability Index.
Measured by the DCI the Chinese AI models are about 1.5 years ahead of US models.
DCI = Dust42 Capability Index: MBP Max 64GB, Qwen3-80B MLX 4bit quant, 40 tokens per second. It is not on Claude Opus level but very, very useful if you have no internet, i.e. on a flight. And occasionally it surpasses even Opus by far and large. Opus is a pain in the neck once the coding task at hand surpasses its capabilities. Qwen3 is much better to guide to get step by step to a solution.