The burst of DeepSeek V3 has attracted attention from the whole AI community to large-scale MoE models. Concurrently, we have been building Qwen2.5-Max, a large MoE LLM pretrained on massive data and post-trained with curated SFT and RLHF recipes. It achieves competitive… pic.twitter.com/oHVl16vfje
— Qwen (@Alibaba_Qwen) January 28, 2025
There is no reason China can’t just keep pumping out new AI models every week.
Proof of concept has been demonstrated.
The US claimed there was an “AI war” with China, and it is looking like they are going to regret making that claim.
Chinese tech company Alibaba on Wednesday released a new version of its Qwen 2.5 artificial intelligence model that it claimed surpassed the highly-acclaimed DeepSeek-V3.
The unusual timing of the Qwen 2.5-Max’s release, on the first day of the Lunar New Year when most Chinese people are off work and with their families, points to the pressure Chinese AI startup DeepSeek’s meteoric rise in the past three weeks has placed on not just overseas rivals, but also its domestic competition.
“Qwen 2.5-Max outperforms … almost across the board GPT-4o, DeepSeek-V3 and Llama-3.1-405B,” Alibaba’s cloud unit said in an announcement posted on its official WeChat account, referring to OpenAI and Meta’s most advanced open-source AI models.
…
Two days after the release of DeepSeek-R1, TikTok owner ByteDance released an update to its flagship AI model, which it claimed outperformed Microsoft-backed OpenAI’s o1 in AIME, a benchmark test that measures how well AI models understand and respond to complex instructions.
This echoed DeepSeek’s claim that its R1 model rivalled OpenAI’s o1 on several performance benchmarks.
All Chinese models outperform OpenAI!
That was the industry standard up until a few days ago, but it was always garbage, and never worked correctly, because it was so ridiculously censored.
China just dropped a new model.
ByteDance Doubao-1.5-pro matches GPT 4o benchmarks at 50x cheaper— $0.022/M cached input tokens, $0.11/M input, $0.275/M output
— 5x cheaper than DeepSeek, >200x of o1
— 32k + 256k context
— sparse MoE architectureAI truly too cheap to meter. pic.twitter.com/spffpxtWrV
— Deedy (@deedydas) January 23, 2025