Advertisement
Alibaba launches maths-specific AI models said to outperform LLMs from OpenAI, Google
- The new Qwen2-Math large language models are expected to help solve complex maths problems
Reading Time:2 minutes
Why you can trust SCMP
1
![Alibaba Group Holding’s maths-specific large language models further burnish the company’s artificial intelligence credentials. Photo: Shutterstock](https://cdn.i-scmp.com/sites/default/files/styles/1020x680/public/d8/images/canvas/2024/08/09/4d87b4b8-454f-4b29-ad5d-052cbe5efcd7_e1da95c4.jpg?itok=6tabGyb0&v=1723202336)
Ann Caoin Shanghai
Alibaba Group Holding is aiming to raise the bar in artificial intelligence (AI) development by launching a group of maths-specific large language models (LLMs) called Qwen2-Math, which the e-commerce giant claims can outperform the capabilities of OpenAI’s GPT-4o in that field.
Advertisement
“Over the past year, we have dedicated significant efforts to researching and enhancing the reasoning capabilities of large language models, with a particular focus on their ability to solve arithmetic and mathematical problems,” the Qwen team, part of Alibaba’s cloud computing unit, said in a post published on developer platform GitHub on Thursday. Alibaba owns the South China Morning Post.
The latest LLMs – the technology underpinning generative AI services like ChatGPT – were built on the Qwen2 LLMs released by Alibaba in June and covers three models based on their scale of parameters – a machine-learning term for variables present in an AI system during training, which helps establish how data prompts yield the desired output.
The model with the largest parameter count, Qwen2-Math-72B-Instruct, outperformed proprietary US-developed LLMs in maths benchmarks, according to the Qwen team’s post. Those included GPT-4o, Anthropic’s Claude 3.5 Sonnet, Google’s Gemini 1.5 Pro and Meta Platforms’ Llama-3.1-405B.
“We hope that Qwen2-Math can contribute to the community for solving complex mathematical problems,” the post said.
![The family of Tongyi Qianwen, also known as Qwen, large language models from Alibaba Group Holding’s cloud computing unit, now includes maths-specific LLMs. Photo: Shutterstock The family of Tongyi Qianwen, also known as Qwen, large language models from Alibaba Group Holding’s cloud computing unit, now includes maths-specific LLMs. Photo: Shutterstock](https://img.i-scmp.com/cdn-cgi/image/fit=contain,width=1024,format=auto/sites/default/files/d8/images/canvas/2024/08/09/4f1a1c04-1c5a-4ff9-907f-d77ce671044f_1126e908.jpg)
The Qwen2-Math AI models were tested on both English and Chinese maths benchmarks, according to the post. These included GSM8K, a data set of 8,500 high-quality linguistically diverse grade school maths problems; OlympiadBench, a high-level bilingual multimodal scientific benchmark; and the gaokao, the mainland’s daunting university entrance examination.
![loading](https://assets-v2.i-scmp.com/production/_next/static/media/wheel-on-gray.af4a55f9.gif)
Advertisement