Alibaba’s Qwen AI models enable low-cost DeepSeek alternatives from Stanford, Berkeley
Stanford’s S1 and Berkeley’s TinyZero are two examples of how researchers are increasingly using Alibaba tech to lower AI training costs
![Alibaba’s open-source Qwen2.5 models are increasingly being used by researchers to find ways of training AI models at even lower costs. Photo: Reuters](https://cdn.i-scmp.com/sites/default/files/styles/1020x680/public/d8/images/canvas/2025/02/10/003b5550-a3bb-4b77-baa8-48c882514363_96a2c3cf.jpg?itok=ZmCZhgWq&v=1739176873)
The S1 reasoning model was developed on top of the Chinese e-commerce giant’s Qwen2.5-32b-Instruct model by researchers from Stanford University, where Li works, and the University of Washington, according to a research paper published last week.
After being trained with answers to 1,000 carefully curated questions and the “thinking process” distilled from Google’s Gemini Thinking Experimental model, the S1 model outperformed OpenAI’s o1-preview on maths and programming skills, according to the paper.
The cost of running just the graphics processing units (GPUs) to develop S1 could be as low as US$14 based on the compute noted in the research, which says the model was trained for 26 minutes on 16 Nvidia H100s. These chips can be rented for US$2 per hour.
![loading](https://assets-v2.i-scmp.com/production/_next/static/media/wheel-on-gray.af4a55f9.gif)