Advertisement
DeepSeek
TechBig Tech

DeepSeek founder shares best paper award at top global AI research conference

More than half of the first-named authors on accepted papers originated from China, up from less than 30 per cent last year

Reading Time:2 minutes
Why you can trust SCMP
2
DeepSeek has focused its efforts on research and was not rushing into fundraising or new commercial ventures. Photo: Shutterstock Images
Hannah Wang
A research paper co-authored by Liang Wenfeng, founder of Chinese artificial intelligence start-up DeepSeek, was honoured with the best paper award at the Association for Computational Linguistics (ACL) conference in Vienna, Austria, widely recognised as the premier global conference for AI researchers.

The paper, titled “Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention,” was published on February 27, with Liang listed as one of 15 authors. The “native sparse attention” mechanism is a core improvement that underpins the high efficiency and low-cost performance of DeepSeek’s AI models.

The paper’s win comes as Chinese scientists and researchers are outperforming US peers in basic research in the field of computational linguistics and natural language processing.

Advertisement

At this year’s ACL conference, more than half of the first-named authors on accepted papers originated from China, up from less than 30 per cent last year. The US ranked second, with 14 per cent of first-named authors, according to ACL data.

Among the four best papers recognised by ACL, two author teams were from China. They included Liang’s DeepSeek team and Yang Yaodong’s team from Peking University.

An undated photo of DeepSeek’s Liang Wenfeng. Photo: Weibo
An undated photo of DeepSeek’s Liang Wenfeng. Photo: Weibo

Yang, an assistant professor at the Institute of Artificial Intelligence and chief scientist of the Peking University-PsiBot Joint Laboratory, led research that explored a possible mechanism explaining the fragility of alignment in language models, attributed to the elasticity of language models.

Advertisement
Select Voice
Choose your listening speed
Get through articles 2x faster
1.25x
250 WPM
Slow
Average
Fast
1.25x