Advertisement
China Telecom develops country’s first MoE models trained entirely on Huawei’s AI chips
The state-owned telecoms network operator trains its TeleChat3 models on Huawei’s Ascend 910B chips
Reading Time:2 minutes
Why you can trust SCMP
15

State-owned China Telecom has developed the country’s first artificial intelligence models with the innovative Mixture-of-Experts (MoE) architecture that are trained entirely on advanced chips from Huawei Technologies.
This marked the first time a Chinese developer publicly validated the feasibility of using only Huawei chips to train AI models with MoE architecture, which has become widely adopted because of its ability to deliver high performance with fewer computational resources.
The TeleChat3 models – ranging from 105 billion to trillions of parameters – were trained on Huawei’s Ascend 910B chips and its open-source deep learning AI framework MindSpore, according to a technical paper published last month by China Telecom’s Institute of Artificial Intelligence (TeleAI).
Advertisement
The Huawei stack was able to meet the “severe demands” of training large-scale MoE models across many different sizes, TeleAI researchers said.
“These contributions collectively address critical bottlenecks in frontier-scale model training, establishing a mature full-stack solution tailored to domestic computational ecosystems,” they said.
Advertisement
The initiative by China Telecom, one of the world’s largest fixed-line and mobile network operators, underscored growing efforts by a number of US-blacklisted Chinese firms – including Huawei and iFlytek – to train AI models using domestically designed semiconductors.

Advertisement
Select Voice
Choose your listening speed
Get through articles 2x faster
1.25x
250 WPM
Slow
Average
Fast
1.25x