DeepSeek innovation speeds up processing of long text, paper says
Chinese firm says its ‘native sparse attention’ (NSA) method offers AI efficiency by focusing only on key words and skipping unnecessary comparisons

The NSA method combined algorithm innovations with improved hardware to improve efficiency without sacrificing performance, according to the paper published on Tuesday on arXiv, a platform for preprint papers that have not been peer reviewed.
It could improve AI’s ability to solve complex problems, write large programs and track long conversations, said the team behind R1, the open-source, low-cost model that shook the AI world last month.
AI models such as ChatGPT use a technique called attention to process text. Just as humans recall earlier words to understand a sentence, AI determines which words are important and how they relate to each other.