
China’s DeepSeek readies new LLM, escalating AI contest
Chinese startup DeepSeek is preparing to ship a new AI model trained on Nvidia’s top chips, intensifying the global LLM race and testing export controls, safety norms and digital power balances.
Chinese AI startup DeepSeek is preparing to launch a new large language model as soon as March 2026, trained on Nvidia’s most advanced chips available to China. The move could further narrow the performance gap with leading U.S. systems while sharpening geopolitical tensions over export controls, data security and who sets the norms for powerful AI. (en.wikipedia.org)
According to company information and recent reporting, DeepSeek — a Hangzhou-based firm backed and owned by hedge fund High-Flyer — has grown from its 2023 founding into one of China’s most important AI model developers. Its existing open-source models, including DeepSeek-R1 and the V3 family, have drawn outsized attention by delivering competitive performance at a fraction of the compute cost of Western flagships, relying largely on Nvidia’s cut-down H800 GPUs and highly optimized training pipelines. (zh.wikipedia.org)
The company now says its next model has been trained on Nvidia’s most advanced AI chip that Chinese firms can access, with release targeted for as early as March 2026. Details on architecture and size have not been made public, but DeepSeek’s previous models have used hybrid designs that switch between high-intensity "thinking" modes and faster, cheaper inference, positioning them as low-cost alternatives for reasoning-heavy tasks like coding and scientific problem solving. (en.wikipedia.org)
DeepSeek’s rapid ascent has already reshaped China’s AI sector. After the open release of DeepSeek-R1 and the V3 line under permissive licenses, Chinese rivals including Alibaba’s Qwen team, search firm Baidu and speech specialist iFlytek have all touted upgrades explicitly benchmarked against DeepSeek’s scores — a sign that the startup, not just U.S. labs, now anchors competitive comparisons inside China. Alibaba’s Qwen2.5 and Qwen3 series, for instance, highlight outperforming or matching DeepSeek-V3 on key benchmarks, while iFlytek has described its Spark X1 reasoning model as rivaling both OpenAI’s o1 and DeepSeek R1. (en.wikipedia.org)
Internationally, DeepSeek has become a key vector for spreading Chinese AI systems into markets where U.S. services are expensive, restricted or politically sensitive. A recent Microsoft report found that DeepSeek’s low-cost, open-access R1 model has gained strong traction in parts of the global south, including Russia, Iran and several African countries, often pre-installed on Chinese-manufactured phones and PCs. That diffusion offers cheaper AI access but also extends China’s technical and regulatory influence, since the models reflect domestic ideological constraints and content rules. (apnews.com)
Some governments are already pushing back. In mid‑2025, Czechia’s national cybersecurity agency formally warned against DeepSeek products and moved to restrict their use in critical systems, citing the risk that Chinese national security and intelligence laws could compel the firm to share user data with state authorities. Similar concerns have led to limits or bans on Chinese AI tools in several U.S. states and allied countries, often folded into broader scrutiny of Chinese telecommunications and cloud providers. (tomshardware.com)
The upcoming model raises three intertwined questions for global policymakers.
First, export controls. DeepSeek’s use of Nvidia’s latest China-compliant chips will test how far U.S. restrictions can actually slow capability gains abroad. If the new model substantially closes the gap with systems trained on more powerful hardware, Washington may face pressure to tighten controls again — or accept that algorithmic efficiency is eroding the leverage of hardware chokepoints.
Second, safety norms. DeepSeek’s prior models have been documented as more tightly aligned with official Chinese Communist Party positions than earlier Chinese systems, particularly around politically sensitive topics. That means the new model is likely to reflect Beijing’s red lines on censorship and national security even as it reaches more users worldwide. For countries worried about information influence operations, an increasingly capable, widely deployed Chinese LLM poses a very different challenge from U.S.- or EU-governed AI platforms. (en.wikipedia.org)
Third, intellectual property and data provenance. In February 2026, Anthropic accused DeepSeek of using thousands of fraudulent accounts to mass‑harvest conversations with its Claude assistant to train DeepSeek models, a claim that, if substantiated, would underline how hard it is to police data flows in a borderless AI training ecosystem. That allegation lands just as regulators in Europe, the United States and Asia are drafting or enforcing new AI rules around data consent, logging and model transparency. (en.wikipedia.org)
For now, DeepSeek’s next release is one more step in a broader pattern: powerful LLMs are no longer a primarily American export. As China’s startups iterate on reasoning, efficiency and open distribution, every new model both expands access to advanced AI and complicates efforts to coordinate on safety, security and economic risk. The model race is increasingly multipolar — and harder to steer.
Tags