Skip to main content
Skip to content
World Today News
  • Home
  • News
  • World
  • Sport
  • Entertainment
  • Business
  • Health
  • Technology
Menu
  • Home
  • News
  • World
  • Sport
  • Entertainment
  • Business
  • Health
  • Technology

Now Available in Preview: DeepSeek V4 Cuts Inference Costs to a Fraction of R1 with Open Weights and Huawei Ascend Support

April 24, 2026 Rachel Kim – Technology Editor Technology

DeepSeek V4’s Ascend Integration: NPU Efficiency Meets Open Weights Pragmatism

DeepSeek’s V4 release arrives not with fanfare but with a quiet recalibration of the LLM cost curve—claiming inference expenses slashed to a fraction of its R1 predecessor while extending native support for Huawei’s Ascend NPU series. This isn’t another benchmark chase; it’s a targeted play for enterprise environments where power envelopes and silicon diversity dictate deployment feasibility. The model, released in preview this week, positions itself as a drop-in alternative for latency-sensitive workloads previously confined to GPU-bound architectures, directly challenging the assumption that competitive LLM performance requires NVIDIA’s ecosystem.

DeepSeek V4's Ascend Integration: NPU Efficiency Meets Open Weights Pragmatism
Ascend Open Weights Huawei

The Tech TL;DR:

  • DeepSeek V4 achieves 28 tokens/sec/Watt on Ascend 910B, outperforming R1’s 9 tokens/sec/Watt on comparable NVIDIA T4 instances in MLPerf Llama 2 7B benchmarks.
  • Native Ascend software stack integration eliminates CUDA translation layers, reducing p99 latency by 40% in Kubernetes-served deployments under 500ms SLOs.
  • Open weights release under MIT license permits fine-tuning on sovereign AI clouds, addressing data residency concerns for EU and APAC enterprises.

The core innovation lies not in novel architecture but in meticulous software-hardware co-design. DeepSeek V4 replaces R1’s mixed-precision FP8 scheme with a dynamic quantization-aware training protocol that exploits Ascend’s native INT4 matrix cores, yielding a 3.2x improvement in energy efficiency per token generated. Crucially, this avoids the accuracy cliff seen in aggressive post-training quantization—V4 maintains 98.7% of R1’s MMLU score (72.1 vs 73.0) while cutting KV cache memory footprint by 65%. For context, running a 7B parameter instance on Ascend 910B consumes 18W sustained versus 56W on an A10G under identical throughput, a differential that shifts the economics of 24/7 inference services.

Under-the-hood transparency remains a strength: V4’s training corpus, tokenizer and ablation studies are fully documented in the official GitHub repository, maintained by DeepSeek’s research team with funding traced to their Series C round led by Highlight Capital and Sequoia China. This contrasts sharply with opaque proprietary models where training data provenance creates liability risks for regulated industries. As

“The real breakthrough isn’t the NPU support—it’s that we finally have an open weights model where the hardware acceleration path is first-class, not an afterthought CUDA port,”

noted Li Wei, Lead ML Engineer at Huawei’s Noah’s Ark Lab, in a recent IEEE Micro paper detailing the Ascend software stack optimizations.

China's DeepSeek previews new model adapted for Huawei chip tech

From a deployment perspective, V4 eliminates a critical friction point for enterprises evaluating AI sovereignty. The model’s compatibility with Huawei’s CANN (Compute Architecture for Neural Networks) software stack means it runs natively on Ascend-enabled servers without requiring containerized GPU emulation layers—a significant advantage for organizations subject to data localization laws. This directly impacts the triage calculus for IT teams: when assessing infrastructure for LLM deployment, the choice now extends beyond cloud architecture consultants optimizing AWS/GCP GPU instances to include specialists in AI hardware integrators who can validate Ascend cluster performance under real-world telecom or manufacturing workloads.

The implementation mandate reveals tangible differences. Where deploying R1 on non-NVIDIA hardware required wrestling with ROCm or oneAPI compatibility layers, V4’s Ascend path is accessible via a simple environment variable switch:

# Deploy DeepSeek V4 on Huawei Ascend 910B via CANN export ASCEND_RT_VISIBLE_DEVICES=0,1,2,3 python -m vllm.entrypoints.api_server  --model deepseek-ai/DeepSeek-V4-Chat  --tensor-parallel-size 4  --dtype auto  --max-model-len 8192  --port 8000 

This simplicity masks substantial engineering: the vLLM integration leverages Ascend’s custom memory allocator to avoid fragmentation during long-context generation, a detail validated in MLPerf Inference v4.1 submissions where V4 achieved 142 tokens/sec on Ascend 910B clusters—competitive with L40S GPUs at 60% lower power draw.

Cybersecurity implications emerge in the model’s attack surface. By reducing reliance on CUDA’s complex driver stack, V4 shrinks the privileged code path exposed to potential exploits like CVE-2025-23293 (NVIDIA GPU driver escape). However, this shifts focus to Ascend’s firmware and CANN runtime, necessitating updated threat models. Enterprises deploying V4 should engage IoT security auditors familiar with NPU-specific side-channel vectors, particularly when processing sensitive inputs in healthcare or financial contexts where model inversion risks persist regardless of hardware.

The editorial kicker: DeepSeek V4’s true significance lies in its challenge to the GPU monoculture. As enterprises diversify AI infrastructure beyond NVIDIA’s walled garden, models with first-class support for alternative accelerators turn into strategic assets—not just for cost savings, but for supply chain resilience. The real test arrives when sovereign AI initiatives demand proof of performance on domestically fabricated silicon; V4’s Ascend integration provides a tangible data point in that ongoing validation.

*Disclaimer: The technical analyses and security protocols detailed in this article are for informational purposes only. Always consult with certified IT and cybersecurity professionals before altering enterprise networks or handling sensitive data.*

Share this:

  • Share on Facebook (Opens in new window) Facebook
  • Share on X (Opens in new window) X

Related

Search:

World Today News

NewsList Directory is a comprehensive directory of news sources, media outlets, and publications worldwide. Discover trusted journalism from around the globe.

Quick Links

  • Privacy Policy
  • About Us
  • Accessibility statement
  • California Privacy Notice (CCPA/CPRA)
  • Contact
  • Cookie Policy
  • Disclaimer
  • DMCA Policy
  • Do not sell my info
  • EDITORIAL TEAM
  • Terms & Conditions

Browse by Location

  • GB
  • NZ
  • US

Connect With Us

© 2026 World Today News. All rights reserved. Your trusted global news source directory.

Privacy Policy Terms of Service