GPUaaS: Hindering or Helping European AI Sovereignty?
Europe is investing heavily in AI, but the reliance on GPU-as-a-Service may be reinforcing the illusion of true AI sovereignty.

Imagine this: you’ve invested heavily in cutting-edge AI infrastructure, betting on the latest processors promising to rival industry giants. Your team is deep into training a complex model, pushing the boundaries of what’s possible. Then, it happens. Persistent processor crashes. Unexplained stability issues. Debugging sessions that feel like deciphering ancient texts, not optimizing algorithms. This isn’t a hypothetical nightmare; it’s the reality for engineers grappling with the undocumented quirks and nascent software ecosystems of emerging AI hardware. The risk is stark: when Western companies and analysts lack transparent access to verify the true capabilities and independence of entities like Huawei’s chip designs, the global tech balance can shift precariously, leaving critical dependencies vulnerable to unforeseen limitations.
Huawei’s dedicated chip research arm, HiSilicon, has become a focal point in the escalating tech cold war. Amidst stringent US sanctions designed to cripple China’s advanced semiconductor ambitions, Huawei has not only survived but aggressively pursued domestic self-sufficiency, particularly in the high-stakes domains of AI and mobile processors. This drive has led to the development of impressive hardware like the Ascend series of AI processors and the revived Kirin mobile chips. However, the story is far from a simple narrative of innovation triumphing over adversity. The true challenge lies in understanding the granular realities of this technology, especially when faced with the failure scenario of unstable performance and limited verification, leaving the global tech community in a state of heightened, yet incomplete, awareness.
Huawei’s ambition is palpable in its Ascend line of AI processors. The Ascend 910D, for instance, is positioned as a direct competitor to Nvidia’s formidable H100. This isn’t mere rhetoric; Huawei claims it outperforms the H100 in specific benchmarks, particularly when integrated into their proprietary Atlas 950/960 SuperPoD clusters. These clusters are designed for massive parallel processing, scaling up to nearly 16,000 Ascend processors interconnected by Huawei’s custom UnifiedBus technology. This interconnect aims to provide the high-bandwidth, low-latency communication essential for large-scale AI training, a critical component that often becomes a bottleneck in advanced computing systems.
The underlying manufacturing process, however, reveals the critical constraints Huawei operates under. While aiming for parity with global leaders, Huawei relies on SMIC’s 7nm N+2 process. This process, while advanced for China, is still based on Deep Ultraviolet (DUV) lithography. The fundamental difference is stark: DUV requires significantly more steps (around 34) to achieve its resolution compared to the 9 steps needed for Extreme Ultraviolet (EUV) lithography, which TSMC and Samsung employ for their most cutting-edge nodes. This difference isn’t just academic; it impacts performance, power efficiency, and the ability to shrink transistors further. Consequently, while Huawei’s roadmap for future AI chips like the Ascend 950PR (inference) and 950DT (training) is ambitious, projected to reach mass production by March 2026 and Q4 2026 respectively, the reliance on DUV technology inherently limits their ultimate performance ceiling compared to EUV-manufactured counterparts. This is a crucial trade-off: the pursuit of domestic supply chains forces compromises on the absolute bleeding edge of semiconductor physics, potentially capping the long-term performance scaling of these chips.
Furthermore, the ecosystem around these processors is a carefully constructed, internally driven effort. Huawei’s CANN (Compute Architecture for Neural Networks) serves as their alternative to Nvidia’s ubiquitous CUDA. The BiSheng Compiler is designed to optimize the performance of Ascend processors, reportedly achieving significant gains in operator performance. This software layer is critical. Without a mature, well-documented, and widely adopted software stack, even the most powerful hardware can falter. The question that looms large for Western observers is the maturity and reliability of this homegrown ecosystem, especially when faced with the demands of complex, cutting-edge AI workloads.
The resilience demonstrated by Huawei and its manufacturing partners like SMIC is undeniable. However, beneath the surface of impressive product announcements lie significant challenges that directly impact performance and cost, and therefore, global competitiveness. Reports suggest that the production yields for certain Ascend chips, such as the 910B/C, are as low as 20% at SMIC. This means a staggering majority of manufactured chips are defective. Low yield rates translate directly into astronomical production costs per functional chip. This isn’t a minor inefficiency; it’s a fundamental impediment to scaling production and achieving price points that can truly disrupt established markets like the high-performance AI training segment.
This isolation also extends to equipment maintenance and spare parts. Sanctions restrict access to the latest fabrication equipment and the specialized engineering support required to keep complex semiconductor machinery running. Reliance on DUV lithography, while a workaround for advanced EUV, means these machines are older and more prone to failure. Sourcing spare parts and obtaining expert maintenance becomes an intricate logistical and geopolitical challenge. When a critical piece of equipment malfunctions, the time required to diagnose, repair, or replace it can be exponentially longer than for a fab integrated into the global supply chain. This creates an inherent fragility in the production line, potentially leading to unpredictable downtime and further exacerbating the yield issues.
These practical limitations have tangible consequences for the software ecosystem, particularly CANN. While CANN is Huawei’s answer to CUDA, early reports from engineers who have experimented with it paint a picture of a system still under heavy development. Anecdotal evidence points to issues ranging from software bugs and system crashes to overheating problems and slower inter-chip communication than advertised, especially when pushing the system to its limits with large-scale AI model training. The complexity of debugging these issues is compounded by what some describe as “poor documentation” and “imperfect coding,” forcing engineers to delve deep into source code to identify and resolve problems. This is the exact opposite of the seamless, highly optimized experience developers have come to expect from mature platforms. The failure scenario of frequent crashes and stability issues, exacerbated by a challenging debugging environment, directly hinders widespread adoption beyond domestic champions. This creates a critical trade-off: while domestic AI inference and mobile processing capabilities are rapidly advancing, the frontier of large-scale, cutting-edge AI training remains a significant hurdle, demanding absolute stability and a mature software ecosystem that is still under construction.
Huawei’s journey in semiconductor design and production is a testament to its strategic imperative for self-reliance, particularly in the face of severe external pressures. The company demonstrates remarkable resilience, especially in developing capable AI inference chips and reviving its mobile processor line. The significant domestic adoption by Chinese cloud providers like Alibaba, ByteDance, and Tencent, along with DeepSeek’s LLM optimization for Ascend/CANN, signals a growing internal ecosystem. Huawei’s projected $12 billion in AI chip revenue for 2026, capturing a notable share of China’s AI computing market, underscores this domestic momentum.
However, to achieve true global parity, especially in the demanding field of AI training, several critical barriers must be overcome. The fundamental limitations of manufacturing scale and the reliance on less advanced process nodes compared to global leaders like TSMC remain significant. More importantly, the maturation of its software ecosystem, epitomized by CANN, is paramount. The reliability, performance, and developer experience offered by CANN must evolve to a point where it can compete not just on paper, but in the trenches of real-world, resource-intensive AI development.
The current landscape presents a clear divergence: for tasks where inference performance and domestic supply chain security are prioritized, Huawei’s chips offer compelling alternatives. However, for high-performance, cutting-edge AI model training where raw performance, exceptional power efficiency, and the benefits of a vast, mature software ecosystem like CUDA are non-negotiable, the trade-offs with current Huawei offerings are significant. Western companies and policymakers must grapple with this evolving reality. The inability for independent verification of the true capabilities and independence of Huawei’s chip designs due to limited access presents a persistent geopolitical risk. As Huawei’s capabilities grow, so does the urgency for transparency and a clear understanding of the global technological balance, which may be shifting faster than current intelligence anticipates. The path forward requires a nuanced understanding of both technological ambition and its practical, often challenging, execution.