Cloud Computing: Returning to AWS and Rediscovering Its Flaws
An engineer shares their experience returning to AWS, detailing the reasons they previously left and what they encountered upon their return.

The promise of European AI sovereignty, bolstered by billions in public investment and ambitious policy directives like the Chips Act, hinges on our ability to independently develop and deploy cutting-edge AI. Yet, a critical bottleneck looms, one that risks turning this aspiration into a perpetual illusion: our burgeoning reliance on GPU-as-a-Service (GPUaaS) offerings, predominantly controlled by non-European entities. This isn’t about lamenting technological dependence; it’s about dissecting how our current GPUaaS strategy actively entrenches it, creating a brittle foundation for truly indigenous AI capabilities.
Consider a hypothetical EU-based AI startup. They’ve secured funding, assembled a crack team of researchers, and developed a groundbreaking model for medical image analysis. To deploy this, they need significant GPU compute for inference. Their ideal scenario would involve leveraging infrastructure that respects GDPR, avoids the specter of the US CLOUD Act, and provides predictable, cost-effective access. However, the reality often forces them into a difficult choice: navigate the opaque availability and exorbitant egress fees of US hyperscalers, or gamble on nascent European sovereign cloud providers whose GPU offerings, while promising, often lag in maturity and raw capacity. This startup’s struggle isn’t an edge case; it’s a microcosm of the systemic challenge Europe faces. We are building on borrowed infrastructure, hoping for independence.
Europe’s narrative of AI sovereignty often begins and ends with chip fabrication and design. The European Chips Act, a monumental undertaking aiming for 20% global market share by 2030, signals a vital recognition of this upstream dependency. Initiatives like VSORA’s Jotunn8 and the Semidynamics/SiPearl collaboration represent genuine efforts to forge sovereign AI silicon. Yet, even these promising European alternatives face a formidable adversary: the de facto software moat constructed by dominant GPU vendors, primarily NVIDIA.
NVIDIA’s CUDA software ecosystem is more than a library; it’s a complex, deeply integrated platform that underpins the vast majority of AI development and deployment globally. For European startups and researchers, switching to a new hardware architecture, even one with superior performance or a more favorable legal jurisdiction, requires a substantial rewrite of their AI pipelines. This includes model training frameworks, optimization libraries, and inference engines. The return on investment for such a migration is often prohibitive, particularly for smaller entities operating on tight budgets.
When we provision GPUaaS from US hyperscalers like AWS, Google Cloud, or Microsoft Azure, we are not merely renting compute cycles. We are implicitly signing up for a deeply integrated stack, heavily optimized for NVIDIA hardware. This creates a powerful inertia. Even if a European provider offers a theoretically sovereign hardware alternative, the effort to port and re-optimize AI workloads for it is a significant barrier to adoption. The risk of duplicated, underutilized GPU clusters arises here, as fragmented Kubernetes environments across various providers lead to economic inefficiencies at scale. This isn’t just a technical challenge; it’s a strategic trap that reinforces existing market dynamics.
We are witnessing a scenario where European investment in AI development is inadvertently flowing into the coffers of companies that control the underlying hardware and software stack, rather than fostering truly independent European hardware and software ecosystems. This is the GPUaaS conundrum: the very service intended to accelerate AI adoption might be inadvertently locking European AI innovation into foreign dependencies.
The allure of the “sovereign cloud” in Europe is undeniable, especially for organizations grappling with GDPR compliance and the implications of the US CLOUD Act. Providers like Stackit, OVH, and IONOS offer a compelling proposition: data centers physically located within EU jurisdictions, promising adherence to local data protection laws. However, when these sovereign clouds lack substantial, readily available, high-performance GPU capacity, their sovereignty becomes a hollow victory for AI development.
The reality is that the lion’s share of advanced GPU inventory remains concentrated within US hyperscalers. European sovereign cloud providers, while improving, often face limitations in acquiring cutting-edge GPUs. This scarcity forces European AI practitioners into a difficult trade-off: prioritize data jurisdiction over compute performance and availability, or vice versa. This leads to the failure scenario where an AI startup, committed to GDPR, finds itself unable to access the necessary GPU resources for production inference, hindering its competitiveness.
Furthermore, the cost structures of US hyperscalers present another significant hurdle. The “crazy egress fees” mentioned by users are not a minor inconvenience; they can dramatically inflate operational expenses for AI inference. As European AI models mature and move towards production deployment, the cost of transferring data out of these platforms can become astronomically high, making truly cost-effective European AI solutions economically unviable.
The CLOUD Act looms large, even for US providers operating EU data centers. This legislation grants US authorities broad powers to demand access to data held by US companies, irrespective of where that data is physically stored. While EU data centers might offer a layer of comfort regarding data residency, they do not negate the potential for extraterritorial data access requests, a critical concern for sensitive AI applications in sectors like healthcare or finance. This is a direct conflict that undermines the very notion of data sovereignty that many European AI initiatives strive to achieve.
The consequence is a fragmented landscape: some European organizations prioritize data jurisdiction and accept limited GPU access and higher operational costs from sovereign providers. Others opt for the raw power and broader ecosystem of US hyperscalers, accepting the associated legal and financial risks. Neither path leads to a robust, independently controlled European AI infrastructure.
The pursuit of European AI sovereignty through a purely GPUaaS-centric approach, especially when reliant on foreign-controlled infrastructure, is an illusion. True sovereignty demands more than just access; it requires control over the entire stack, from silicon design to software deployment.
What breaks at scale or under concurrent load? The current model breaks when demand outstrips supply, leading to prolonged provisioning times for GPUs. It breaks economically when egress fees cripple profitability. It breaks legally when the CLOUD Act trumps GDPR. It breaks strategically when lock-in to foreign software stacks prevents the adoption of emerging European hardware.
When should organizations avoid this approach?
The path forward necessitates a multifaceted strategy that extends beyond mere chip manufacturing. It demands:
Europe’s AI sovereignty is not an insurmountable goal, but it requires a fundamental recalibration of our strategy. Continuing down the path of GPUaaS reliance, without a clear plan to build genuine indigenous capabilities across the entire AI value chain, risks solidifying our position as a consumer of AI technology rather than a creator. The illusion of sovereignty will persist, but the reality of dependence will deepen.