The AI Infrastructure Arms Race: Why the Shift to ‘AI Factories’ is Redefining Global Business
For years, data centres were viewed as the “digital warehouses” of the internet—quiet, sterile environments where servers stored data and hosted websites. But that era is over. We are witnessing a fundamental pivot toward what industry insiders are calling “AI Factories.”
The recent launch of NEXTDC’s KL1 facility in Kuala Lumpur is a prime example of this shift. This isn’t just another colocation site; it is a purpose-built engine designed for high-performance computing (HPC) and artificial intelligence. When a company invests AUD$1 billion into a single regional hub, they aren’t betting on storage—they are betting on the massive compute power required to fuel the next decade of generative AI.
The Rise of Digital Sovereignty and ‘Sovereign-Ready’ Cloud
As AI integrates into government services, healthcare, and national security, the question is no longer just “Does it work?” but “Where does the data live?” This is the birth of digital sovereignty.

Businesses are increasingly wary of sending sensitive data across borders where it may be subject to foreign laws. This trend is driving a surge in demand for “sovereign-ready” environments—infrastructure that allows companies to scale AI systems while maintaining strict control over governance and compliance within their own borders.
We are seeing this play out across Southeast Asia, where nations are competing to become the primary hub for AI. By establishing local, high-tier infrastructure, providers allow enterprises to satisfy regulatory requirements without sacrificing the speed of the cloud. This “local-first” approach to global scale is becoming the blueprint for multinational expansion.
Beyond Colocation: The Move Toward GPU-as-a-Service (GPUaaS)
The hardware requirements for AI are vastly different from traditional cloud computing. Standard CPUs cannot handle the parallel processing needed for Large Language Models (LLMs); you need GPUs (Graphics Processing Units), specifically high-end chips like those from NVIDIA.
However, GPUs are expensive and difficult to source. This has led to the rise of GPU-as-a-Service (GPUaaS). Instead of building their own data centres, companies are partnering with infrastructure providers to rent massive GPU clusters on demand.
A real-world example is the partnership between SharonAI and NEXTDC, where GPUaaS was deployed to achieve rapid scalability without the capital expenditure of building a private facility. In the future, You can expect “AI-Ready” data centres to function less like landlords and more like utility providers, delivering raw compute power as a scalable resource.
The Southeast Asian ‘Data Gold Rush’
While Singapore has long been the digital heart of Asia, constraints on land and energy have opened the door for neighbors. Malaysia, Indonesia, and Thailand are now in a fierce competition to attract the world’s tech giants.

Malaysia, in particular, is positioning itself as a strategic alternative. The investment in the Klang Valley indicates a broader trend: the decentralization of the Asian cloud. By offering a combination of regulatory clarity, available land, and aggressive energy policies, Malaysia is attracting “AI Factories” that require more space and power than a dense city-state can provide.
This regional shift is further bolstered by diplomatic and economic strategies, such as Australia’s Southeast Asia Economic Strategy to 2040, which encourages cross-border capital flow to build sustainable digital ecosystems.
Future Trends to Watch
- Liquid Cooling Integration: As GPUs get hotter, traditional air conditioning will fail. Expect a massive shift toward immersion cooling and direct-to-chip liquid cooling in new builds.
- Edge AI Convergence: While massive hubs like KL1 handle the “training” of AI, we will see a rise in smaller “Edge” data centres that handle the “inference” (the actual running of the AI) closer to the end-user to reduce latency.
- Green AI: The energy demand of AI is staggering. The next competitive advantage for data centres won’t be just speed, but the ability to prove Net Zero operations through renewable energy integration.
Frequently Asked Questions
What is a Tier IV data centre?
A Tier IV facility is the highest level of data centre certification from the Uptime Institute. It is fully fault-tolerant, meaning any single failure in the power or cooling systems will not affect the critical load.

Why is Malaysia becoming a hub for AI infrastructure?
Malaysia offers a strategic balance of available land, power capacity, and government support (such as the AI Nation 2030 vision), making it an attractive alternative to the more constrained markets like Singapore.
What is the difference between traditional cloud and AI-ready infrastructure?
Traditional cloud is designed for general-purpose workloads (web hosting, databases). AI-ready infrastructure is built for high-density power, specialized cooling for GPUs, and massive interconnectivity to handle the huge data flows required by machine learning.
Join the Conversation: Do you think the shift toward digital sovereignty will unhurried down global AI innovation, or will regional hubs like KL1 actually accelerate it? Let us know your thoughts in the comments below or subscribe to our newsletter for more deep dives into the future of digital infrastructure.
