TLDR
- Microsoft deployed Maia 200, its latest AI chip manufactured on TSMC’s 3nm process for Azure cloud infrastructure
- The processor delivers over 10 petaFLOPS at 4-bit precision with 216GB HBM3e memory in a 750W design
- Nvidia stock dipped 0.64% following the announcement as Wall Street assesses long-term competitive pressure
- Tech giants including Google and Amazon are building proprietary chips to control costs and diversify suppliers
- The chip runs in Microsoft’s Iowa facility and will support Copilot plus OpenAI’s GPT-5.2 language models
Microsoft launched Maia 200 this week. The chip represents the company’s latest effort to reduce reliance on external semiconductor suppliers.
The processor handles inference operations across Azure’s cloud network. Microsoft built it to lower costs while maintaining performance for AI workloads.
TSMC fabricated the chip using its advanced 3nm manufacturing technology. The design includes 216GB of HBM3e memory and 272MB of on-chip SRAM.
Microsoft says Maia 200 delivers over 10 petaFLOPS when running 4-bit operations. The chip hits more than 5 petaFLOPS at 8-bit precision.
All this computing power fits within a 750W thermal budget. Microsoft positions the chip as faster than competing solutions from Amazon and Google.
The processor already runs production workloads in Iowa. OpenAI’s GPT-5.2 models and Microsoft’s Copilot services use the new hardware.
Market Reaction Stays Calm
Nvidia shares dropped less than 1% after Microsoft’s reveal. The modest decline suggests investors see limited near-term risk to Nvidia’s business.
Supply constraints still limit Nvidia chip availability. Customers continue waiting months for orders as AI infrastructure buildouts accelerate globally.
Microsoft designed Maia 200 for specific use cases rather than complete Nvidia replacement. The company still relies on Nvidia hardware for training large AI models.
Cloud providers face mounting electricity bills as AI adoption spreads. Custom chips offer a path to better power efficiency and lower operating expenses.
The Broader Custom Chip Movement
Microsoft isn’t pioneering this approach. Google has used Tensor Processing Units in its datacenters since 2016.
Amazon developed Trainium and Inferentia processors for AWS customers. Both companies cite cost control and supply chain diversification as key motivations.
These proprietary chips handle targeted workloads where efficiency matters most. They complement rather than replace third-party processors for demanding tasks.
Building custom silicon gives cloud providers negotiating power with external suppliers. It also reduces exposure to disruptions in the global chip supply chain.
Microsoft released a software development kit for Maia 200. The tools help engineers optimize applications to extract maximum performance from the hardware.
Deployment Plans Moving Forward
Maia 200 is processing real customer workloads right now. Microsoft plans broader rollouts across additional datacenter locations during 2026.
The company will balance Maia chips with Nvidia processors based on specific job requirements. Training complex models still demands Nvidia’s specialized capabilities.
Inference work requires less computing power than training. That makes it an ideal candidate for cost-optimized custom processors like Maia 200.
Microsoft’s chip strategy mirrors moves by other hyperscale cloud operators. As AI computing demands surge, controlling more hardware becomes strategically valuable.
The Iowa datacenter serves as the testing ground for expanded Maia 200 adoption. Production performance data will guide decisions about future deployment scale.


