Microsoft just announced Maia 200, an AI chip set to compete with the best from Amazon and Google. The chip is designed specifically for large-scale AI workloads.
Maia 200 is made on TSMC’s 3nm process. Microsoft claims the chip is “the most performant, first-party silicon from any hyperscaler.” The tech giant added that Maia 200 has three times the FP4 performance of Amazon’s third-gen Trainium chip and better FP8 performance than Google’s seventh-gen TPU.
Why this matters
Training AI models takes a tremendous amount of computing power, but that is far from the only cost of operating LLMs. Even after a model is trained, running that model costs money for each use.
Microsoft claims that Maia 200 is the most efficient inference system the company has deployed and that it delivers 30% better performance per dollar than the latest generation of hardware used by the tech giant.
Those figures could drastically reduce operation costs for OpenAI and other companies that use Maia 200.
“Maia 200 is part of our heterogenous AI infrastructure and will serve multiple models, including the latest GPT-5.2 models from OpenAI, bringing performance per dollar advantage to Microsoft Foundry and Microsoft 365 Copilot,” said Microsoft.
Maia 200 is an inference accelerator. It is engineered to make AI token generation more affordable. Training a model is a massive one-time expense. Inference is the process of the AI actually answering a user. This is a continuous daily cost. Microsoft optimized this chip specifically for that task. The goal is to slash the “per-token” cost of running services like ChatGPT and Copilot.
If Maia 200 delivers in the way Microsoft described, the chip will help reduce the costs of operating LLMs, which could significantly help OpenAI and other companies that use the chip.
Microsoft says this chip will save money in token costs. Do you think custom silicon is enough to make AI profitable by 2027? Let us know in the comments.
Follow Windows Central on Google News to keep our latest news, insights, and features at the top of your feeds!