The rise of custom silicon for the data center

Richard Brown
3 min readApr 15, 2024

Last week, Google and Meta made landmark announcements regarding their in-house chip development efforts, marking a growing shift away from reliance on off-the-shelf solutions toward custom silicon development in the cloud computing and AI market. Google unveiled Axion, its first custom Arm-based CPU for the data center, while Meta revealed details of its second-generation Meta Training and Inference Accelerator (MTIA).

Why Custom Chips Matter

The increasing adoption of custom silicon by tech giants like Google, Meta, and AWS (Amazon Web Services) represents a strategic response to the unique demands of AI and cloud workloads.

By designing their own chips, these three companies gain the ability to optimize performance for specific workloads while minimizing power consumption. Control over the entire hardware stack also allows the companies to implement new features and optimizations at their own pace rather than being constrained by the traditional release cycles of merchant silicon vendors such as Nvidia and Intel.

Google’s Arm-based Thrust with Axion

Built on Arm’s latest Neoverse V2 technology, the Google Axion processor is designed for general-purpose workloads including web and app servers, containerized microservices, open-source databases, data analytics engines, media processing, and CPU-based AI training and inferencing.

Google claims that the Axion processor delivers instances with up to 30% better performance than the fastest general-purpose Arm-based instances available in the cloud, as well as up to 50% better performance and up to 60% better energy-efficiency than comparable current-generation x86-based instances.

The company plans to open access to the Axion processor for Google Cloud services such as Google Compute Engine, Google Kubernetes Engine, Dataproc, Dataflow, and Cloud Batch. It says that Google Cloud data centers are 1.5X more efficient than the industry average and deliver 3X more computing power with the same amount of electrical power compared with five years ago.

Meta’s Evolving AI Hardware Strategy

The Meta Training and Inference Accelerator (MTIA) is part of the company’s growing custom silicon development program, which aims to power its AI services and reduce its reliance on external suppliers. Meta claims that its MTIA v2 boasts significant improvements in performance, on-chip memory, and bandwidth compared to its predecessor, enhancing the company’s ability to optimize its recommendation and ranking models for advertising, a core component of its business.

The Future: Competition and a Diversifying Landscape

With continuing shortages of high-end AI server chips, growing concerns about data center energy consumption, and the rising cost of delivering cloud-based AI services, it is no big surprise that Google, Meta, and AWS are accelerating custom silicon development. In addition to reducing their dependence on Nvidia, Intel, and AMD, the design and deployment of home-grown chips will enable them to reduce costs, minimize energy consumption, and further boost the efficiency of their data center operations.

--

--

Richard Brown

I live in Taiwan and am interested in exploring what ancient Chinese philosophy can tell us about technology and the rise of modern China.