Each Ironwood chip packs 192GB of memory and can work in massive clusters of up to 9,216 chips. Google says it doubles the performance-per-watt compared to its previous chip, Trillium.

The chip specializes in "inference" - running existing AI models rather than training new ones. This focus matters as companies rush to deploy AI applications at scale.

Google built Ironwood specifically for cloud customers who need to run large language models efficiently. The chip includes a specialized core for recommendation systems and ranking tasks.

The launch highlights Google's push to compete with Nvidia, which dominates the AI chip market. Unlike Nvidia's products, Google's chips are only available through its cloud service.

Google plans to integrate Ironwood into its AI Hypercomputer system later this year, though it hasn't named the manufacturer producing the chips.

Why this matters:

Google Unleashes Gemma 3: The AI That Fits in Your Pocket
Google just launched Gemma 3, the latest version of its “open” AI model. This new release packs a serious punch - it can analyze text, images, and videos while running on devices as small as your phone. The tech giant claims Gemma 3 outsmarts rivals like Facebook’s Llama and OpenAI
AI News
Marcus Schuler

Marcus Schuler

San Francisco

Tech translator with German roots who fled to Silicon Valley chaos. Decodes startup noise from San Francisco. Launched implicator.ai to slice through AI's daily madness—crisp, clear, with Teutonic precision and sarcasm. E-Mail: [email protected]