AI Hypercomputer updates from Google Cloud Next 25

Share This Post


Our AI Hypercomputer underpins our Cloud customers’ most demanding AI workloads. Its hardware and software layers are optimized to deliver more intelligence per dollar for training and inference.

Today at Google Cloud Next 25, we introduced updates throughout the AI Hypercomputer stack:

  • AI-optimized hardware: Our new seventh-generation TPU, Ironwood, is designed specifically for thinking and inferential AI models. Ironwood offers five times more peak compute capacity and six times the high-bandwidth memory (HBM) capacity compared to the prior-generation TPU.
  • Software advances for inference: Updates to our AI Hypercomputer’s software layer help developers optimize compute resources, while speeding up AI workloads. These advances are shortening the time between training and inference.
  • Flexible consumption options: There are more ways for businesses to control costs with flexible consumption models in Dynamic Workload Scheduler.

Learn more about these AI infrastructure updates on the Google Cloud blog.



Source link

Related Posts

Why the AI era is forcing a redesign of the entire compute backbone

Want smarter insights in your inbox? Sign up...

Blue Origin flies crypto entrepreneur, five others on 14th crewed New Shepard flight

WASHINGTON — Blue Origin conducted its third crewed...

Canyon’s bike customisation programme makes me want to sell my car

If you’re someone who’s ever spent hours daydreaming...

I want a Steam Deck, but I’m scared to buy one

I’ve wanted a Steam Deck from the first...

VinFast opens largest showroom in Chennai; plans 35 outlets across India by year-end

Electric vehicle manufacturer VinFast Auto India, a subsidiary...
- Advertisement -spot_img