← Blog
explainx / blog

Google Cloud Next 2026: TPU 8t / TPU 8i, Gemini Enterprise Agent Platform, and the “agentic enterprise”

At Cloud Next ‘26, Google split its eighth-generation TPUs into training (8t) and inference (8i) silicon, launched Gemini Enterprise Agent Platform atop Vertex, and published striking usage stats—3× training pod compute vs Ironwood, 80% better inference $/$, 1,152-chip inference pods, 75% AI-generated new code at Google, 16B+ customer tokens per minute. Primary sources: Google and Google Cloud official posts.

4 min readExplainX Team
Google CloudGeminiTPUAI infrastructureEnterprise AIGoogle Cloud Next

Includes frontmatter plus an attribution block so copies credit explainx.ai and the canonical URL.

Google Cloud Next 2026: TPU 8t / TPU 8i, Gemini Enterprise Agent Platform, and the “agentic enterprise”

Google Cloud Next ‘26 (Las Vegas, April 2026) packaged one narrative: custom TPUs, Gemini (and partner models in places), and Gemini Enterprise as the end-to-end “agentic enterprise” layer. The same story echoed across X and trade press; the notes below are tied to primary Google and Google Cloud posts—not Grok- or second-hand summaries alone.

Read first (official, in order):


TPU 8t (training) and TPU 8i (inference)

Per Google’s TPU 8 post, 8t and 8i are purpose-split for the agent era: training needs huge scale-up; inference needs memory bandwidth, low latency, and efficiency when many small steps chain together.

TPU 8t (training):

  • ~3× compute per pod vs the prior generation (Google names Ironwood in the same post).
  • Up to 9,600 chips and 2 petabytes of shared HBM in a superpod; 121 exaFLOPS FP4; 2× interchip bandwidth; 10× faster storage to the fabric (TPUDirect); Virgo and JAX / Pathways for large jobs. Pichai’s shorter post also references scaling to on the order of one million 8t chips in one logical cluster for frontier training.
  • Google targets >97% “goodput” (productive training time) via RAS, rerouting, and OCS.

TPU 8i (inference):

  • Pichai: 1,152 TPUs in one 8i pod; 3× more on-chip SRAM than the prior generation; aimed at latency-sensitive and agent workloads.
  • The long TPU post adds 288 GB HBM and 384 MB on-chip SRAM per chip, doubled Axion hosts, Boardfly topology, a Collectives Acceleration Engine, and about 80% better performance per dollar vs the prior inference generation—Google’s own efficiency claim, not a cross-vendor GPU benchmark.

GA: public posts say later in 2026; request TPU information for commercial follow-up.


Gemini Enterprise Agent Platform and ecosystem

The new Gemini Enterprise (April 22, 2026) frames Gemini Enterprise as end-to-end: models, product surfaces, governance, and deployment. The Agent Platform evolves Vertex into a build/tune/govern stack with MCP support, Model Armor, Agent Identity, paths into the Gemini Enterprise app, and a governed agent gallery for employees.

Partner agents name Oracle, Salesforce, ServiceNow, Workday, Adobe, Accenture, and others. Salesforce and Google’s joint PR (Cloud Next ‘26) covers Agentforce Sales in Gemini Enterprise and cross-platform Slack / Workspace work. Workspace Intelligence and Agentic Data Cloud are summarized in the Next ‘26 news post.


75% new internal code, 16B+ tokens per minute, customer scale

From Pichai’s blog: 75% of new code at Google is AI-generated and engineer-approved (up from 50% last fall); a separate item highlights agentic workflows and a faster complex migration vs a year prior; and first-party Cloud models process more than 16 billion tokens per minute from direct customer API use, up from 10 billion the prior quarter.

Next ‘26 and the Cloud welcome add adoption scale: e.g. nearly 75% of Google Cloud customers using AI products; 330 customers each with >1 trillion tokens in 12 months; 35 above 10 trillion. Business press also tied the event to Alphabet share moves; this post is not financial advice.


ExplainX: multicloud, skills, and verification

  1. Google is packaging silicon, models, governance, and SaaS partners in one story—compelling for GCP-centric shops.
  2. Portable patterns still matter: agent skills, MCP, and explainx.ai/skills help connectors and workflows survive model and host changes.
  3. The 75% figure is a process metric at one company; pair any keynote stat with hallucination literacy and your own evals.
  4. Courses teach the same primitives: trust boundaries, registries, and verification first.

Read next: Claude Code /ultrareview · Agent skills and security · Chrome “Skills” vs SKILL.md

SKUs, dates, and claims evolve. Re-verify on Cloud Next and product pages before plans or procurement.

Related posts