The tempo of technological change since final 12 months’s Cloud Subsequent has by no means been sooner, and Google Cloud has unbelievable momentum.
Our first-party fashions now course of greater than 16 billion tokens per minute through direct API use by our prospects, up from 10 billion final quarter. To help and drive this development, in 2026, simply over half of our general machine studying compute funding is predicted to go in the direction of the Cloud enterprise to profit our cloud prospects and companions.
You possibly can learn all about our momentum and the extraordinary vary of partnerships and improvements we’re saying at Cloud Subsequent.
I need to spotlight simply 4 key areas.
1. We’re firmly within the agentic Gemini period
Final fall we launched Gemini Enterprise, the end-to-end system for the agentic period — the connective tissue between your information, your individuals and your objectives.
It has nice momentum: In Q1, we noticed 40% development in paid month-to-month lively customers quarter-over-quarter.
Via this speedy development, we’ve seen how each worker in each group can develop into a builder. That is an unbelievable shift, but it surely comes with complexity. The dialog has gone from “Can we construct an agent?” to “How will we handle hundreds of them?”
That’s why we’re introducing our new Gemini Enterprise Agent Platform. It offers the safe, full-stack connective tissue it’s essential construct, scale, govern and optimize your brokers with confidence — a mission management for the agentic enterprise.
2. Utilizing AI to defend towards safety threats
Whereas AI can improve safety dangers, our Cloud prospects now have AI on their facet to guard their organizations. At this time we’re unveiling a spread of latest agentic options for menace detection, as a part of an AI-powered cybersecurity platform that mixes Google’s Risk Intelligence and Safety Operations with Wiz’s Cloud and AI Safety Platform.
As well as, we’re launching Wiz’s new AI Software Safety Platform (AI-APP), which offers autonomous safety, from code to cloud to runtime, throughout multicloud, hybrid and AI environments.
3. Introducing our eighth-generation TPUs
Within the period of AI brokers, infrastructure must evolve to tackle probably the most demanding AI workloads. This 12 months, we’re bringing the eighth technology of our Tensor Processing Items with a twin chip method:
- TPU 8t, optimized for coaching, scales as much as 9,600 TPUs and a couple of petabytes of shared, high-bandwidth reminiscence in a single superpod. It achieves thrice the processing energy of Ironwood and delivers as much as 2x extra efficiency/watt.
- TPU 8i, optimized for inference, connects 1,152 TPUs in a single pod, dramatically decreasing latency, with 3x extra on-chip SRAM, to ship the huge throughput and low latency wanted to concurrently run tens of millions of brokers cost-effectively.
We’ll provide these to Cloud prospects as a core a part of our number of compute processors, together with a portfolio of NVIDIA GPU cases. Learn extra in our weblog submit.
4. Staying on the cutting-edge as “buyer zero”
To be the very best companion, we all the time need to be “buyer zero” for our personal applied sciences. This helps us think about, take a look at, construct and scale the very best Google applied sciences for our cloud prospects, for as we speak and tomorrow. Our database service Bigtable, which powers so many Google companies, and our TPUs, which have been so necessary in coaching and powering our Gemini fashions, are nice examples.
Listed here are a couple of more moderen ones:
First, coding.
- We’ve been utilizing AI to generate code internally at Google for some time. At this time, 75% of all new code at Google is now AI-generated and accepted by engineers, up from 50% final fall.
- We’re now shifting to really agentic workflows. Our engineers are orchestrating totally autonomous digital activity forces, firing off brokers and conducting unbelievable issues.
- Just lately, a very advanced code migration completed by brokers and engineers working collectively was accomplished six occasions sooner than was attainable a 12 months in the past with engineers alone.
- And with our current launch of the Gemini app on MacOS, the group constructed the preliminary launch with our agentic growth platform Antigravity, going from an thought to a local Swift app prototype in a couple of days.
Second, safety.
- We’ve lengthy led the business in safety. Now, our Safety Operations Heart brokers mechanically triage tens of hundreds of unstructured menace stories every month, decreasing menace mitigation time by greater than 90%. And we’ve got constructed and actively use Gemini-based AI brokers (like CodeMender) to search out and, importantly, repair essential software program flaws.
Third, our operations.
- For the launch of Gemini in Chrome, our advertising groups used our fashions to quickly generate hundreds of variations of our inventive belongings, which might traditionally take weeks. Utilizing AI led to 70% sooner turnaround and a 20% improve in conversions, getting us to market sooner and extra successfully.
Congratulations to our Google Cloud group, and an enormous due to our companions who’re constructing the long run with us. We’ll have much more to share on how we’re bringing the most recent know-how to everybody at Google I/O on Might 19.










