Only a 12 months in the past, the narrative round Google and enterprise AI felt caught. Regardless of inventing core applied sciences just like the Transformer, the tech large appeared perpetually on the again foot, overshadowed by OpenAI‘s viral success, Anthropic‘s coding prowess and Microsoft‘s aggressive enterprise push.
However witness the scene at Google Cloud Subsequent 2025 in Las Vegas final week: A assured Google, armed with benchmark-topping fashions, formidable infrastructure and a cohesive enterprise technique, declaring a shocking turnaround. In a closed-door analyst assembly with senior Google executives, one analyst summed it up. This feels just like the second, he stated, when Google went from “catch up, to catch us.”
This sentiment that Google has not solely caught up with however even surged forward of OpenAI and Microsoft within the enterprise AI race prevailed all through the occasion. And it’s extra than simply Google’s advertising and marketing spin. Proof suggests Google has leveraged the previous 12 months for intense, centered execution, translating its technological property right into a performant, built-in platform that’s quickly profitable over enterprise decision-makers. From boasting the world’s strongest AI fashions operating on hyper-efficient customized silicon, to a burgeoning ecosystem of AI brokers designed for real-world enterprise issues, Google is making a compelling case that it was by no means truly misplaced – however that its stumbles masked a interval of deep, foundational improvement.
Now, with its built-in stack firing on all cylinders, Google seems positioned to steer the subsequent part of the enterprise AI revolution. And in my interviews with a number of Google executives at Subsequent, they stated Google wields benefits in infrastructure and mannequin integration that opponents like OpenAI, Microsoft or AWS will wrestle to copy.
The shadow of doubt: acknowledging the latest previous
It’s unattainable to understand the present momentum with out acknowledging the latest previous. Google was the birthplace of the Transformer structure, which sparked the fashionable revolution in giant language fashions (LLMs). Google additionally began investing in specialised AI {hardware} (TPUs), which at the moment are driving industry-leading effectivity, a decade in the past. And but, two and a half years in the past, it inexplicably discovered itself taking part in protection.
OpenAI’s ChatGPT captured the general public creativeness and enterprise curiosity at breathtaking velocity and have become the fastest-growing app in historical past. Rivals like Anthropic carved out niches in areas like coding.
Google’s personal public steps generally appeared tentative or flawed. The notorious Bard demo fumbles in 2023 and the later controversy over its picture generator producing traditionally inaccurate depictions fed a story of an organization doubtlessly hampered by inside forms or overcorrection on alignment. It felt like Google was misplaced: The AI stumbles appeared to suit a sample, first proven by Google’s preliminary slowness within the cloud competitors, the place it remained a distant third in market share behind Amazon and Microsoft. Google Cloud CTO Will Grannis acknowledged the early questions on whether or not Google Cloud would stand behind in the long term. “Is it even a real thing?,” he recalled individuals asking him. The query lingered: May Google translate its simple analysis brilliance and infrastructure scale into enterprise AI dominance?
The pivot: a aware resolution to steer
Behind the scenes, nonetheless, a shift was underway, catalyzed by a aware resolution on the highest ranges to reclaim management. Mat Velloso, VP of product for Google DeepMind’s AI Developer Platform, described sensing a pivotal second upon becoming a member of Google in Feb. 2024, after leaving Microsoft. “When I came to Google, I spoke with Sundar [Pichai], I spoke with several leaders here, and I felt like that was the moment where they were deciding, okay, this [generative AI] is a thing the industry clearly cares about. Let’s make it happen,” Velloso shared in an interview with VentureBeat throughout Subsequent final week.
This renewed push wasn’t hampered by a feared “brain drain” that some outsiders felt was depleting Google. Actually, the corporate quietly doubled down on execution in early 2024 – a 12 months marked by aggressive hiring, inside unification and buyer traction. Whereas opponents made splashy hires, Google retained its core AI management, together with DeepMind CEO Demis Hassabis and Google Cloud CEO Thomas Kurian, offering stability and deep experience.
Furthermore, expertise started flowing in direction of Google’s centered mission. Logan Kilpatrick, as an example, returned to Google from OpenAI, drawn by the chance to construct foundational AI inside the firm, creating it. He joined Velloso in what he described as a “zero to one experience,” tasked with constructing developer traction for Gemini from the bottom up. “It was like the team was me on day one… we actually have no users on this platform, we have no revenue. No one is interested in Gemini at this moment,” Kilpatrick recalled of the place to begin. Folks acquainted with the inner dynamics additionally credit score leaders like Josh Woodward, who helped begin AI Studio and now leads the Gemini App and Labs. Extra not too long ago, Noam Shazeer, a key co-author of the unique “Attention Is All You Need” Transformer paper throughout his first tenure at Google, returned to the corporate in late 2024 as a technical co-lead for the essential Gemini mission
This concerted effort, combining these hires, analysis breakthroughs, refinements to its database know-how and a sharpened enterprise focus total, started yielding outcomes. These cumulative advances, mixed with what CTO Will Grannis termed “hundreds of fine-grain” platform parts, set the stage for the bulletins at Subsequent ’25, and cemented Google’s comeback narrative.
Pillar 1: Gemini 2.5 and the period of considering fashions
It’s true {that a} main enterprise mantra has turn into “it’s not just about the model.” In spite of everything, the efficiency hole between main fashions has narrowed dramatically, and tech insiders acknowledge that true intelligence is coming from know-how packaged across the mannequin, not simply the mannequin itself – for instance, agentic applied sciences that enable a mannequin to make use of instruments and discover the net round it.
Regardless of this, to own the demonstrably best-performing LLM is a crucial feat – and a strong validator, an indication that the model-owning firm has issues like superior analysis and essentially the most environment friendly underlying know-how structure. With the discharge of Gemini 2.5 Professional simply weeks earlier than Subsequent ’25, Google definitively seized that mantle. It rapidly topped the impartial Chatbot Area leaderboard, considerably outperforming even OpenAI’s newest GPT-4o variant, and aced notoriously tough reasoning benchmarks like Humanity’s Final Examination. As Pichai acknowledged within the keynote, “It’s our most intelligent AI model ever. And it is the best model in the world.” The mannequin had pushed an 80 % improve in Gemini utilization inside a month, he Tweeted individually.
For the primary time, Google’s Gemini demand was on fireplace. As I detailed beforehand, apart from Gemini 2.5 Professional’s uncooked intelligence, what impressed me is its demonstrable reasoning. Google has engineered a “thinking” functionality, permitting the mannequin to carry out multi-step reasoning, planning, and even self-reflection earlier than finalizing a response. The structured, coherent chain-of-thought (CoT) – utilizing numbered steps and sub-bullets – avoids the rambling or opaque nature of outputs from different fashions from DeepSeek or OpenAI. For technical groups evaluating outputs for vital duties, this transparency permits validation, correction, and redirection with unprecedented confidence.
However extra importantly for enterprise customers, Gemini 2.5 Professional additionally dramatically closed the hole in coding, which is without doubt one of the largest software areas for generative AI. In an interview with VentureBeat, CTO Fiona Tan, the CTO of main retailer Wayfair, stated that after preliminary assessments, the corporate discovered it “stepped up quite a bit” and was now “pretty comparable” to Anthropic’s Claude 3.7 Sonnet, beforehand the popular alternative for a lot of builders.
Google additionally added an enormous 1 million token context window to the mannequin, enabling reasoning throughout complete codebases or prolonged documentation, far exceeding the capabilities of the fashions of OpenAI or Anthropic. (OpenAI responded this week with fashions that includes equally giant context home windows, although benchmarks counsel Gemini 2.5 Professional retains an edge in total reasoning). This benefit permits for complicated, multi-file software program engineering duties.
Complementing Professional is Gemini 2.5 Flash, introduced at Subsequent ’25 and launched simply yesterday. Additionally, a “thinking” mannequin, Flash is optimized for low latency and cost-efficiency. You may management how a lot the mannequin causes and stability efficiency together with your finances. This tiered method additional displays the “intelligence per dollar” technique championed by Google executives.
Velloso confirmed a chart revealing that throughout the intelligence spectrum, Google fashions provide one of the best worth. “If we had this conversation one year ago… I would have nothing to show,” Velloso admitted, highlighting the fast turnaround. “And now, like, across the board, we are, if you’re looking for whatever model, whatever size, like, if you’re not Google, you’re losing money.” These charts have been up to date to account for OpenAI’s newest mannequin releases this week. See under:
For any given worth, Google’s fashions provide extra intelligence than 90 % of the time. Supply: Pierre Bongrand.
Wayfair’s Tan stated she additionally noticed promising latency enhancements with 2.5 Professional: “Gemini 2.5 came back faster,” making it viable for “more customer-facing sort of capabilities,” she stated, one thing she stated hasn’t been the case earlier than with different fashions. Gemini may turn into the primary mannequin Wayfair makes use of for these buyer interactions, she stated.
The Gemini household’s capabilities prolong to multimodality, integrating seamlessly with Google’s different main fashions like Imagen 3 (picture technology), Veo 2 (video technology), Chirp 3 (audio), and the newly introduced Lyria (text-to-music), all accessible through Google’s platform for Enterprise customers, Vertex. Google is the one firm that provides its personal generative media fashions throughout all modalities on its platform. Microsoft, AWS and OpenAI need to associate with different firms to do that.
Pillar 2: Infrastructure prowess – the engine underneath the hood
The flexibility to quickly iterate and effectively serve these highly effective fashions stems from Google’s arguably unparalleled infrastructure, honed over a long time of operating planet-scale companies. Central to that is the Tensor Processing Unit (TPU).
At Subsequent ’25, Google unveiled Ironwood, its seventh-generation TPU, explicitly designed for the calls for of inference and “thinking models.” The dimensions is immense, tailor-made for demanding AI workloads: Ironwood pods pack over 9,000 liquid-cooled chips, delivering a claimed 42.5 exaflops of compute energy. Google’s VP of ML Techniques Amin Vahdat stated on stage at Subsequent that that is “more than 24 times” the compute energy of the world’s present #1 supercomputer.
Google acknowledged that Ironwood gives 2x perf/watt relative to Trillium, the earlier technology of TPU. That is vital since enterprise prospects more and more say power prices and availability constrain large-scale AI deployments.
Google Cloud CTO Will Grannis emphasised the consistency of this progress. Yr over 12 months, Google is making 10x, 8x, 9x, 10x enhancements in its processors, he instructed VentureBeat in an interview, creating what he referred to as a “hyper Moore’s law” for AI accelerators. He stated prospects are shopping for Google’s roadmap, not simply its know-how.
Google’s place fueled this sustained TPU funding. It must effectively energy large companies like Search, YouTube, and Gmail for greater than 2 billion customers. This necessitated growing customized, optimized {hardware} lengthy earlier than the present generative AI increase. Whereas Meta operates at an identical client scale, different opponents lacked this particular inside driver for decade-long, vertically built-in AI {hardware} improvement.
Now these TPU investments are paying off as a result of they’re driving the effectivity not just for its personal apps, however in addition they enable Google to supply Gemini to different customers at a greater intelligence per greenback, all the things equal.
Why can’t Google’s opponents purchase environment friendly processors from Nvidia, you ask? It’s true that Nvidia’s GPU processors dominate the method pre-training of LLMs. However market demand has pushed up the value of those GPUs, and Nvidia takes a wholesome minimize for itself as revenue. This passes vital prices alongside to customers of its chips. And likewise, whereas pre-training has dominated the utilization of AI chips thus far, that is altering now that enterprises are literally deploying these functions. That is the place ” inference” is available in, and right here TPUs are thought of extra environment friendly than GPUs for workloads at scale.
While you ask Google executives the place their essential know-how benefit in AI comes from, they normally fall again to the TPU as an important. Mark Lohmeyer, the VP who runs Google’s computing infrastructure, was unequivocal: TPUs are “certainly a highly differentiated part of what we do… OpenAI, they don’t have those capabilities.”
Considerably, Google presents TPUs not in isolation, however as a part of the broader, extra complicated enterprise AI structure. For technical insiders, it’s understood that top-tier efficiency hinges on integrating more and more specialised know-how breakthroughs. Many updates have been detailed at Subsequent. Vahdat described this as a “supercomputing system,” integrating {hardware} (TPUs, the newest Nvidia GPUs like Blackwell and upcoming Vera Rubin, superior storage like Hyperdisk Exapools, Wherever Cache, and Speedy Storage) with a unified software program stack. This software program consists of Cluster Director for managing accelerators, Pathways (Gemini’s distributed runtime, now accessible to prospects), and bringing optimizations like vLLM to TPUs, permitting simpler workload migration for these beforehand on Nvidia/PyTorch stacks. This built-in system, Vahdat argued, is why Gemini 2.0 Flash achieves 24 instances increased intelligence per greenback, in comparison with GPT-4o.
Google can also be extending its bodily infrastructure attain. Cloud WAN makes Google’s low-latency 2-million-mile non-public fiber community accessible to enterprises, promising as much as 40% quicker efficiency and 40% decrease complete value of possession (TCO) in comparison with customer-managed networks.
Moreover, Google Distributed Cloud (GDC) permits Gemini and Nvidia {hardware} (through a Dell partnership) to run in sovereign, on-premises, and even air-gapped environments – a functionality Nvidia CEO Jensen Huang lauded as “utterly gigantic” for bringing state-of-the-art AI to regulated industries and nations. At Subsequent, Huang referred to as Google’s infrastructure one of the best on this planet: “No company is better at every single layer of computing than Google and Google Cloud,” he stated.
Pillar 3: The built-in full stack – connecting the dots
Google’s strategic benefit grows when contemplating how these fashions and infrastructure parts are woven right into a cohesive platform. Not like opponents, which frequently depend on partnerships to bridge gaps, Google controls almost each layer, enabling tighter integration and quicker innovation cycles.
So why does this integration matter, if a competitor like Microsoft can merely associate with OpenAI to match infrastructure breadth with LLM mannequin prowess? The Googlers I talked with stated it makes an enormous distinction, they usually got here up with anecdotes to again it up.
Take the numerous enchancment of Google’s enterprise database BigQuery. The database now gives a information graph that permits LLMs to go looking over knowledge way more effectively, and it now boasts greater than 5 instances the purchasers of opponents like Snowflake and Databricks, VentureBeat reported yesterday. Yasmeen Ahmad, Head of Product for Knowledge Analytics at Google Cloud, stated the huge enhancements have been solely potential as a result of Google’s knowledge groups have been working intently with the DeepMind crew. They labored by means of use instances that have been arduous to resolve, and this led to the database offering 50 % extra accuracy based mostly on widespread queries, no less than in response to Google’s inside testing, in attending to the suitable knowledge than the closest opponents, Ahmad instructed VentureBeat in an interview. Ahmad stated this form of deep integration throughout the stack is how Google has “leapfrogged” the {industry}.
This inside cohesion contrasts sharply with the “frenemies” dynamic at Microsoft. Whereas Microsoft companions with OpenAI to distribute its fashions on the Azure cloud, Microsoft can also be constructing its personal fashions. Mat Velloso, the Google govt who now leads the AI developer program, left Microsoft after getting annoyed making an attempt to align Home windows Copilot plans with OpenAI’s mannequin choices. “How do you share your product plans with another company that’s actually competing with you… The whole thing is a contradiction,” he recalled. “Here I sit side by side with the people who are building the models.”
This integration speaks to what Google leaders see as their core benefit: its distinctive potential to attach deep experience throughout the complete spectrum, from foundational analysis and mannequin constructing to “planet-scale” software deployment and infrastructure design.
Vertex AI serves because the central nervous system for Google’s enterprise AI efforts. And the mixing goes past simply Google’s personal choices. Vertex’s Mannequin Backyard gives over 200 curated fashions, together with Google’s, Meta’s Llama 4, and quite a few open-source choices. Vertex gives instruments for tuning, analysis (together with AI-powered Evals, which Grannis highlighted as a key accelerator), deployment, and monitoring. Its grounding capabilities leverage inside AI-ready databases alongside compatibility with exterior vector databases. Add to that Google’s new choices to floor fashions with Google Search, the world’s finest search engine.
Integration extends to Google Workspace. New options introduced at Subsequent ’25, like “Help Me Analyze” in Sheets (sure, Sheets now has an “=AI” system), Audio Overviews in Docs and Workspace Flows, additional embed Gemini’s capabilities into day by day workflows, creating a robust suggestions loop for Google to make use of to enhance the expertise.
Whereas driving its built-in stack, Google additionally champions openness the place it serves the ecosystem. Having pushed Kubernetes adoption, it’s now selling JAX for AI frameworks and now open protocols for agent communication (A2A) alongside help for current requirements (MCP). Google can also be providing lots of of connectors to exterior platforms from inside Agentspace, which is Google’s new unified interface for workers to search out and use brokers. This hub idea is compelling. The keynote demonstration of Agentspace (beginning at 51:40) illustrates this. Google gives customers pre-built brokers, or workers or builders can construct their very own utilizing no-code AI capabilities. Or they’ll pull in brokers from the surface through A2A connectors. It integrates into the Chrome browser for seamless entry.
Pillar 4: Give attention to enterprise worth and the agent ecosystem
Maybe essentially the most vital shift is Google’s sharpened give attention to fixing concrete enterprise issues, notably by means of the lens of AI brokers. Thomas Kurian, Google Cloud CEO, outlined three causes prospects select Google: the AI-optimized platform, the open multi-cloud method permitting connection to current IT, and the enterprise-ready give attention to safety, sovereignty, and compliance.
Brokers are key to this technique. Apart from AgentSpace, this additionally consists of:
Constructing Blocks: The open-source Agent Improvement Package (ADK), introduced at Subsequent, has already seen vital curiosity from builders. The ADK simplifies creating multi-agent techniques, whereas the proposed Agent2Agent (A2A) protocol goals to make sure interoperability, permitting brokers constructed with completely different instruments (Gemini ADK, LangGraph, CrewAI, and many others.) to collaborate. Google’s Grannis stated that A2A anticipates the dimensions and safety challenges of a future with doubtlessly lots of of 1000’s of interacting brokers.
This A2A protocol is actually essential. In a background interview with VentureBeat this week, the CISO of a significant US retailer, who requested anonymity due to the sensitivity round safety points. However they stated the A2A protocol was useful as a result of the retailer is searching for an answer to differentiate between actual individuals and bots who’re utilizing brokers to purchase merchandise. This retailer desires to keep away from promoting to scalper bots, and with A2A, it’s simpler to barter with brokers to confirm their proprietor identities.
Goal-built Brokers: Google showcased professional brokers built-in into Agentspace (like NotebookLM, Thought Era, Deep Analysis) and highlighted 5 key classes gaining traction: Buyer Brokers (powering instruments like Reddit Solutions, Verizon’s help assistant, Wendy’s drive-thru), Artistic Brokers (utilized by WPP, Brandtech, Sphere), Knowledge Brokers (driving insights at Mattel, Spotify, Bayer), Coding Brokers (Gemini Code Help), and Safety Brokers (built-in into the brand new Google Unified Safety platform).
This complete agent technique seems to be resonating. Conversations with executives at three different giant enterprises this previous week, additionally talking anonymously because of aggressive sensitivities, echoed this enthusiasm for Google’s agent technique. Google Cloud COO Francis DeSouza confirmed in an interview: “Every conversation includes AI. Specifically, every conversation includes agents.”
Kevin Laughridge, an govt at Deloitte, an enormous person of Google’s AI merchandise, and a distributor of them to different firms, described the agent market as a “land grab” the place Google’s early strikes with protocols and its built-in platform provide vital benefits. “Whoever is getting out first and getting the most agents that actually deliver value – is who is going to win in this race,” Laughridge stated in an interview. He stated Google’s progress was “astonishing,” noting that customized brokers Deloitte constructed only a 12 months in the past may now be replicated “out of the box” utilizing Agentspace. Deloitte itself is constructing 100 brokers on the platform, focusing on mid-office features like finance, threat and engineering, he stated.
The client proof factors are mounting. At Subsequent, Google cited “500 plus customers in production” with generative AI, up from simply “dozens of prototypes” a 12 months in the past. If Microsoft was perceived as method forward a 12 months in the past, that doesn’t appear so clearly the case anymore. Given the PR warfare from all sides, it’s tough to say who is actually profitable proper now definitively. Metrics range. Google’s 500 quantity isn’t immediately akin to the 400 case research Microsoft promotes (and Microsoft, in response, instructed VentureBeat at press time that it plans to replace this public depend to 600 shortly, underscoring the extraordinary advertising and marketing). And if Google’s distribution of AI by means of its apps is critical, Microsoft’s Copilot distribution by means of its 365 providing is equally spectacular. Each at the moment are hitting hundreds of thousands of builders by means of APIs.
[Editor’s note: Understanding how enterprises are navigating this ‘agent land grab,’ and successfully deploying these complex AI solutions, will be central to the discussions at VentureBeat’s Transform event this June 24-25 in San Francisco.]
However examples abound of Google’s traction:
Wendy’s: Deployed an AI drive-thru system to 1000’s of places in only one 12 months, bettering worker expertise and order accuracy. Google Cloud CTO Will Grannis famous that the AI system is able to understanding slang and filtering out background noise, considerably decreasing the stress of reside buyer interactions. That frees up employees to give attention to meals prep and high quality — a shift Grannis referred to as “a great example of AI streamlining real-world operations.”
Salesforce: Introduced a significant growth, enabling its platform to run on Google Cloud for the primary time (past AWS), citing Google’s potential to assist them “innovate and optimize.”
Honeywell & Intuit: Corporations beforehand strongly related to Microsoft and AWS, respectively, now partnering with Google Cloud on AI initiatives.
Main Banks (Deutsche Financial institution, Wells Fargo): Leveraging brokers and Gemini for analysis, evaluation, and modernizing customer support.
Retailers (Walmart, Mercado Libre, Lowe’s): Utilizing search, brokers, and knowledge platforms.
This enterprise traction fuels Google Cloud’s total development, which has outpaced AWS and Azure for the final three quarters. Google Cloud reached a $44 billion annualized run charge in 2024, up from simply $5 billion in 2018.
Navigating the aggressive waters
Google’s ascent doesn’t imply opponents are standing nonetheless. OpenAI’s fast releases this week of GPT-4.1 (centered on coding and lengthy context) and the o-series (multimodal reasoning, instrument use) show OpenAI’s continued innovation. Furthermore, OpenAI’s new picture technology characteristic replace in GPT-4o fueled large development over simply the final month, serving to ChatGPT attain 800 million customers. Microsoft continues to leverage its huge enterprise footprint and OpenAI partnership, whereas Anthropic stays a robust contender, notably in coding and safety-conscious functions.
Nonetheless, it’s indeniable that Google’s narrative has improved remarkably. Only a 12 months in the past, Google was considered as a stodgy, halting, blundering competitor that maybe was about to blow its probability at main AI in any respect. As a substitute, its distinctive, built-in stack and company steadfastness has revealed one thing else: Google possesses world-class capabilities throughout your entire spectrum – from chip design (TPUs) and international infrastructure to foundational mannequin analysis (DeepMind), software improvement (Workspace, Search, YouTube), and enterprise cloud companies (Vertex AI, BigQuery, Agentspace). “We’re the only hyperscaler that’s in the foundational model conversation,” deSouza acknowledged flatly. This end-to-end possession permits for optimizations (like “intelligence per dollar”) and integration depth that partnership-reliant fashions wrestle to match. Rivals usually must sew collectively disparate items, doubtlessly creating friction or limiting innovation velocity.
Google’s second is now
Whereas the AI race stays dynamic, Google has assembled all these items on the exact second the market calls for them. As Deloitte’s Laughridge put it, Google hit some extent the place its capabilities aligned completely “where the market demanded it.” When you have been ready for Google to show itself in enterprise AI, you might have missed the second — it already has. The corporate that invented lots of the core applied sciences powering this revolution seems to have lastly caught up – and greater than that, it’s now setting the tempo that opponents must match.
Within the video under, recorded proper after Subsequent, AI professional Sam Witteveen and I break down the present panorama and rising developments, and why Google’s AI ecosystem feels so sturdy:
Day by day insights on enterprise use instances with VB Day by day
If you wish to impress your boss, VB Day by day has you coated. We provide the inside scoop on what firms are doing with generative AI, from regulatory shifts to sensible deployments, so you’ll be able to share insights for optimum ROI.
An error occured.