What Google Cloud announced in AI this month – 2025

Editor’s note: Want to keep up with the latest from Google Cloud? Check back here for a monthly recap of our latest updates, announcements, resources, events, learning opportunities, and more.
2025 in review – your month-by-month Google Cloud AI highlight reel
From the early foundation of Gemini 2.0 to the breakthrough reasoning capabilities of Gemini 3 and Nano Banana Pro, this year was defined by velocity, creativity, and bringing AI to where your teams already work. 
Whether you were building your first autonomous agent, generating cinematic video with Veo 3.1, or deploying on our newest Ironwood TPUs, our focus has been on giving you the tools to move faster. Here is your month-by-month guide to the innovations that defined 2025.
Keeping pace with AI innovation is a challenge for every organization. Developers and decision-makers this year faced specific hurdles: moving from prototype to production, managing the costs of inference, and solving the “agent orchestration” puzzle. The updates we released from January to December were answers to these friction points, designed to help you build applications that are faster, smarter, and more cost-effective.
Let’s take a look.
Setting the foundation with models and security
We kicked off the year by expanding the Gemini family and reinforcing the security required for enterprise adoption. In February, we addressed the gap between demos and production performance by launching the Vertex AI RAG Engine. This managed service helped developers reduce hallucinations and build trustworthy applications using their own enterprise data.
By March, the model landscape grew significantly with the introduction of Gemini 2.5 and Gemma 3, our most capable open model to date. To ensure these powerful tools could be deployed safely, we also launched AI Protection, a set of capabilities to safeguard AI workloads across clouds, proving that security is not an afterthought—it’s the foundation.
Agents, I/O, and the developer experience
Spring brought a major shift toward agentic workflows. At Google Cloud Next ’25 in April, we solved a major interoperability challenge with the launch of the Agent2Agent (A2A) protocol. This open standard allowed AI agents to discover and collaborate with one another, complementing Anthropic’s Model Context Protocol (MCP).
Creativity took center stage at Google I/O in May. We unveiled Veo 3 for video, Imagen 4 for images, and Lyria 2 for music, giving creators unprecedented control. We also introduced Jules, an autonomous AI coding agent designed to help developers with complex tasks like bug fixing. By June, we brought Gemini closer to where developers live—the terminal. The launch of the Gemini CLI allowed developers to generate code and debug issues without leaving their command line.
Creativity goes bananas, 1,000 real-world use cases, and payments
The second half of the year focused on proving the value of AI through ROI and enabling agents to take real-world actions. In July, security researchers gained a new ally with Big Sleep, an agent developed with Google DeepMind that successfully identified zero-day vulnerabilities in the wild.
In August, we launched Nano Banana, which quickly went viral overnight because it makes it easy to blend multiple images into a single image, maintain character consistency for rich storytelling, make targeted transformations using natural language, and use Gemini’s world knowledge to generate and edit images.
Finally, to help businesses move from “what if” to “how to,” we published 1,001 real-world gen AI use cases in August, complete with 101 technical blueprints. We then closed the loop on AI commerce in September with the Agent Payments Protocol (AP2). Developed with leading payments companies, AP2 gave agents the ability to securely initiate transactions, turning them from digital assistants into active business participants.
Gemini Enterprise, breakthrough reasoning, and heavy-duty infrastructure
We ended the year with massive leaps in model intelligence and the silicon required to power it. In October, we deepened our partnership with NVIDIA and expanded our Axion processors, enabling Google production services to run simultaneously on x86 and Arm-based machines for better efficiency.
We also announced Gemini Enterprise, the new front door for AI in the workplace. You gain instant access to Google’s cutting-edge Gemini models. Use the latest multimodal breakthrough of Google’s best AI to solve your most complex business problems.
Finally, the year culminated in November with the arrival of Gemini 3, our most intelligent model featuring a 1501 Elo score on the LMArena Leaderboard. To power these advanced workloads, we announced the general availability of Ironwood TPUs and the Antigravity agentic development platform.
2025 set a new pace for innovation, but 2026 will be about what you build with it. If you haven’t yet experienced our most capable model, now is the time to start.
Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud customers, read our monthly recap, Cool stuff customers built.

November
The fastest way to transform your business is here. This November, we took major jumps in both software and silicon. On the model front, we brought Gemini 3 to builders and businesses, securing a breakthrough 1501 Elo score on the LMArena Leaderboard, and redefining what’s possible with agentic workflows. Plus, we built Nano Banana Pro using Gemini 3, which will make it even easier for businesses to bring their creativity to life. 
On the infrastructure side, we announced the general availability of Ironwood TPUs to handle the most demanding workloads imaginable. Whether you’re looking for the world’s best model or the heavy-duty hardware to run it at scale, we’ve got you covered.
Top hits

Bringing Gemini 3 to Enterprise. We brought Gemini 3, our most intelligent model, to every developer and enterprise team. It’s the best model in the world for multimodal understanding, and our most powerful agentic and vibe-coding model yet. Plus, Gemini 3 Pro tops the LMArena Leaderboard with a breakthrough score of 1501 Elo. You can learn more about the model capabilities here.
Announcing Nano Banana Pro for every builder and business. Nano Banana Pro (Gemini 3 Pro Image), our state-of-the art image generation and editing model, is available in Vertex AI and Google Workspace, and coming soon to Gemini Enterprise.

Build with Google Antigravity, our new agentic development platform: Development is lifting off. Google Antigravity is a new agentic development platform designed to help you operate at a higher, task-oriented level.
Announcing Ironwood TPUs General Availability and new Axion VMs to power the age of inference: Ironwood, our seventh generation TPU, will be generally available in the coming weeks. Ironwood is purpose-built for the most demanding workloads: from large-scale model training and complex reinforcement learning (RL) to high-volume, low-latency AI inference and model serving.
More ways to build, scale, and govern AI agents with Vertex AI Agent Builder. We shared new capabilities across the entire agent lifecycle to help you build, scale, and govern AI agents. 
Expanding support for AI developers on Hugging Face. For those building with AI, most are in it to change the world — not twiddle their thumbs. So when inspiration strikes, the last thing anyone wants is to spend hours waiting for the latest AI models to download to their development environment. So, we announced a deeper partnership between Hugging Face and Google Cloud that reduces Hugging Face model download times through Vertex AI and Google Kubernetes Engine, offers native support for TPUs on all open models sourced through Hugging Face, and more.
Google Named a Leader in the Gartner® Magic Quadrant™ for AI Application Development Platforms. We are proud to announce that Google has been recognized as a Leader in the inaugural 2025 Gartner Magic Quadrant for AI Application Development Platforms for our Ability to Execute and Completeness of Vision.

Standout customer success

Mattel’s game changer: How AI is turning customer feedback into real-time product updates. Read about how Mattel is deploying generative AI across their business, including marketing, forecasting, customer service, design, and e-commerce. It’s all aimed at reimagining the way Mattel makes toys, and dreams, come true.

News you can use

5 things to try with Gemini 3 Pro in Gemini CLI: We’ve integrated Gemini 3 Pro directly into Gemini CLI to unlock a new level of performance and productivity in the terminal. This powerful combination delivers state-of-the-art reasoning for executing better commands, enhances support for complex engineering work through agentic coding, and enables smarter, more tailored workflows via advanced tool use.
From silicon to softmax: Inside the Ironwood AI stack: This blog details the core components of Google’s AI software stack that are woven into Ironwood, demonstrating how this deep co-design unlocks performance, efficiency, and scale. We cover the JAX and PyTorch ecosystems, the XLA compiler, and the high-level frameworks that make this power accessible.
Why GKE & Gemini CLI are better together. Dive into how the Gemini CLI and Google Kubernetes Engine (GKE) are coming together with the new open-sourcing of the GKE Gemini CLI extension. This extension brings GKE directly into the Gemini CLI ecosystem, and can also be used as an MCP server with any other MCP client. 

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud customers, read our monthly recap, Cool stuff customers built.

October
From announcing the new front door for enterprise AI and supercharged infrastructure, to powerful new tools for creators, this month’s updates are focused on making generative AI more powerful and accessible.
We kicked off the month by announcing Gemini Enterprise, our new agentic platform designed to bring the best of Google AI to every workflow. We also expanded our AI infrastructure, deepening our NVIDIA partnership with new A4X Max instances, and advanced our multi-arch capabilities with Axion processors.
At the same time, we continue to see incredible creative momentum. Since launching our popular Nano Banana model, consumers have created 13 billion images and 230 million videos. Now, enterprises can harness that same power by combining Gemini 2.5 Pro with our generative media models—Lyria, Chirp, Imagen, and Veo.
Here’s a rundown of top announcements, and helpful how-to guides to help you get started.
Top hits 

Introducing Gemini Enterprise. To help all of our customers succeed with AI, we’re introducing Gemini Enterprise. It’s our advanced agentic platform that brings the best of Google AI to every employee, for every workflow. Sign up for a 30-day free trial, or contact sales.  
Expanding our NVIDIA partnership: Now shipping A4X Max, Vertex AI Training, and more. We deepened our partnership with NVIDIA with a suite of new capabilities that strengthens our platform for the entire AI lifecycle:

New A4X Max instances powered by NVIDIA’s GB300 NVL72, purpose-built for multimodal AI reasoning tasks

Google Kubernetes Engine (GKE), now supporting Dynamic Resource Allocation Kubernetes Network Driver (DRANET), boosting bandwidth in distributed AI/ML workloads

GKE Inference Gateway, now integrating with NVIDIA NeMo Guardrails 

Vertex AI Model Garden to feature NVIDIA Nemotron models

Vertex AI Training recipes on top of the NVIDIA NeMo Framework and NeMo-RL

Announcing new capabilities in Vertex AI Training for large-scale training: Our new managed training features, aimed at developers training with hundreds to thousands of AI accelerators, builds on the best of Google Cloud’s AI infrastructure offerings, including Cluster Director for a fully managed and resilient Slurm environment, and adds sophisticated management tools. This includes pre-built data science tooling and optimized recipes integrated with frameworks like NVIDIA NeMo for specialized, massive-scale model building.
Google Skills: Your new home for Google AI learning and more. A new platform, Google Skills, that will bring together nearly 3,000 courses and labs in one place — including content from across Google Cloud, Google DeepMind, Grow with Google and Google for Education. Sign up here

Standout thought leadership

At Google, the future is multiarch; AI and automation are helping us get there: Parthasarathy Ranganathan, VP, Engineering Fellow and Wolff Dobson, Developer Relations Engineer, put Axion processors to the test: running Google production services. Now that our clusters contain both x86 and Axion Arm-based machines, Google’s production services are able to run tasks simultaneously on multiple instruction-set architectures (ISAs). Today, this means most binaries that compile for x86 now need to compile to both x86 and Arm at the same time — no small thing when you consider that the Google environment includes over 100,000 applications!

News you can use

The ultimate prompting guide for Veo 3.1: This guide is a framework for directing Veo 3.1, our latest model that marks a shift from simple generation to creative control. Veo 3.1 builds on Veo 3, with stronger prompt adherence and improved audiovisual quality when turning images into videos. 
5 ad agencies used Gemini 2.5 Pro and gen media models to create an “impossible ad”: To us, generative media is a canvas to explore ideas that were previously constrained by time, budget, or the limits of conventional production. To test this, we briefed several top agencies to use Google’s AI to create an “impossible” ad — a campaign that pushes the boundaries of what’s creatively and technically feasible.
Building scalable AI agents: Design patterns with Agent Engine on Google Cloud: This post details how to build, scale, and manage enterprise-grade agentic systems using Google Cloud AI products to enable SI Partners to offer these transformative solutions to enterprise clients.
Use Gemini CLI to deploy cost-effective LLM workloads on GKE: With Inference Quickstart, you can replace months of manual trial-and-error with out-of-the-box manifests and data-driven insights. Inference Quickstart integrates with the Gemini CLI through native Model Context Protocol (MCP) support to offer tailored recommendations for your LLM workload cost and performance needs. 

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud customers, read our monthly recap, Cool stuff customers built.

September
This month, we’re buzzing with the news of AP2, the Agent Payments Protocol, an open protocol  developed with 60+ leading payments and technology companies. AP2 is a blueprint for interoperable, AI-native commerce, designed to work seamlessly as an extension of the Agent2Agent (A2A) protocol and Model Context Protocol (MCP). 
We’ve also seen “bananas” momentum with our generative media models, especially through nano-banana, or Gemini 2.5 Flash Image. Let’s dive in! 
Top hits

Building next-gen visuals with Gemini 2.5 Flash Image (aka nano-banana) on Vertex AI: We announced native image generation and editing in Gemini 2.5 Flash (aka nano-banana) to deliver higher-quality images and more powerful creative control. Gemini 2.5 Flash Image is State-of-the-Art (SOTA) for both generation and image editing. For creative use cases, this means you can create richer, more dynamic visuals and edit images until they’re just right.

Adobe Firefly uses Google's Gemini 2.5 Flash Image

Powering AI commerce with the new Agent Payments Protocol (AP2): We announced the Agent Payments Protocol (AP2), an open protocol developed with leading payments and technology companies to securely initiate and transact agent-led payments across platforms. The protocol can be used as an extension of the Agent2Agent (A2A) protocol and Model Context Protocol (MCP). In concert with industry rules and standards, it establishes a payment-agnostic framework for users, merchants, and payments providers to transact with confidence across all types of payment methods.

Intro to Google Agent Payments Protocol (AP2)

The ROI of AI: How agents are delivering for business: The AI story is evolving. According to our 2025 ROI of AI Report, for the 52% of executives who report their organizations are now deploying AI agents in production, this represents a fundamental shift in how business gets done.
AI agent security: How to protect digital sidekicks (and your business): The promise of AI agents is immense. Powerful software that can plan, make decisions, and accurately act on your behalf at work? Yes please. But what if your digital assistant goes rogue? Without robust, flexible security in place, an AI agent could leak sensitive company data, cause a system outage, and even expose your business to security threats.
Scaling high-performance inference cost-effectively: GKE Inference Gateway is generally available, and we are launching new capabilities that deliver even more value. This underscores our commitment to helping companies deliver more intelligence, with increased performance and optimized costs for both training and serving. To learn more and get started, visit our AI Hypercomputer site.

News you can use 

Agent Factory Recap: Keith Ballinger on AI, The Future of Development, and Vibe Coding: In Episode #6 of the Agent Factory podcast, Vlad Kolesnikov joined Keith Ballinger, VP and General Manager at Google Cloud, for a deep dive into the transformative future of software development with AI. We explore how AI agents are reshaping the developer’s role and boosting team productivity.

Announcing the new Practical Guide to Data Science on Google Cloud: We’ve designed our new guide for practitioners looking to use Google Cloud’s capabilities across BigQuery, Vertex AI, and Google Cloud Serverless for Apache Spark. We’ll show you how to use an AI-first approach in your data science workflows, use previously untapped unstructured and multimodal data, and achieve new levels of efficiency.
Automate app deployment and security analysis with new Gemini CLI extensions: We’re closing the gap between your terminal and the cloud with a first look at the future of Gemini CLI, delivered through two new extensions: security extension and Cloud Run extension:

/security:analyze performs a comprehensive scan right in your local repository, with support for GitHub pull requests coming soon. This makes security a natural part of your development cycle.
/deploy deploys your application to Cloud Run, our fully managed serverless platform, in just a few minutes.

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.

August
This August, we’re closing the gap between inspiration and implementation. We published 101 real-world gen AI use cases with technical blueprints, the technical counterpart to our popular 601 real-world gen AI use cases from the world’s leading organizations. Each blueprint provides the Google Cloud tech stack you might need to bring these ideas to life. So take a look, play around, and share what you build. 
If you’re not sure which Google AI developer tool to use to get started, we provided a helpful guide to show you which tool to use, and when. 
Let’s dive in! 
Top hits

Building next-gen visuals with Gemini 2.5 Flash Image on Vertex AI: We announced native image generation and editing in Gemini 2.5 Flash to deliver higher-quality images and more powerful creative control. Gemini 2.5 Flash Image is State of the Art (SOTA) for both generation and image editing. 

How to get started: Developers and enterprises can access Gemini 2.5 Flash Image in preview today on Vertex AI, with built-in SynthID watermarking for responsible and transparent use.

Adobe Firefly uses Google's Gemini 2.5 Flash Image

How much energy does Google AI use? We did the math: Did you know the estimated per-prompt energy impact is equivalent to watching TV for less than nine seconds? Take a look at the technical paper to learn more.

Calculating our AI energy consumption

How Wells Fargo is using Google Cloud AI to empower its workforce with agentic tools: Wells Fargo, an early adopter of Google Agentspace, is transforming how individuals and teams work, collaborate, and serve customers. 

At the Google Cloud Security Summit 2025, we shared details around new capabilities designed to help you secure your AI initiatives, and to help you use AI to make your organization more secure. 

News you can use

Here’s which Google AI developer tool to use for each situation: From Jules to Firebase Studio, we mapped out our developer tool landscape to help you choose the right product for your project. This is a must read for any developer who wants to get started with Google AI developer tools.

Run OpenAI’s new gpt-oss model at scale with Google Kubernetes Engine: This guide walks you through deploying the new gpt-oss model from OpenAI on GKE for scalable and efficient performance.

Build a real-time voice agent with Gemini and Google ADK: Learn how to use the Agent Development Kit (ADK) and Gemini to create sophisticated, real-time voice agents for your applications.

The latest in AI Hypercomputer: Catch up on the latest updates and advancements in our AI Hypercomputer architecture, designed to handle the most demanding AI workloads.

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.

July
Do you know since its preview launch on Vertex AI in June, enterprise customers have already generated over 6 million videos? 
This July, we focused on helping you build what’s next, and do it with confidence. We announced Veo 3 and Veo 3 Fast on Vertex AI to bring your ideas to life. This is exciting, especially for marketers, who want to rapidly iterate on high-quality creative. Plus, we’ve pulled together over 25+ of our best guides – from model deployment to building gen AI apps – to help you find ways to get started.
Top hits

Veo 3 and Veo 3 Fast are now available for everyone on Vertex AI: Veo 3 Fast is a faster way to turn text to video, from narrated product demos to short films. 

How to get started: Go here to learn more about Veo 3 and Veo 3 Fast on Vertex AI, and try it on Vertex AI Media Studio.

Stained glass workshop

Our Big Sleep agent makes a big leap: Developed by Google DeepMind and Google Project Zero, Big Sleep can help security researchers find zero-day (previously unknown) software security vulnerabilities. We believe this is the first time an AI agent has been used to directly foil efforts to exploit a vulnerability in the wild.

Announcing a complete developer toolkit for scaling A2A agents on Google Cloud: We announced version 0.3 of the A2A protocol, which brings a more stable interface to build against and is critical to accelerating enterprise adoption. This version introduces several key capabilities, including gRPC support, the ability to sign security cards, and extended client side support in the Python SDK, which provide more flexible use, better security and easier integration.

The global endpoint offers improved availability for Anthropic’s Claude on Vertex AI: Anthropic’s Claude models on Vertex AI now have improved overall availability with the global endpoint for Claude models. Now generally available, the global endpoint unlocks the ability to dynamically route your requests to any region with available capacity supported by the Claude model you’re using.

Our customers are building cool things: For our latest edition, we explore Box’s AI agents extracting insights with cross-platform agent integration; Schroders uses multiple connected agents to build a complex investment research system; Hypros’ IoT device can monitor patient distress in hospitals without individual monitoring; a Formula E exhibition of whether regenerative braking can power an EV supercar for an entire lap.

News you can use

Bookmark our 25+ generative AI how-to guides: We gathered over 25 how-to guides for enterprise generative AI, covering everything from faster model deployment and building multi-agent systems to fine-tuning, evaluation, and RAG.

Build a conversational analytics agent with BigQuery: We released a new, first-party toolset for BigQuery that works with Google’s Agent Development Kit (ADK) and the open-source MCP Toolbox for Databases. Learn how to build AI agents that can securely and intelligently interact with your enterprise data.

Take an open model from discovery to endpoint on Vertex AI: This guide walks you through the process of selecting, fine-tuning, evaluating, and deploying an open model on Vertex AI. It covers everything from the Model Garden to a production-ready endpoint.

Enable Secure Boot for your AI workloads: Learn how to enable Secure Boot for your GPU-accelerated AI workloads on Google Cloud. Our Secure Boot capability can be opted into at no additional charge, and now there’s a new, easier way to set it up for your GPU-accelerated machines.

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.

June
From the command line to cinematic screens, June was about making AI practical and powerful. Developers can now access Gemini right from their terminal with our new command-line interface, Gemini CLI, and creators can produce stunning, narrative-driven video that combines audio and visuals with Veo 3. Plus, we published several helpful guides and resources for you to bookmark and use as you start building. 
Top hits: Gemini comes to everyone – from your command line to Vertex AI 

In your terminal: Introducing Gemini CLI, an open-source AI agent that brings the power of Gemini directly into your terminal. It provides lightweight access to Gemini, giving you the most direct path from your prompt to our model. Check it out on GitHub.

How to use it: To use Gemini CLI free-of-charge, simply login with a personal Google account to get a free Gemini Code Assist license. That free license gets you access to Gemini 2.5 Pro and its massive 1 million token context window. If you’re a professional developer who needs to run multiple agents simultaneously, or if you prefer to use specific models, you can use a Google AI Studio or Vertex AI key for usage-based billing or get a Gemini Code Assist Standard or Enterprise license.

In Vertex AI: Gemini 2.5 Flash and 2.5 Pro now stable and generally available. Our most intelligent models for speed and advanced reasoning are production-ready providing organizations with the stability, reliability and scalability needed to confidently deploy the most advanced AI capabilities into mission-critical applications.  Learn more. 

The story unfolds: You dream it, Veo creates it
Any creative class will tell you the same axiom of truth: A great story shows rather than tells. We gave our text-to-video model, Veo 3, a simple, evocative prompt to show how it can turn a creative spark into a cinematic scene. Veo 3 is now available for all Google Cloud customers and partners in public preview on Vertex AI.

Why this matters: Veo 3 not only brings stunning visual quality, but now adds sound from background sounds to dialogue. Learn more. 

The prompt: “A medium shot frames an old sailor, his knitted blue sailor hat casting a shadow over his eyes, a thick grey beard obscuring his chin. He holds his pipe in one hand, gesturing with it towards the churning, grey sea beyond the ship’s railing. ‘This ocean, it’s a force, a wild, untamed might. And she commands your awe, with every breaking light'”
The result:

Introducing Veo 3: Old Sailor

AI agents – they’re taking autonomous action, but we need to keep them secure 
As AI moves from answering questions to taking action, securing these autonomous systems is paramount. This month, we’re not only highlighting how to build powerful agents but also how to implement a security-first approach to their deployment. 
Exec perspective: Earlier this month, Anton Chuvakin, security advisor for Google Cloud’s Office of the CISO, discussed a new Google report on securing AI agents, and the new security paradigm they demand. Here’s a snippet: With great power comes a great responsibility for agent developers. To help mitigate the potential risks posed by rogue agent actions, we should invest in a new field of study focused specifically on securing agent systems.
Keeping the conversation going: What are the unique security threats posed by agentic AI, and what can leaders do to start securing their workloads? To help answer that question, we sat down with Archana Ramamoorthy, Senior Director, Google Cloud Security, to ask her how businesses should protect their AI workloads by making security more collaborative. Here’s a snippet: 
“Just as leaders are getting a handle on the security risks of generative AI, the ground is shifting again. The emergence of AI agents — which can execute tasks on your company’s behalf — is demanding a new security paradigm: agent security. This new layer of autonomy can magnify blind spots in an organization’s existing security posture…The bottom line is that we need to be prepared, and the whole organization should invest in keeping systems secure.”
News you can use
Build smarter, multimodal AI applications

Build a RAG-capable app using Vertex AI services. This is a new architecture guide that helps understand the role of Vertex AI and Vector Search in a generative AI app. It includes diagrams, design considerations, and much more.
Check out this digestible tutorial to create multimodal agents for complex tasks like object detection by using a new tutorial featuring Gemini, LangChain, and LangGraph. Matthew and May show you which decisions you need to make to combine Gemini, LangChain and LangGraph to build multimodal agents that can identify objects. 
Bookmark this quick guide to fine-tune video inputs on Vertex AI. If your work involves content moderation, video captioning, and detailed event localization, this guide is for you. 

Deploy your AI efficiently and at scale

How good is your AI? In this explainer, we dive into the new features of the Gen AI Evaluation Service, designed to help you scale your evaluations, evaluate your autorater, customize your autorater with rubrics and evaluate your agents in production. Check out this helpful rubric:

New recipe! Learn how to access Llama4 and DeepSeek models today on AI Hypercomputer. A quick but essential read. 

Ground your AI in your enterprise data

Read this guide on MCP integrations with Google Cloud Databases. Now with Toolbox, any MCP-compatible AI assistant (including Claude Code, Cursor, Windsurf, Cline, and many more) can help you write application code that queries your database, designs a schema for a new application, refactors code when the data model changes, generates data for integration testing, explores the data in your database, etc.

Stay tuned
Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.

May
This month has been one of our biggest yet – from over 100 announcements at I/O, to announcing critical new partnerships such as Anthropic’s Claude Opus 4 and Claude Sonnet 4 on Vertex AI, to launching a first-of-its kind generative AI certification for non-technical learners. 
Today, we’re taking you through the biggest news this month, helpful guides to get started, and some inspiration for your next project with the May edition of our new series, Cool stuff Google Cloud customers built.

Created with Veo: Lost Island

Top announcements
Google I/O brought a fresh wave of tools from Google Cloud, all designed to help businesses and developers build what’s next. These updates bring new ways for organizations to work with AI, code more easily, create media, and manage intelligent agents. Here are the highlights: 

We introduced new generative AI models for media, including Veo 3 for video, Imagen 4 for images, and Lyria 2 for music on Vertex AI. These models give you excellent ways to create visual and audio content from text prompts. Learn more in our blog here.

We’ve expanded Gemini 2.5 Flash and Pro model capabilities to help enterprises build more sophisticated and secure AI-driven applications and agents. With thought summaries, businesses get clarity and auditability of a model’s raw thoughts — including key details and tool usage. The new Deep Think mode uses research techniques that enable the model to consider multiple hypotheses before responding. 

Gemini 2.5 is now powering all Gemini Code Assist editions! We also launched Jules, a new autonomous AI coding agent, now in public beta, designed to understand user intent and perform coding tasks like writing tests and fixing bugs

Firebase Studio is a cloud-based, AI workspace powered by Gemini 2.5 that lets you turn your ideas into a full-stack app in minutes. Now you can import Figma designs directly into Firebase Studio using the builder.io plugin, and then add features and functionality using Gemini in Firebase without having to write any code. 

We’re making AI application deployment significantly easier with Cloud Run, launching three key updates: first, you can now deploy applications built in Google AI Studio directly to Cloud Run with just a single button click; second, we enabled direct deployment of Gemma 3 models from AI Studio to Cloud Run, complete with GPU support for scalable, pay-per-use endpoints; and third, we’ve introduced a new Cloud Run MCP server, which empowers MCP-compatible AI agents (like AI assistants, IDE integrations, or SDKs) to programmatically deploy applications. Read more here. 

The news didn’t stop with I/O – we announced several important announcements to help you deploy AI at scale:

Introducing the next generation of AI inference, powered by llm-d: We’re making inference even easier and more cost-effective, by making vLLM fully scalable with Kubernetes-native distributed and disaggregated inference. This new project is called llm-d. Google Cloud is a founding contributor alongside Red Hat, IBM Research, NVIDIA, and CoreWeave, joined by other industry leaders AMD, Cisco, Hugging Face, Intel, Lambda, and Mistral AI.

Mistral AI’s Le Chat Enterprise and Mistral OCR 25.05 model are available on Google Cloud. Available today on Google Cloud Marketplace, Mistral AI’s Le Chat Enterprise is a generative AI work assistant designed to connect tools and data in a unified platform for enhanced productivity. 

Anthropic’s Claude Opus 4 and Claude Sonnet 4 on Vertex AI. Claude Opus 4 and Claude Sonnet 4 are generally available as a Model-as-a-Service (MaaS) offering on Vertex AI. For more information on the newest Claude models, visit Anthropic’s blog. 

…and made strides in security: 

What’s new with Google Cloud’s Risk Protection Program. We unveiled at Google Cloud Next major updates to our Risk Protection Program, an industry-first collaboration between Google and insurers that provides competitively priced cyber-insurance and broad coverage for Google Cloud customers. We’re now including Affirmative AI insurance coverage for your Google-related AI workloads. Here’s what’s new.
How Confidential Computing lays the foundation for trusted AI. Our latest Confidential Computing innovations highlight the creative ways our customers are using Confidential Computing to protect their most sensitive workloads — including AI.
How governments can use AI to improve threat detection and reduce cost. In the latest Cloud CISO Perspectives newsletter, our Office of the CISO’s Enrique Alvarez, public sector advisor, explains how government agencies can use AI to improve threat detection — and save money. 

News you can use: Actionable ways to get started
Get fluent in generative AI: 
62% of employers now expect candidates and employees to possess at least some familiarity with AI. That’s why we launched a first-of-its kind generative AI certification for non-technical learners—plus a new suite of no-cost training to help you prepare for that certification. That means you — and your company — can be among the first to take advantage of this opportunity to validate your strategic acumen in gen AI. Become a generative AI leader today. 
Then, put generative AI to work 
At I/O, we expanded generative AI media on Vertex AI. But how do you get started, today? To help you make the most of all the latest generative AI media announcements, we redesigned Vertex AI Studio. The developer-first experience will be your source for generative AI media models across all modalities. You’ll have access to Google’s powerful generative AI media models such as Veo, Imagen, Chirp and Lyria in the Vertex AI Media Studio.

Redesigned Vertex AI Studio

To help you turn your generative AI ideas into real web applications, we published this guide to create gen AI apps in less than 30 seconds with Vertex AI and Cloud Run. Any developer knows it’s a complex process to build shareable, interactive applications: you have to set up infrastructure, wire APIs, and build a front-end. It’s usually a complex process. What if you could skip the heavy lifting and turn your generative AI concept into a working web app with just a few clicks?
New how-to series alert: Text-to-SQL agents 
Recently, powerful large language models (LLMs) like Gemini, with their abilities to reason and synthesize, have driven remarkable advancements in the field of text-to-SQL. In this blog post, the first entry in a series, we explore the technical internals of Google Cloud’s text-to-SQL agents.
Real-life demo: What if we turned Gemini into an AI basketball coach? 
We rounded out this month with a deep-dive into a demo we showcased at Google Cloud Next and most recently, at I/O.  In this article, we showed an AI experiment that turns Gemini 2.5 Pro into a jump shot coach. By combining a ring of Pixel cameras with Vertex AI, the coaching system connects AI motion capture, biomechanical analytics, and Gemini-powered coaching via text and voice.
“It’s like we always say: AI is only as good as the information you give it. For the AI basketball coach to be accurate, we knew we had to talk to actual, real-life professionals. So we talked to our partners at the Golden State Warriors and came up with essential criteria for helping you shoot like the pros.”
Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.

April
From Agent2Agent Protocol to Gemini 2.5 Pro, it’s been an exciting month. We hosted Google Cloud Next in Las Vegas on April 9th with over 30,000 people, announcing incredible innovations from Ironwood TPU to Agent2Agent Protocol. We also expanded Vertex AI in dizzying ways – now, it’s the only platform with generative media models across video, image, speech, and music and makes it possible for multiple agents to coexist in your enterprise.
If you missed the livestream, take a look at our Day 1 recap and summary of the developer keynote. It’s been incredible to see how customers have been applying AI to hundreds of use cases. In fact, we’ve counted more than 600 examples. 
Top announcements
Agents
Our recently launched Agent2Agent (A2A) protocol is getting a lot of attention, and for good reason. This open interoperability protocol is designed to make it easy for AI agents to communicate with one another, no matter its foundation. It’s also a powerful complement to Anthropic’s Model Context Protocol (MCP). Together, these two open protocols are the foundation for complex multi-agent systems.
Since the launch, we’ve been excited to see all the discussions and reactions across social media and YouTube. The A2A GitHub repository (13k stars) has grown quickly, which underscores the industry’s eagerness for a standardized agent communication protocol, especially these features:

Easy agent discovery: A key feature of A2A is the ability for agents to discover the capabilities of other agents through standardized “Agent Cards”. 

Complementary to MCP: As we say on our A2A ❤️ MCP topic page, A2A complements MCP.  MCP. MCP equips an individual agent with the right context and tools to make it more capable A2A enables multiple, diverse agents to communicate and work with each other

Open and community driven: A2A is open source, inviting contributions from the broader community to refine and expand its functionality. To learn more, check out our GitHub here. 

And, speaking of agents, we made a host of updates to Google Agentspace – starting with unified search. Now, Agentspace is integrated with Chrome Enterprise, which means you can search and access all of your enterprise’s resources — right from the search box in Chrome. You can also discover and adopt agents quickly and easily with Agent Gallery, and create agents with our new no-code Agent Designer.
Models
This month, we announced six new models (some in preview, some generally available) for our customers:

Gemini 2.5 Pro is available in preview on Vertex AI and the Gemini app.    

Gemini 2.5 Flash — our workhorse model optimized specifically for low latency and cost efficiency — is coming soon to Vertex AI, AI Studio, and the Gemini app. 

Imagen 3, our highest quality text-to-image model, now has improved image generation and inpainting capabilities for reconstructing missing or damaged portions of an image. 

Chirp 3, our groundbreaking audio generation model, now includes a new way to create custom voices with just 10 seconds of audio input.

Lyria, the industry’s first enterprise-ready, text-to-music model, can transform simple text prompts into 30-second music clips.

Veo 2, our industry-leading video generation model, is expanding with new features that help organizations create videos, edit them, and add visual effects.

Security
Hot on the heels of our AI Protection news, we introduced Google Unified Security at Next ‘25, which lays the foundation for superior security outcomes, followed by the RSA Conference in San Francisco. We showcased there:

Three new ways you can use AI as your security sidekick — complete with prompts to get you started.

Two new security agents for malware analysis and alert triage (plus two agents for customers who opt into the new SecOps Labs.)

Our singular vision for how agents can revolutionize security operations centers.

New open-source MCP servers for Google Security Operations, Google Threat Intelligence, and Security Command Center. 

News you can use:
In just the last year alone, we’ve seen over 40x growth in Gemini use on Vertex AI, now with billions of API calls per month. So what’s new and better with Vertex AI?
To start:
Meta’s Llama 4 is generally available on Vertex AI. 

Vertex AI Dashboards: These help you monitor usage, throughput, latency, and troubleshoot errors, providing you with greater visibility and control.

Vertex AI Model Optimizer: This capability uses Google’s unique understanding of Gemini to automatically direct your query to the most performant model and tools, based on your quality, speed and cost preferences. 

Live API: To enable truly conversational interactions, Live API offers streaming audio and video directly into Gemini.

Worth a second read: 

Agent Development Kit (ADK) is a new open-source framework for designing agents built on the same framework that powers Google Agentspace and Google Customer Engagement Suite (CES) agents. Many powerful examples and extensible sample agents are readily available in Agent Garden. 

Agent Engine is a fully managed runtime in Vertex AI that helps you deploy your custom agents to production with built-in testing, release, and reliability at a global, secure scale.

Next month, we’ll be releasing several guides and deep-dives into how to use all these features, so stay tuned!
Hear from our leaders: 
As usual, we ended this month with our monthly column, The Prompt. In this installment, we hear from Logan Kilpatrick, AI product leader on Google DeepMind, on how multimodal capabilities – especially audio and vision – are bringing about a new UX paradigm. Here’s a snippet:
“There’s the age-old quote that a picture is worth a thousand words. This matters double in the world of multimodal. If I look at my computer right now and try to describe everything I see, it would take 45 minutes. Or, I could just take a picture. Use cases for vision could be something from as simple as object tracking to image detection. For example, a factory watching an assembly line to make sure there’s no impurities in the product they’re creating. Or analyzing dozens of pictures of your farm and trying to understand crop yields. There’s huge breadth and opportunity by blending these modalities together.”
Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.

March
In March, we made big strides in our models and accessibility – from announcing Gemini 2.5, our most intelligent AI model, to Gemma 3, our most capable and advanced version of the Gemma open-model family. We also introduced Gemini Code Assist for individuals, a free version of our AI coding assistant. 
In the industry, we shone light on telecom and gaming. We went to Mobile World Congress, where we displayed key agent use cases where AI is becoming an imperative in telecom. At the same time, we talked to leaders in the games industry, who told us how they’re using Google Cloud AI to drive unprecedented advancements in game development, including smarter, faster, and more immersive gaming experiences. 
Top announcements 
The Gemini family is growing – we introduced Gemini 2.5, a thinking model designed to tackle increasingly complex problems. Gemini 2.5 Pro Experimental leads common benchmarks by meaningful margins and showcases strong reasoning and code capabilities. Learn all about it here.
We also introduced Gemma 3, the most capable model you can run on a TPU or GPU. To help you get started, we shared guides on how to deploy your AI workloads on Cloud Run and Vertex AI.
In the world of open-source, we announced Claude 3.7 Sonnet, Anthropic’s most intelligent model to date and the first hybrid reasoning model on the market, is available in preview on Vertex AI Model Garden. Claude 3.7 Sonnet can produce quick responses or extended, step-by-step thinking that is made visible to the user. Explore our sample notebook and documentation to start building.
Finally, we took a step forward in security. We introduced AI Protection, a set of capabilities designed to safeguard AI workloads and data across clouds and models — irrespective of the platforms you choose to use. Our Office of the CISO suggests these five do-it-today, actionable tips on how leaders can help their organizations adopt AI securely.
News you can use
Wondering how to make the most of your AI? When it comes to Gemini 2.0, we created a helpful guide that teaches you how to optimize one of enterprises’ most time-intensive tasks: document extraction. 
From an infrastructure perspective, we broke down the four top use cases for AI Hypercomputer and ways to get started. In this guide, you’ll learn everything from affordable inference to reducing delivery bottlenecks. Take Moloco, for example. Using the AI Hypercomputer architecture they achieved 10x faster model training times and reduced costs by 2-4x.
And, speaking of cost – do you know the true cost of enterprise AI? Enterprises need ways to optimize large AI workloads because these resources can still be quite expensive. Learn how to calculate your AI cost and dig into these five tips to optimize your workflow on Google Cloud Platform.
Hear from our leaders 
We closed off this month with our monthly column, The Prompt. In this installment, we hear from Suraj Poozhiyil, AI product leader, on how AI agents depend on “enterprise truth” – your enterprise’s unique context – to be successful. Here’s a snippet:
We’ve heard it before – AI is only as good as the data you put into it. When it comes to agents, AI agents are only as good as the context you give them. Enterprise truth is the answer to questions like, “What is our company’s policy for creating a purchase order?” and “What is the specific workflow for approvals and compliance?”
Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.

February
2025 is off to a racing start. From announcing strides in the new Gemini 2.0 model family to retailers accelerating with Cloud AI, we spent January investing in our partner ecosystem, open-source, and ways to make AI more useful. We’ve heard from people everywhere, from developers to CMOs, about the pressure to adapt the latest in AI with efficiency and speed – and the delicate balance of being both conservative and forward-thinking. We’re here to help. Each month, we’ll post a retrospective that recaps Google Cloud’s latest announcements in AI – and importantly, how to make the most of these innovations. 
Top announcements: Bringing AI to you 
This month, we announced agent evaluation in Vertex AI. A surprise to nobody, AI agents are top of mind for many industries looking to deploy their AI and boost productivity. But closing the gap between impressive model demos and real-world performance is crucial for successfully deploying generative AI. That’s why we announced Vertex AI’s RAG Engine, a fully managed service that helps you build and deploy RAG implementations with your data and methods. Together, these new innovations can help you build reliable, trustworthy models.
From an infrastructure perspective, we announced new updates to AI Hypercomputer. We wanted to make it easier for you to run large multi-node workloads on GPUs by launching A3 Ultra VMs and Hypercompute Cluster, our new highly scalable clustering system. This builds on multiple advancements in AI infrastructure, including Trillium, our sixth-generation TPU.
What’s new in partners and open-source 
This month, we invested in our relationship with our partners. We shared how Gemini-powered content creation in Partner Marketing Studio will help partners co-market faster. These features are designed to streamline marketing efforts across our entire ecosystem, empowering our partners to unlock new levels of success, efficiency, and impact. 
At the same time, we shared several important announcements in the world of open-source. We announced Mistral AI’s Mistral Large 24.11 and Codestral 25.01 models on Vertex AI. These models will help developers write code and build faster – from high-complexity tasks to reasoning tasks, like creative writing. To help you get started, we provided sample code and documentation.
And, most recently, we announced the public beta of Gen AI Toolbox for Databases in partnership with LangChain, the leading orchestration framework for developers building LLM applications. Toolbox is an open-source server that empowers application developers to connect production-grade, agent-based generative AI applications to databases. You can get started here.
Industry news: Google Cloud at the National Retail Federation (NRF) 
The National Retail Federation kicked off the year with their annual NRF conference, where Google Cloud showed how AI agents and AI-powered search are already helping retailers operate more efficiently, create personalized shopping experiences, and use AI to get the latest products and experiences to their customers. Check our new AI tools to help retailers build gen AI search and agents. 
As an example, Google Cloud worked with NVIDIA to empower retailers to boost their customer engagements in exciting new ways, deliver more hyper-personalized recommendations, and build their own AI applications and agents. Now with NVIDIA’s AI Enterprise software available on Google Cloud, retailers can handle more data and more complex AI tasks without their systems getting bogged down.
News you can use 
This month, we shared several ways to better implement fast-moving AI, from a comprehensive guide on Supervised Fine Tuning (SFT), to how developers can help their LLMs deliver more accurate, relevant, and contextually aware responses, minimizing hallucinations and building trust in AI applications by optimizing their RAG retrieval.
We also published new documentation to use open models in Vertex AI Studio. Model selection isn’t limited to Google’s Gemini anymore. Now, choose models from Anthropic, Meta, and more when writing or comparing prompts.
Hear from our leaders
We closed out the month with The Prompt, our monthly column that brings observations from the field of AI. This month, we heard from Warren Barkley, AI product leader, who shares some best practices and essential guidance to help organizations successfully move AI pilots to production. Here’s a snippet:
More than 60% of enterprises are now actively using gen AI in production, helping to boost productivity and business growth, bolster security, and improve user experiences. In the last year alone, we witnessed a staggering 36x increase in Gemini API usage and a nearly 5x increase of Imagen API usage on Vertex AI — clear evidence that our customers are making the move towards bringing gen AI to their real-world applications.
Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.
Quelle: Google Cloud Platform

What’s new with Google Cloud – 2025

Want to know the latest from Google Cloud? Find it here in one handy location. Check back regularly for our newest updates, announcements, resources, events, learning opportunities, and more. Tip: Not sure where to find what you’re looking for on the Google Cloud blog? Start here: Google Cloud blog 101: Full list of topics, links, and resources.
Dec 22 – Dec 26

To design production-ready AI agents, you must choose the right tools for memory, reasoning, and orchestration. To simplify this process, Choose your agentic AI architecture components provides an iterative framework to help you select products and tools that best match your workload characteristics. To build and deploy secure and reliable single-agent AI systems on a scalable serverless platform, see our Single-agent AI system using ADK and Cloud Run architecture guide.

Dec 15 – Dec 19

Announcing Advanced Governance Capabilities to Vertex AI Agent Builder: Today, with the integration of the Cloud API Registry, we’re excited to bring enhanced tool governance capabilities to Vertex AI Agent Builder. With this latest update, administrators can now manage available tools for developers across your organization directly in Vertex AI Agent Builder Console, and developers can leverage tools managed by the registry with a new ApiRegistry. Following last month’s expansion of our Agent Builder platform, we are also introducing new capabilities across the entire agent lifecycle to help developers build faster using new ADK capabilities and visual tools, and scale with high performance through the expansion of Agent Engine services, including the general availability of support for sessions and memory. Read more below.Vertex AI Agent Builder provides the unified platform to manage the entire agent lifecycle, helping you close the gap from prototype to a production-ready agent. To explore these new features, visit the updated Agent Builder documentation and release notes.
Single-tenant Cloud HSM is now Generally Available: We’re thrilled to announce the General Availability (GA) of Single-tenant Cloud HSM –  a standards compliant, highly available, and scalable HSM cluster that provides you complete control over your cryptographic keys for highly sensitive workloads in the cloud for general purpose applications. Customers have complete control over their cryptographic keys and the ability to manage their own admin credentials through our gcloud APIs, which establish a cryptographically isolated cluster of dedicated HSM partitions for each customer. Single-tenant Cloud HSM is integrated with Cloud KMS, allowing its use with Customer-Managed Encryption Keys (CMEK). Single-tenant Cloud HSM is available in the following regions: us-central1, us-east4, europe-west1, and europe-west4.
Advanced AI, data, and compliance security capabilities are now available to Security Command Center (SCC) Premium pay-as-you-go (PayGo) customers. Previously exclusive to Enterprise and Premium subscriptions, we now offer to PayGo customers the AI Security Dashboard, Data Security Posture Management (DSPM), Compliance Manager, and Security Graph, including Graph Search and Correlated Threats. This update can help you address novel risks from generative AI and autonomous agents by offering integrated, automated protection for both traditional and AI workloads in Google Cloud. Customers can start a 30-day free trial to access the full SCC Premium experience.

Dec 8 – Dec 12

Application Design Center is now Generally AvailableWe’re excited to announce the General Availability (GA) of Application Design Center, enabling platform teams and developers to streamline cloud application infrastructure design, deployment, and evolution, ensuring security and best practices.This GA launch includes powerful new capabilities such as enterprise-grade governance with public APIs and gcloud CLI support; bring your own Terraform, full compatibility with VPC Service Controls; and simplified onboarding with app-managed project support. To learn more, read the Application Design Center GA launch blog.

Apigee Feature Templater Simplifies API Proxy Development for EveryoneThe new open-source Apigee Feature Templater (AFT) streamlines API proxy authoring by turning complex policies into reusable building blocks called “features.” Non-experts can quickly assemble robust proxies—including AI Gateways, security, and data masking—using simple CLI or REST commands. AFT accelerates time-to-market by enabling expert developers to delegate feature creation and empowering a broader team to compose APIs. Read the full release details.

Navigating the Industry Shift in Client Authentication for Apigee mTLSAn industry policy change is phasing out the Client Authentication Extended Key Usage (EKU) in public certificates, directly impacting server-to-server mTLS for Apigee. This shift forces organizations away from Public CAs to manage their own Private PKI to maintain service continuity by mid-2026. This article presents the two paths: implementing a Private Certificate Authority (Private CA), ideally using Google Cloud Certificate Authority Service (CAS) for immediate mTLS continuity; or modernizing long-term with Demonstrating Proof of Possession (DPoP) for maximum operational efficiency. Read about the two paths to mTLS continuity.

Learn how to implement Kubernetes Secrets in Apigee hybridApigee hybrid introduces direct, read-only access to custom Kubernetes Secrets within API proxies. This exclusive feature offers a superior way to handle highly sensitive credentials (like private keys and backend passwords) compared to KVMs. Secrets never leave the cluster boundary, ensuring enhanced compliance and security. It enables a clean separation of duties, allowing cluster operators to manage credentials via GitOps workflows while API developers securely reference them using flow variables, without ever viewing the raw sensitive data. Read the full article.

Don’t let your AI program fail at the final hurdle. Our new guide, Successful Chatbot: 5 Steps from ROI to Rollout, outlines the essential practices for rigorous customer testing and strategic deployment. Learn how to align testing with business goals, define clear evaluation criteria, and drive actionable insights. The post emphasizes that delivering a successful AI program requires more than just domain expertise, highlighting the importance of clear scoping, strategic staffing, and disciplined financial planning. This is crucial for maximizing confidence in your AI’s long-term impact, especially in regulated industries like healthcare.Useful Product links – Google Cloud’s Vertex AI & Agentspace, Google Cloud’s Healthcare API, and, Google Cloud’s Natural Language API

Apigee Now Supports Model Context Protocol (MCP) for Secure Agentic ToolsGoogle has expanded support for Model Context Protocol (MCP) with the release of fully-managed, remote MCP servers, giving developers worldwide consistent and enterprise-ready access to Google and Google Cloud services. This includes support for MCP in Apigee, which makes it possible for agents to use your secure, governed APIs and custom workflows cataloged in Apigee API hub as tools to complete tasks for end users. With Apigee’s support for MCP, you don’t need to make any changes to your existing APIs, write any code, or deploy and manage any local or remote MCP servers. Apigee uses your existing API specifications and manages the underlying infrastructure and transcoding, so that you can focus on the business logic for your agents. Read the full announcement.

Introducing Fully-Managed MCP Servers to Power Agentic AIGoogle Cloud is announcing fully-managed, remote Model Context Protocol (MCP) servers, enhancing Google’s API infrastructure to provide a unified, enterprise-ready layer for AI agents. This eliminates the burden on developers to install and manage individual MCP servers. Now, AI agents can reliably use trusted Google services like Google Maps, BigQuery, Compute Engine, and GKE as tools to perform complex tasks. This unified approach, managed via Cloud IAM and Apigee API Hub, ensures rigorous governance, security, and observability for all agentic interactions. Read the full announcement.

Marketplace Customer Credit Program now available for Marketplace Channel Private OffersGoogle Cloud’s Marketplace Customer Credit Program offers up to 3% in Google Cloud credits when customers purchase an eligible cloud marketplace solution for the first time, whether directly through an ISV or via a chosen channel partner. Learn more.

Two-step control plane minor upgrades with rollback safety in Public previewUpgrading a production Kubernetes cluster can be a stressful, high-stakes event. GKE’s new two-step control plane minor upgrade with rollback safety gives you a safe window to validate a new minor version before committing, with a simple rollback option if you find any issues. By decoupling control plane binary changes from new API and feature changes, you can easily revert to the previous minor version if issues arise during the validation period. Learn more about this new, safer upgrade process.

Google named a Leader in the 2025 IDC MarketScape for Worldwide Hyperscaler MarketplacesIDC Marketscape has positioned Google as a Leader in the 2025 IDC MarketScape for Worldwide Hyperscaler Marketplaces. We believe this recognition underscores our commitment to deliver a cloud marketplace experience that fuels the AI agent economy and accelerates innovation. This achievement reflects our dedication to creating an open and interoperable agentic AI ecosystem for our customers and partners. Learn more.

Dec 1 – Dec 5

Unlock the full potential of your data with Object Contexts in Google Cloud StorageThis new feature provides a foundation for semantic storage and actions, allowing you to integrate Gemini with GCS objects and enrich your objects in a more intelligent and meaningful way. Learn how to get started with Object Contexts and revolutionize your data workflows. Learn more.

Nov 24 – Nov 28

Boost API Security: IP Allowlisting and ML Enhancements for Apigee Abuse DetectionTo keep your applications safe, it’s critical to detect and block attacks on your APIs as quickly as possible. In the past few months, we’ve made some changes to Advanced API Security’s Abuse Detection feature to make it easier and faster to identify legitimate attacks and take action. Get all the details on Apigee’s new IP allowlisting.
Apigee AI Gateway Deep Dive on December 11Join the final Apigee Community Tech Talk of the year for a deep dive into the Apigee AI Gateway. This session provides practical details on integrating, proxying, and converting complex MCP protocol services with traditional REST backends. Learn specific techniques for securing, monitoring, and gaining technical control over MCP backends to meet enterprise-grade governance requirements. Register now

Nov 10 – Nov 14

Deploy n8n to Cloud RunWith just a few commands, you can deploy n8n to Cloud Run and have it up and running, ready to supercharge your business with AI workflows that can manage spreadsheets, read and draft emails, and more. n8n and Cloud Run are both easy to use and powerful tools that empower developers to do more with AI. Learn more here.
GKE Node Memory Swap in Public previewYou can now configure swap space on your GKE Standard nodes to provide a crucial buffer against Out-of-Memory (OOM) errors for memory-intensive applications, especially during unexpected usage spikes. Enabling swap can improve workload resilience, reduce pod evictions due to memory pressure, and enhance overall application stability and cost-effectiveness. This feature is currently available in a public preview.

Nov 3 – Nov 7

Announcing the Data Engineering AgentData teams can now automate complex SQL pipeline tasks with the new Data Engineering Agent for BigQuery, available in Preview. This agent simplifies development, maintenance, and troubleshooting, allowing engineers to focus on strategic initiatives. It supports natural language pipeline creation, intelligent modification, and seamless migration from legacy tools.Transform your data engineering workflows today!
From Threat Model to TTX: Bringing a New Design Partner to the TableGain an overview of threat modeling, how threat models can be performed rapidly, and why threat model scenarios make excellent tabletop scenarios – especially for products that are still in development.To get more information about threat modeling or tabletop exercises, check out The Defender’s Advantage or reach out to a Mandiant cybersecurity expert for specialized assistance.
Application Monitoring now includes a Topology.Application Monitoring now includes a graphical representation of runtime dependencies (i.e Topology) for your App Hub defined application. This now allows you to quickly understand your app architecture, spot anomalous runtime interactions and resolve issues flagged from alerts quicker. Runtime dependencies are extracted from the OpenTelemetry traces you send to Cloud Trace from your App Hub registered workload.Follow the outline here to register your app and unlock the benefits of Application Monitoring and its newly launched Topology
Supercharge AI Agents: Apply Enterprise Governance to GenAI Workflows with ApigeeAs Generative AI agents move to production, you need control over cost, reliability, and security. A powerful new pattern introduces Apigee as the unified AI Agent Gateway for Large Language Model (LLM) calls. Route agent traffic through Apigee to gain immediate enterprise-grade governance, including dynamic circuit breaking, token consumption quotas, and sensitive data masking. A new Apigee wrapper for the Agent Development Kit (ADK) simplifies implementation. Turn your agents into manageable, secure AI products.Read the full article and explore the new pattern.

Oct 20 – Oct 24

Dataframe visualization in Colab Enterprise. Use visualization cells to create custom, stylized visualizations of your DataFrames: no coding required! Choose your fields, chart type, aggregation, and color scheme, then see a visualization of your data without leaving your notebook. Check out the tutorial and get started with data visualization today.

Oct 13 – Oct 17

Build Serverless AI in the Cloud Run HackathonReady to go from idea to global scale in minutes? The Cloud Run Hackathon is here! Build serverless AI apps with AI Studio, orchestrate intelligent agents, or harness the power of GPUs. Compete for a share of $50,000+ in prizes!

Submissions are open from Oct 6, 2025 to Nov 10, 2025.
Learn more and register: run.devpost.com

Oct 6 – Oct 10

Multi-agent AI systems help you optimize complex and dynamic processes by segmenting them into discrete tasks that multiple specialized AI agents collaboratively execute. To get started with building secure and reliable multi-agent AI systems, see this reference architecture guide: Design a multi-agent AI system in Google cloud. The example architecture in this guide showcases a couple of agent patterns: sequential, and loop. For a comprehensive review of all the possible agent design patterns and for help with choosing patterns that are appropriate for your use cases, see this design guide: Choose a design pattern for your agentic AI system.

Sept 29 – Oct 3

Announcing Koog Supports for Agent2Agent protocol (A2A)The future of interconnected AI is here. We’re thrilled to announce that Koog now supports A2A, a protocol that lets agents talk directly, securely, and seamlessly across companies and clouds. For Kotlin developers, this unlocks a new era of powerful, enterprise-grade AI. Build sophisticated agents that automatically discover and collaborate with other services, all while calling on Google Cloud’s state-of-the-art models like Gemini directly from your workflows. Stop building bridges and start creating truly intelligent, interconnected systems today. Learn more about building with Koog, A2A, and Google Cloud.

Sept 15 – 19

Your AI is Now a Local Expert: Grounding with Google Maps is GAWe are excited to announce the General Availability (GA) of Grounding with Google Maps in Vertex AI. This feature lets developers build generative AI applications that are connected to real-world, up-to-date information from Google Maps, using its data on over 250 million places worldwide.To learn more and get started, visit our documentation and check out our demo.
Production-ready YOLO model training serving workflow on Vertex AIThis guide walks you through a complete, automated workflow for training a custom YOLO model on Vertex AI. You’ll learn how to use a custom training job, package the model in a custom prediction container, and register it in the Vertex AI Model Registry, making it ready for easy deployment. Best of all, this approach is designed to work directly with existing Vertex AI managed datasets for object detection, meaning you can reuse the same data you’re already using for AutoML models.Checkout details on developer forums

Sept 8 – 12

Scaling Inference To Billions of Users And AI Agents: Discover the architecture required to serve AI models at a planetary scale. This article details how Google Cloud’s ecosystem—from the GKE Inference Gateway for smart load balancing to the power of custom TPUs and open-source engines like vLLM—provides a production-ready path. Move beyond the hype and learn how to build for the next wave of AI. Explore the technical deep-dive.
We’re celebrating the one-year anniversary of bringing Confidential Computing with Intel TDX to Google Cloud. We’ve been shipping new capabilities to help you protect your most sensitive data while it’s in use. Now Generally Available:

Confidential GKE Nodes with Intel TDX: Secure entire Kubernetes clusters, node pools, and workloads.
Confidential Space with Intel TDX: Build secure data clean rooms for collaboration on sensitive information.
Confidential GPUs: Protect cutting-edge AI workloads with Confidential NVIDIA H100s GPUs on GCE and GKE.We’ve also expanded Intel TDX to more regions! Read the blog

Aug 25 – 29

Applied AI for Modern Manufacturers: New original growth series, hosted by Jake Hall, The Manufacturing Millennial, that dives into leading trends, best practices, and what companies are doing right now with AI in manufacturing. Hear from industry thought leaders – Rick Bullotta, Jonathan Wise, Walker Reynolds and Berardino Baratta – and Google Cloud experts – Praveen Rao, Eric Lam, Dave Nguyen Ph.D., Geoffrey Hirschheim, and Jim Anderson. Watch Modules 1 and  2 now, where we delve into the AI Innovation and trends and AI Costs and ROI in the Era of Digital Manufacturing. Next module kicks off Tuesday, Sep 2. Join now

Firestore with MongoDB compatibility is now generally available (GA): Developers can now build cost-effective, scalable, and highly reliable apps on Firestore’s serverless database using a familiar MongoDB-compatible API. With the general availability of Firestore with MongoDB compatibility, the 600,000 active developers within the Firestore community can now use existing MongoDB application code, drivers, and tools, as well as the open-source MongoDB ecosystem, with Firestore’s serverless service. Firestore offers benefits like multi-region replication, virtually unlimited scalability, up to 99.999% SLA, single-digit millisecond read performance, integrated Google Cloud governance, and pay-as-you-go pricing. Register now for the webinar on September 9th for a deep dive into Firestore with MongoDB compatibility.

Aug 18 – 22

Earth Engine in BigQuery is now Generally Available, bringing advanced geospatial analytics directly to your BigQuery workflows. Unlock insights with satellite data!

Aug 11 – Aug 15

New HPC VM and Slurm-gcp Images: A new HPC VM Image (under the project cloud-hpc-image-public) is now available, featuring a Rocky Linux 8-based image, IntelMPI v2021.16, and RDMA drivers. In partnership with SchedMD, new Slurm images (Slurm 25.05) have also been released. These are based on the latest HPC VM Image and are available for Ubuntu 22.04/24.04 Accelerator Images (ARM/AMD64) and Debian 12. These releases allow for the deployment of Slurm-ready clusters on GCP, providing the advantages of an HPC-optimized and performance-tested foundation. Read more.
Scaling our Gemini Embedding model in Vertex AI. Following increased popularity from its General Availability launch in May, we’ve recently increased quota and input size limits for customers of Vertex AI’s most powerful text embedding model, gemini-embedding-001.

Customers can now send up to 250 input texts per request (generating 250 embeddings) instead of only a single piece of text, bringing improved throughput and decreased round-trip network latency to large-scale embedding applications.
We’ve increased quota limits for this model by 10x for most users, allowing hassle-free scaling of embedding applications to millions of tokens per minute and beyond.Get started with Gemini Embeddings today!

Aug 4 – Aug 8

GKE Node Memory Swap in private preview: You can now configure swap space on your GKE Standard nodes to provide a crucial buffer against Out-of-Memory (OOM) errors for memory-intensive applications, especially during unexpected usage spikes. Enabling swap can improve workload resilience, reduce pod evictions due to memory pressure, and enhance overall application stability and cost-effectiveness. This feature is currently available in a private preview.

Contact your Google Cloud account team for more information and to request access.

If you’d like to see more configurations, please contact your account team or make a feature request on our issue tracker!

Unlock Peak Performance: GKE Topology Manager is Now Generally Available: For customers running performance-sensitive workloads like AI/ML and HPC, GKE Topology Manager is now GA and ready to optimize your performance through NUMA alignment. By ensuring CPU, memory, and GPU resources are allocated on the same NUMA node, the Topology Manager minimizes cross-socket latency and maximizes throughput for your most demanding applications. Configure your alignment policies via the NodeConfig API to achieve significant performance gains.

Achieve these performance gains by configuring your alignment policies via the NodeConfig API.
If you’d like to see more expansion of Topology manager, please contact your account team or make a feature request on our issue tracker!

Fine-Tune at Scale: A Massive GKE NodeConfig Expansion for All Workloads: GKE has massively expanded node customization capabilities, adding nearly 130 new Sysctl and Kubelet configurations. This gives you finer-grained control for any workload needing node customization, performance requirements, or application-specific tuning. By replacing complex DaemonSets with native controls, you can benefit from enhanced security, high flexibility, faster node startup times, and less operational management.

Check out our public documentation to learn how to consume these new NodeConfig options.
If you’d like to see more configurations, please contact your account team or make a feature request on our issue tracker!

New capability for managing licenses in Compute Engine: We are announcing a new capability in Compute Engine which allows users to easily change the OS licenses on their VMs. Users can now append, remove, or replace OS licenses, enabling seamless transitions between license types—such as converting Red Hat Enterprise Linux from pay-as-you-go (PAYG) to bring-your-own subscription (BYOS), or upgrading from Ubuntu to Ubuntu Pro—without needing to redeploy instances. This feature empowers customers to meet their evolving licensing with speed and flexibility. To learn more, read about managing licenses on Compute Engine.

GKE Turns 10 Hackathon: Calling all developers! Google Kubernetes Engine (GKE) is turning 10, and we’re celebrating with a hackathon! Join us to build powerful AI agents that interact with microservice applications using Google Kubernetes Engine and Google AI models. Compete for over $50,000 in prizes and demonstrate the power of building agentic AI on GKE.

Submissions are open from Aug 18, 2025 to Sept, 22 2025
Learn more and register: gketurns10.devpost.com

Jul 28 – Aug 1

Now GA: C4 VMs with Local SSD, bare metal, and larger shapes, on Intel Xeon 6: C4’s expanded shapes are now GA! This expansion introduces C4 shapes with Google’s next-gen Titanium Local SSD, C4 bare metal instances, and new extra-large shapes, all powered by the latest Intel Xeon 6 processors, Granite Rapids. We’re excited to be the first leading hyperscaler to bring Xeon 6 to customers, delivering performance gains of up to 30% for general compute and up to 60% for ML recommendation workloads, and up to 35% lower access latency on Titanium Local SSD shapes. Learn more here!

Jul 14 – 18

DMS SQL Server to PostgreSQL migrations are now generally available! Accelerate your SQL Server modernization to Cloud SQL for PostgreSQL or AlloyDB for PostgreSQL with:

Automatic database schema and code conversion 
Gemini augmented code conversion 
Gemini assisted PostgreSQL training and code improvements
Low-downtime, CDC based data movement

Learn more and start your migration journey today!
Jul 7 – 11

Level up your AI Agent game with “The Agent Factory,” a new video podcast for developers! We’re going beyond the buzz to explore practical design, build, deploy, & management strategies for production-ready AI agents using Google Cloud. Expect code snippets, architecture deep dives, and integrations with open-source frameworks. Subscribe now!

Jun 23 – 27

Announcing partnership between Maxim AI and Google Cloud’s Vertex AI to evaluate agentic applications — Maxim AI offers a comprehensive platform to help teams build, evaluate, and observe their AI agents with greater speed and confidence, covering the entire AI lifecycle from prompt engineering to production monitoring. This new partnership deeply integrates Vertex AI’s Gen AI evaluation service directly within the Maxim AI environment, allowing users to leverage Gemini to power assistant responses and evaluate them using Vertex AI’s comprehensive suite of evaluators. This provides access to metrics such as helpfulness, relevance, safety, and trajectory. The setup allows users to simulate, evaluate, and trace complex multi-turn interactions on Maxim, helping teams bring reliable AI products to market faster through a seamless developer experience. To learn more, check out this blog from Maxim AI

Run non-request workloads at scale with Cloud Run Worker Pools, now in Public Preview — Looking for the ease-of-use and scalability of serverless, without being limited to HTTP request-driven workloads? Cloud Run Worker Pools provide the same elasticity and high-quality developer experience as Cloud Run Services, but are designed for non-request workloads. Worker Pools are ideal for pull-based use cases like processing messages from Pub/Sub or Kafka, and other backend processing.  Check out the public documentation to learn more about how to choose between Services, Jobs, and Worker Pools. Then give Worker Pools a try by deploying a sample Worker Pool.

Building a Multi-Agent Research Assistant for Financial Analysis with Schroders & Google Cloud — Financial analysts spend hours grappling with ever-increasing volumes of market and company data to extract key signals, combine diverse data sources, and produce company research. To maximise its edge as an active manager, Schroders wants to enable its analysts to shift from data collection to the higher-value strategic thinking that is critical for business scalability and client investment performance.  To achieve this, Schroders and Google Cloud collaborated to build a multi-agent research assistant prototype using Vertex AI Agent Builder. Find out more here.

Jun 16 – 20

Simplify Your Multi-Cloud Strategy with Cloud Location Finder, now in Public Preview: As cloud environments expand beyond traditional architectures to include multiple clouds, managing your infrastructure effectively becomes more complex. Imagine effortlessly accessing consistent and up-to-date location information across different cloud providers, so your multi-cloud applications are designed and optimized with performance, security, and regulatory compliance in mind. Today, we are making this a reality with Cloud Location Finder, a new Google Cloud service which provides up-to-date location data across Google Cloud, Amazon Web Services (AWS), Azure, and Oracle Cloud Infrastructure (OCI). Now, you can strategically deploy workloads across different cloud providers with confidence and control. Cloud Location Finder is accessible via REST APIs and gcloud CLI, explore the Cloud Location Finder documentation and blog to learn more.

SOTA Gemini Text Embedding is Now Generally Available in Vertex AI: We recently launched a new Gemini Embedding text model (gemini-embedding-001) through the Vertex AI GenAI API. This groundbreaking model, leveraging Gemini’s core language understanding, sets a new benchmark for text embeddings. It’s the first unified model to excel across English, multilingual text, and code, outperforming previous models (text-embedding-005, text-multilingual-embedding-002) and achieving top ranking on the MTEB Multilingual leaderboard (100+ tasks). Our internal benchmarks demonstrate substantial performance improvements across various industry verticals, including retail, news, finance, healthcare, legal, and code. Detailed results are available in our technical report.

Backup vaults now support disk backups and multi-regions: We’ve added exciting new features to Google Cloud Backup and Disaster Recovery service! You can now secure your Persistent Disk and Hyperdisk backups in backup vaults, protecting them from cyber attacks and accidental data loss. In addition, backup vaults can now be created in multi-region storage locations, maximizing your data resilience and supporting compliance with business continuity requirements. Check out the blog to learn more!

DeepSeek R1, a powerful 671B parameters model, is now available as a fully managed API on Vertex AI in Preview, making advanced AI capabilities more accessible to developers. This Model as a Service (MaaS) offering eliminates the need for extensive GPU resources and infrastructure management, allowing developers to focus on building applications. DeepSeek R1 on Vertex AI provides a simple, scalable API with features like transparent “chain-of-thought” reasoning and enterprise-ready security. It’s currently available at no additional cost during the preview, and can be accessed via UI, REST API, or the OpenAI Python API Client Library. Learn more.

Jun 9 – 13

Serverless Spark Now GA in BigQuery: Unified Analytics, Accelerated: Google Cloud Serverless for Apache Spark is now generally available in BigQuery, offering a unified developer experience in BigQuery Studio. Run Spark and SQL side-by-side on the same data, powered by the Lightning Engine for up to 3.6x faster performance and enhanced with Gemini productivity. Simplify your data pipelines and accelerate insights with this deeply integrated, zero-ops solution.
Cloud Pub/Sub introduced Pub/Sub Single Message Transforms (SMTs) to make it easy to perform simple data transformations right within Pub/Sub: An overarching goal of Pub/Sub is to simplify streaming architectures. We already greatly simplified data movement with Import Topics and Export Subscriptions, which removed the need to use additional services for ingesting raw streaming data through Pub/Sub into destinations like BigQuery. Pub/Sub Single Message Transforms (SMTs), designed to be a suite of features making it easy to validate, filter, enrich, and alter individual messages as they move in real time. The first SMT is available now: JavaScript User-Defined Functions (UDFs), which allows you to perform simple, lightweight modifications to message attributes and/or the data directly within Pub/Sub via snippets of JavaScript code. JavaScript UDFs as the first Single Message Transform is generally available starting today for all users. You’ll find the new “Add Transform” option in the Google Cloud console when you create a topic or subscription in your Google Cloud project. You can also use gcloud CLI to start using JavaScript Single Message Transforms today.
This analysis evaluates the efficiency of fine-tuning a Llama 3-8B model on Vertex AI using both a single A100 GPU and a distributed four-A100 setup with Axolotl. While both methods achieved similar model convergence, the results underscore the power of distributed training. The process, which took 1 day and 20 hours on a single device, was completed in just 11 hours in the distributed environment—a dramatic acceleration. This speed was achieved with consistently high GPU utilization (94%), though at the cost of higher system and GPU memory overhead. For a detailed breakdown of the methodology, resource utilization metrics, and performance curves, you can review the complete work here.

May 26 – 30

Cloud Run GPUs are now GA: NVIDIA GPU support for Cloud Run is now generally available, offering a powerful runtime for a variety of use cases that’s also remarkably cost-efficient. Developers can now get on-demand access to GPUs with our serverless runtime, Cloud Run. Follow the footsteps of customers like MidJourney, vivo, and Wayfair. Read blog.
Datastream now supports MongoDB as a source! Seamlessly ingest data from MongoDB (Replica Sets, Sharded Clusters, self-hosted, AtlasDB) into BigQuery/Cloud Storage. Enjoy scalable, fully-managed data streaming with backfill and CDC, enabling real-time insights and data-driven decisions. Link

May 19 – May 23

Beyond cuts and fades: Understanding narrative flow with Gemini for accurate scene transition detection — Google Cloud’s Gemini models are revolutionizing video understanding by accurately detecting narrative scene transitions, moving beyond simple cuts and fades. This breakthrough technology understands the holistic context of videos by analyzing visual, audio, and textual elements simultaneously. Media companies can now convert passive video assets into structured data, enabling intelligent content discovery, strategic ad placement, and personalized viewing experiences. The result? Up to 38% increased viewer engagement and 27% reduced abandonment rates. 

Read more on the medium blog. 

Learn more and access the code repository: View Code Repo

Announced at I/O: Deploy AI apps to Cloud Run from AI Studio and MCP — We are making AI deployments easier and more accessible by introducing new ways to deploy your apps to Cloud Run.

You can deploy applications developed in AI Studio with a click of a button to Cloud Run, including Gemma 3. 

Model Context Protocol(MCP) is becoming a popular open protocol standardizing how AI agents interact with other tools. Now with Cloud Run MCP server, you can deploy apps from compatible AI agents like from Claude or VS Code Copilot.

Read blog to learn more.

May 12 – May 16

Google for Startups Accelerator: AI For Energy now accepting applications!Applications are now open for startups headquartered in Europe and Israel, working on solutions for utilities, grid operators and energy developers; solutions for residential and commercial end-use customers focused on demand flexibility and solutions for industrial customers. This equity-free program offers 10 weeks of intensive mentorship and technical project support to startups integrating AI into their core energy services or products. Selected startups will collaborate with a cohort of peer founders and engage with leaders across Google and the energy sector. The curriculum will provide founders with access to AI tools and include workshops on tech and infrastructure, UX and product, growth, sales, leadership and more. Learn more and apply before June 30th, 2025. 

Extending Google Cloud Workstations containers to run any GUI based programAre you having difficulty customizing Google Cloud Workstations to run a GUI program outside of the supported configurations of IDE’s? If so, you’re not alone. In this article we discuss how to use the base Workstations Docker image and build it to run a terminal and Google Chrome.

Google Cloud Marketplace simplifies deals and improves economics. Announcing three initiatives that build upon Google Cloud Marketplace as a growth engine for customers and partners:

Improving partner deal economics to help partners retain more earnings by moving to a variable revenue share model

Simplifying commit drawdown for purchases through channel partners

Unlocking new workloads with the Marketplace Customer Credit Program incentiveLearn more

2025 Google Cloud DORA Awards are now open for submission!Has your team achieved remarkable success through DORA principles? It’s time to shine. We’re thrilled to announce the launch of the 2025 Google Cloud DORA Awards, celebrating outstanding achievements in technology delivery and operational performance. Submit your story today!

May 5 – May 9

AI assisted development with MCP Toolbox for DatabasesWe are excited to announce new updates to MCP Toolbox for Databases. Developers can now use Toolbox from their preferred IDE, such as Cursor, Windsurf, Claude Desktop, more and leverage our new pre-built tools such as execute_sql and list_tables for AI-assisted development with Cloud SQL for PostgreSQL, AlloyDB and self-managed PostgreSQL.

Get Started with MCP Toolbox for Databases

Apr 28 – May 2

Itching to build AI agents? Join the Agent Development Kit Hackathon with Google Cloud! Use ADK to build multi-agent systems to solve challenges around complex processes, customer engagement, content creation, and more. Compete for over $50,000 in prizes and demonstrate the power of multi-agent systems with ADK and Google Cloud.

Submissions are open from May 12, 2025 to June 23, 2025.
Learn more and register here.

Apr 21 – Apr 25

Iceland’s Magic: Reliving Solo Adventure through GeminiEmbark on a journey through Iceland’s stunning landscapes, as experienced on Gauti’s Icelandic solo trip. From majestic waterfalls to the enchanting Northern Lights, Gautami then takes these cherished memories a step further, using Google’s multi-modal AI, specifically Veo2, to bring static photos to life. Discover how technology can enhance and dynamically relive travel experiences, turning precious moments into immersive short videos. This innovative approach showcases the power of AI in preserving and enriching our memories from Gauti’s unforgettable Icelandic travels. Read more.

Introducing ETLC – A Context-First Approach to Data Processing in the Generative AI Era: As organizations adopt generative AI, data pipelines often lack the dynamic context needed. This paper introduces ETLC (Extract, Transform, Load, Contextualize), adding semantic, relational, operational, environmental, and behavioral context. ETLC enables Dynamic Context Engines for context-aware RAG, AI co-pilots, and agentic systems. It works with standards like the Model Context Protocol (MCP) for effective context delivery, ensuring business-specific AI outputs. Read the full paper.

Apr 14 – Apr 18

What’s new in Database CenterWith general availability, Database Center now provides enhanced performance and health monitoring for all Google Cloud databases, including Cloud SQL, AlloyDB, Spanner, Bigtable, Memorystore, and Firestore. It delivers richer metrics and actionable recommendations, helps you to optimize database performance and reliability, and customize your experience. Database Center also leverages Gemini to deliver assistive performance troubleshooting experience. Finally, you can track the weekly progress of your database inventory and health issues. 
Get started with Database Center today

Access Database Center in Google Cloud console

Review the documentation to learn more

Apr 7 – Apr 11

This week, at Google Cloud Next, we announced an expansion of Bigtable’s SQL capabilities and introduced continuous materialized views. Bigtable SQL and continuous materialized views empower users to build fully-managed, real-time application backends using familiar SQL syntax, including specialized features that preserve Bigtable’s flexible schema — a vital aspect of real-time applications. Read more in this blog.
DORA Report Goes Global: Now Available in 9 Languages!Unlock the power of DevOps insights with the DORA report, now available in 9 languages, including Chinese, French, Japanese, Korean, Portuguese, and Spanish. Global teams can now optimize their practices, benchmark performance, and gain localized insights to accelerate software delivery. The report highlights the significant impact of AI on software development, explores platform engineering’s promises and challenges, and emphasizes user-centricity and stable priorities for organizational success. Download the DORA Report Now
New Google Cloud State of AI Infrastructure Report ReleasedIs your infrastructure ready for AI? The 2025 State of AI Infrastructure Report is here, packed with insights from 500+ global tech leaders. Discover the strategies and challenges shaping the future of AI and learn how to build a robust, secure, and cost-effective AI-ready cloud. Download the report and enhance your AI investments today. Download the 2025 AI infrastructure report now
Google Cloud and Oracle Accelerate Enterprise Modernization with New Regions, Expanded CapabilitiesAnnouncing major Oracle Database@Google Cloud enhancements! We’re launching the flexible Oracle Base Database Service and powerful new Exadata X11M machines. We’re rapidly expanding to 20 global locations, adding new Partner Cross-Cloud Interconnect options, and introducing Cross-Region Disaster Recovery for Autonomous Database. Benefit from enhanced Google Cloud Monitoring, integrated Backup & DR, plus expanded support for enterprise applications like SAP. Customers can run critical Oracle workloads with more power, resilience, and seamless Google Cloud integration. Get started right away from your Google Cloud Console or learn more here.

Mar 17 – Mar 21

Cloud CISO Perspectives: 5 tips for secure AI success – To coincide with new AI Protection capabilities in Security Command Center, we’re offering 5 tips to set up your organization for secure AI success.
Our 4-6-3 rule for strengthening security ties to business: The desire to quickly transform a business can push leaders to neglect security and resilience, but prioritizing security can unlock value. Follow these 4 principles, 6 steps, and 3 metrics to use a security-first mindset to drive business results.
The new Data Protection Tab in Compute Engine ensures your resources are protected: Not only have we co-located your backup options, but we also have introduced smart default data protection for any Compute Engine instance created via Cloud Console. Here’s how it works.
DORA report – Impact of Generative AI in Software DevelopmentThis report builds on and extends DORA’s research into AI. We review the current landscape of AI adoption, look into its impact on developers and organizations, and outline a framework and practical guidance for successful integration, measurement, and continuous improvement. Download the report!

Mar 10 – Mar 14

Protecting your APIs from OWASP’s top 10 security threats: We compare OWASP’s top 10 API security threats list to the security capabilities of Apigee. Here’s how we hold up.

Project Shield makes it easier to sign up, set up, automate DDoS protection: It’s now easier than ever for vulnerable organizations to apply to Project Shield, set up protection, and automate their defenses. Here’s how.

How Google Does It: Red teaming at Google scale – The best red teams are creative sparring partners for defenders, probing for weaknesses. Here’s how we do red teaming at Google scale.

AI Hypercomputer is a fully integrated supercomputing architecture for AI workloads – and it’s easier to use than you think. Check out this blog, where we break down four common use cases, including reference architectures and tutorials, representing just a few of the many ways you can use AI Hypercomputer today. 

Transform Business Operations with Gemini-Powered SMS-iT CRM on Google Cloud: SMS-iT CRM on Google Cloud unifies SMS, MMS, email, voice, and 22+ social channels into one Smart Inbox. Enjoy real-time voice interactions, AI chatbots, immersive video conferencing, AI tutors, AI operator, and unlimited AI agents for lead management. Benefit from revenue-driven automation, intelligent appointment scheduling with secure payments, dynamic marketing tools, robust analytics, and an integrated ERP suite that streamlines operations from project management to commerce. This comprehensive solution is designed to eliminate inefficiencies and drive exponential growth for your business. Experience the Future Today.

Join us for a new webinar, Smarter CX, Bigger Impact: Transforming Customer Experiences with Google AI, where we’ll explore how Google AI can help you deliver exceptional customer experiences and drive business growth. You’ll learn how to:

Transform Customer Experiences:  With conversational AI agents that provide personalized customer engagements.

Improve Employee Productivity & Experience: With AI that monitors customers sentiment in real-time, and assists customer service representatives to raise customer satisfaction scores.

Deliver Value Faster: With  30+ data connectors and 70+ action connectors to the most commonly used CRMs and information systems.Register here

Mar 3 – Mar 7

Hej Sverige! Google Cloud launches new region in Sweden – More than just another region, it represents a significant investment in Sweden’s future and Google’s ongoing commitment to empowering businesses and individuals with the power of the cloud. This new region, our 42nd globally and 13th in Europe, opens doors to opportunities for innovation, sustainability, and growth — within Sweden and across the globe. We’re excited about the potential it holds for your digital transformations and AI aspirations.
[March 11th webinar] Building infrastructure for the Generative AI era: insights from the 2025 State of AI Infra report: Staying at the forefront of AI requires an infrastructure built for AI. Generative AI is revolutionizing industries, but it demands a new approach to infrastructure. In this webinar, we’ll unveil insights from Google Cloud’s latest research report and equip tech leaders with a practical roadmap for building and managing gen AI workloads, including: the top gen AI use cases driving the greatest return on investment, current infrastructure approaches and preferences for Generative AI workloads, the impact of performance benchmarks, scalability, and security on cloud provider selection. Register today.
Cloud CISO Perspectives: Why PQC is the next Y2K, and what you can do about it: Much like Y2K 25 years ago, post-quantum cryptography may seem like the future’s problem — but it will soon be ours if IT doesn’t move faster, explains Google Cloud’s Christiane Peters. Here’s how business leaders can get going on PQC prep.
How Google Does It: Using threat intelligence to uncover and track cybercrime — How does Google use threat intelligence to uncover and track cybercrime? Google Threat Intelligence Group’s Kimberly Goody takes you behind the scenes.
5 key cybersecurity strategies for manufacturing executives — Here are five key governance strategies that can help manufacturing executives build a robust cybersecurity posture and better mitigate the evolving risks they face.
Datastream now offers Salesforce source in Preview. Instantly connect, capture changes, and deliver data to BigQuery, Cloud Storage, etc. Power real-time insights with flexible authentication and robust backfill/CDC. Unlock Salesforce data for Google Cloud analytics, reporting, and generative AI. Read the documentation to learn more.
Find out how much you can save with Spanner – According to a recent Forrester Total Economic Impact™ study, by migrating to Spanner from a traditional database, a $1 billion per year B2C organization could get a 132% return on investment (ROI) with a 9-month payback period, and realize $7.74M in total benefits over the three years. To see how, check out the blog or download the report. 
GenAI Observability for Developers series: The Google Cloud DevRel team hosted a four-part webinar series, “Gen AI Observability for Developers,” demonstrating observability best practices in four programming languages. Participants learned to instrument a sample application deployed on Cloud Run for auditing Vertex AI usage, writing structured logs, tracking performance metrics, and utilizing OpenTelemetry for tracing. The series covered Go, Java, NodeJS, and Python, using common logging and web frameworks. Missed it? Recordings and hands-on codelabs are available to guide you at:

Gen AI O11y for Go Developers
Gen AI O11y for Java Developers
Gen AI O11y for NodeJS Developers
Gen AI O11y for Python DevelopersStay tuned for future events at cloudonair.withgoogle.com.

Feb 24 – Feb 28

Rethinking 5G: Ericsson and Google Cloud are collaborating to redefine 5G mobile core networks with a focus on autonomous operations. By leveraging AI and cloud infrastructure, we aim to enhance efficiency, security, and innovation in the telecommunications industry. This partnership addresses the increasing demands of 5G and connected devices, paving the way for a more dynamic and intelligent network future, and setting the stage for next-generation technologies like 6G. Learn more here.
Adopt a principles-centered well-architected framework to design, build, deploy, and manage Google Cloud workloads that are secure, resilient, efficient, cost-efficient, and high-performing. Also get industry and technology-focused well-architected framework guidance, like for AI and ML workloads.

Feb 17 – Feb 21

Easier Default Backup Configuration for Compute Engine Instances – The Create a Compute Instance page in the Google Cloud console now includes enhanced data protection options to streamline backup and replication configurations. By default, an option to back up data is pre-selected, ensuring recoverability in case of unforeseen events. Learn more here.

Feb 10 – Feb 14

[Webinar] Generative AI for Software Delivery: Strategies for IT Leaders: Generative AI is transforming the way organizations build and deploy software. Join Google Cloud experts on February 26th to learn how organizations can leverage AI to streamline their software delivery, including: the role of gen AI in software development, how to use gen AI for migration and modernization, best practices for integrating gen AI into your existing workflows, and real-world applications of gen AI in software modernization and migration through live demos. Register here.

Feb 3 – Feb 7

SQL is great but not perfect. We’d like to invite you to reimagine how you write SQL with Google’s newest invention: pipe syntax (public available to all BigQuery and Cloud Logging users). This new extension to GoogleSQL brings a modern, streamlined approach to data analysis. Now you can write simpler, shorter and more flexible queries for faster insights. Check out this video to learn more. 

Jan 13 – Jan 17

C4A virtual machines with Titanium SSD—the first Axion-based, general-purpose instance with Titanium SSD, are now generally available. C4A virtual machines with Titanium SSDs are custom designed by Google for cloud workloads that require real-time data processing, with low-latency and high-throughput storage performance. Titanium SSDs enhance storage security and performance while offloading local storage processing to free up CPU resources. Learn more here.

Jan 6 – Jan 10

A look back on a year of Earth Engine advancements: 2024 was a landmark year for Google Earth Engine, marked by significant advancements in platform management, cloud integration, and core functionality and increased interoperability between Google Cloud tools and services. Here’s a round up of 2024’s top Earth Engine launches.
Get early access to our new Solar API data and features: We’re excited to announce that we are working on 2 significant expansions to the Solar API from Google Maps Platform and are looking for trusted testers to help us bring them to market. These include improved and expanded buildings coverage and greater insights for existing solar installations with Detected Arrays. Learn more.
Google for Startups Accelerator: Women Founders applications are now open for women-led startups headquartered in Europe and Israel. Discover why this program could be the perfect fit for your startup and apply before January 24th, 2025.
Best of N: Generating High-Quality Grounded Answers with Multiple Drafts – We are excited to announce that Check Grounding API has released a new helpfulness score feature. Building on top of our existing groundedness score, we now enable users to implement Best of N to improve RAG response quality without requiring extensive model retraining. Learn more about Best of N and how it can help you here.

Quelle: Google Cloud Platform

The Year in Google Cloud — 2025

In the AI era, when one year can feel like 10, you’re forgiven for forgetting what happened last month, much less what happened all the way back in January. To jog your memory, we pulled the readership data for top product and company news of 2025. And because we publish a lot of great thought leadership and customer stories, we pulled that data too. Long story short: the most popular stories largely mapped to our biggest announcements. But not always — there were more than a few sleeper hits on this year’s list. Read on to relive this huge year, and perhaps discover a few gems that you may have missed. 

Building tomorrow, today: 2025 customer AI innovation highlights with Google Cloud

January
2025 started strong with important new virtual machine offerings, foundational AI tooling, and tools for both Kubernetes and data professionals. We also launched our “How Google Does It” series, looking at the internal systems and engineering principles behind how we run a modern threat-detection pipeline. We showed developers how to get started with JAX and made AI predictions for the year ahead. Readers were excited to learn about how L’Oréal built its MLOps platform and Deutsche Börse’s pioneering work on cloud-native financial trading.
Product news

Simplify the developer experience on Kubernetes with KRO

Blackwell is here — new A4 VMs powered by NVIDIA B200 now in preview

Introducing Vertex AI RAG Engine: Scale your Vertex AI RAG pipeline with confidence

Introducing BigQuery metastore, a unified metadata service with Apache Iceberg support

C4A, the first Google Axion Processor, now GA with Titanium SSD

Thought leadership:

How Google Does It: Making threat detection high-quality, scalable, and modern

2025 and the Next Chapter(s) of AI

Customer stories

How L’Oréal Tech Accelerator built its end-to-end MLOps platform

Trading in the Cloud: Lessons from Deutsche Börse Group’s cloud-native trading engine

FebruaryThere are AI products, and then there are products enhanced by AI. This month’s top launch, Gen AI Toolbox for Databases, falls into the latter category. This was also the month readers got serious about learning, with blogs about upskilling, resources, and certifications topping the charts. The fruits of our partnership with Anthropic made an appearance in our best-read list, and engineering leaders detailed Google’s extensive efforts to optimize AI system energy consumption. Execs ate up an opinion piece about how agents will unlock insights into unstructured data (which makes up 90% of enterprises’ information assets), and digested a sobering report on AI and cybercrime. During the Mobile World Congress event, we saw considerable interest in our work with telco leaders like Vodafone Italy and Amdocs.Product and company news:Announcing public beta of Gen AI Toolbox for DatabasesGet Google Cloud certified in 2025—and see why the latest research says it mattersDiscover Google Cloud careers and credentials in our new Career DreamerAnnouncing Claude 3.7 Sonnet, Anthropic’s first hybrid reasoning model, is available on Vertex AIThought leadershipDesigning sustainable AI: A deep dive into TPU efficiency and lifecycle emissionsFrom dark data to bright insights: How AI agents make data simpleNew AI, cybercrime reports underscore need for security best practicesCustomer storiesTransforming data: How Vodafone Italy modernized its data architecture in the cloudAI-powered network optimization: Unlocking 5G’s potential with Amdocs

MarchBack when we announced it, our intent to purchase cybersecurity startup Wiz was Google’s largest deal ever, and the biggest tech deal of the year. We built on that security momentum with the launch of AI Protection. We also spread our wings to the Nordics with a new region, and announced the Gemma 3 open model on Vertex AI. Meanwhile, we explained the threat that North Korean IT workers pose to employers, gave readers a peek under the hood of the Colossus file system, and reminisced about what we’ve learned over 25 years of building data centers. Readers were interested in Levi’s approach to data and weaving it into future AI efforts, and in honor of the GDC Festival of Gaming, our AI partners shared some new perspectives on “living games.”Product and company newsGoogle + Wiz: Strengthening Multicloud SecurityAnnouncing AI Protection: Security for the AI eraHej Sverige! Google Cloud launches new region in SwedenAnnouncing Gemma 3 on Vertex AIThought leadershipThe ultimate insider threat: North Korean IT workersColossus under the hood: How we deliver SSD performance at HDD prices3 key lessons from 25 years of warehouse scale computingCustomer storiesLevi’s seamless data strategy: How tailor-made AI keeps an icon from getting hemmed inCo-op mode: New partners driving the future of gaming with AI

AprilWith April came Google Cloud Next, our flagship annual conference. From Firebase Studio, Ironwood TPUs, and Google Agentspace, to Vertex AI, Cloud WAN, and Gemini 2.5, it’s hard to limit ourselves to just a few stories, there were so many bangers (for the whole list, there’s always the event recap). Meanwhile, our systems team discussed innovations to keep data center infrastructure’s thermal envelope in check. And at the RSA Conference, we unveiled our vision for the agentic security operations center of the future. On the customer front, we highlighted the startups who played a starring role at Next, and took a peek behind the curtain of The Wizard of Oz at Sphere.Product and company newsIntroducing Firebase Studio and agentic developer tools to build with GeminiIntroducing Ironwood TPUs and new innovations in AI HypercomputerVertex AI offers new ways to build and manage multi-agent systemsScale enterprise search and agent adoption with Google AgentspaceCloud WAN: Connect your global enterprise with a network built for the AI eraGemini 2.5 brings enhanced reasoning to enterprise use casesThe dawn of agentic AI in security operations at RSAC 2025Thought leadershipAI infrastructure is hot. New power distribution and liquid cooling infrastructure can help3 new ways to use AI as your security sidekickCustomer storiesGlobal startups are building the future of AI on Google CloudThe AI magic behind Sphere’s upcoming ‘The Wizard of Oz’ experience

MaySchool was almost out, but readers got back into learning mode to get certified as generative AI leaders. You were also excited about new gen AI media models in Vertex AI, the availability of Anthropic’s Claude Opus 4 and Claude Sonnet 4. We also learned that you’re very excited to use AI to generate SQL code, and about using Cloud Run as a destination for your AI apps. We outlined the steps for building a well-defined data strategy, and showed governments how AI can actually improve their security posture. And on the customer front, we launched our “Cool Stuff Customers Built” round-ups, and ran stories from Formula E and MLB.Google Cloud announces first-of-its-kind generative AI leader certificationExpanding Vertex AI with the next wave of generative AI media modelsAnnouncing Anthropic’s Claude Opus 4 and Claude Sonnet 4 on Vertex AIThought leadershipGetting AI to write good SQL: Text-to-SQL techniques explainedAI deployments made easy: Deploy to Cloud Run from AI Studio or any MCP clientBuilding a data strategy for the AI eraHow governments can use AI to improve threat detection and reduce costCustomer storiesCool Stuff Customers Built: May EditionPushing the limits of electric mobility: Formula E’s Mountain RechargeTuning in with AI: How MLB My Daily Story creates truly personalized highlight videos

JuneUp until this point, the promise of generative AI was largely around text and code. The launch of Veo 3 changed all that. Developers writing and deploying AI apps saw the availability of GPUs on Cloud Run as a big win, and we continued our steady drumbeat of Gemini innovation with 2.5 Flash and Flash-Lite. We also shared our thoughts on securing AI agents. And to learn how to actually build these agents, readers turned to stories about Box, the British real estate firm Schroders, and French luxury conglomerate LVMH (home of Louis Vuitton, Channel, Sephora and more).You dream it, Veo creates it: Veo 3 is now available for everyone in public preview on Vertex AICloud Run GPUs, now GA, makes running AI workloads easier for everyoneGemini momentum continues with launch of 2.5 Flash-Lite and general availability of 2.5 Flash and Pro on Vertex AIThought leadershipAsk OCTO: Making sense of AI agentsCloud CISO Perspectives: How Google secures AI agentsCustomer storiesThe secret to document intelligence: Box builds Enhanced Extract Agents with A2A frameworkHow Schroders built its multi-agent financial analysis research assistantInside LVMH’s perfectly manicured data estate, where luxury AI agents are taking root

JulyReaders took a break from reading about AI to read about network infrastructure — the new Sol transatlantic cable, to be precise. Then it was back to AI: new video generation models in Vertex; a crucial component for building stateful, context-aware agents; and a new toolset for connecting BigQuery data to Agent Development Kit (ADK) and Multi-Cloud Protocol (MCP) environments. Developers cheered the integration between Cloud Run and Docker Compose, and executive audiences enjoyed a listicle on actionable, real-world uses for AI agents.On the security front, we took a back-to-basics approach this month, exploring the persistence of some cloud security problems. And then, back to AI again, with our Big Sleep agent. Readers were also interested in how AI is alleviating record-keeping for nurses at HCA Healthcare, Ulta Beauty’s data warehousing and mobile record keeping initiatives, and how SmarterX migrated from Snowflake to BigQuery.Strengthening network resilience with the Sol transatlantic cableVeo 3 and Veo 3 Fast are now generally available on Vertex AIAnnouncing Vertex AI Agent Engine Memory Bank available for everyone in previewBigQuery meets ADK & MCP: Accelerate agent development with BigQuery’s new first-party toolsetFrom localhost to launch: Simplify AI app deployment with Cloud Run and Docker ComposeThought leadershipSecure cloud. Insecure use. (And what you can do about it)Our Big Sleep agent makes a big leapCustomer storiesHow nurses are charting the future of AI at America’s largest hospital network, HCA HealthcareUlta Beauty redefines beauty retail with BigQuerySmarterX’s migration from Snowflake to BigQuery accelerated model building and cut costs in half

AugustAI is compute- and energy-intensive; in a new technical paper, we released concrete numbers about our AI infrastructure’s power consumption. Then people went [nano] bananas for Gemini 2.5 Flash Image on Vertex AI, and developers got a jump on their AI projects with a wealth of technical blueprints to work from. The summer doldrums didn’t stop our security experts from tackling the serious challenge of cyber-enabled fraud. We also took a closer look at the specific agentic tools empowering workers at Wells Fargo, and how Keeta processes 11 million blockchain transactions per second with Spanner.How much energy does Google’s AI use? We did the mathBuilding next-gen visuals with Gemini 2.5 Flash Image (aka nano-banana) on Vertex AI101+ gen AI use cases with technical blueprintsThought leadershipNew Threat Horizons report details evolving risks — and defensesHow CISOs and boards of directors can help fight cyber-enabled fraudHow AI-powered weather forecasting can transform energy operationsCustomer storiesHow Wells Fargo is using Google Cloud AI to empower its workforce with agentic toolsHow Keeta processes 11 million financial transactions per second on the blockchain with Spanner

SeptemberAI is cool tech, but how do you monetize it? One answer is the Agent Payment Protocol, or AP2. Developers and data scientists preparing for AI flocked to blogs about new Data Cloud offerings, the 2025 DORA Report, and new trainings. Executives took in our thoughts on building an agentic data strategy, and took notes on the best prompts with which to kickstart their AI usage. And because everybody is impacted by the AI era, including business leaders, we explained what it means to be “bilingual” in AI and security. Then, at Google’s AI Builders Forum, startups described how Google’s AI, infrastructure, and services are supporting their growth. Not to be left out, enterprises like Target and Mr. Cooper also showed off their AI chops.Powering AI commerce with the new Agent Payments Protocol (AP2)The new data scientist: From analyst to agentic architectAnnouncing the 2025 DORA Report: State of AI-Assisted Software DevelopmentBack to AI school: New Google Cloud training to future-proof your AI skillsThought leadershipBuilding better data platforms, for AI and beyondBoards should be ‘bilingual’ in AI, security to gain advantageA leader’s guide to five essential AI promptsCustomer storiesHow Google Cloud’s AI tech stack powers today’s startupsFrom query to cart: Inside Target’s search bar overhaul with AlloyDB AIHow Mr. Cooper assembled a “team” of AI agents to handle complex mortgage questions

OctoberWelcome to the Gemini Enterprise era, which brings enhanced security, data control, and advanced agent capabilities to large organizations. To help you prepare, we relaunched a variety of enhancements to our learning platform, and added new commerce and security programs. And while developers versed themselves on the finer points of Veo prompts, we discussed securing the AI supply chain, building AI agents for cybersecurity and defense, and a new vision on economic threat modeling. We partnered with PayPal to enable commerce in AI chats, Germany’s Planck Institute showed how AI can help share deep scientific expertise, and DZ Bank pioneered ways to make blockchain-based finance more reliable.Introducing Gemini EnterpriseGoogle Skills: Your new home for cloud learningEnabling a safe agentic web with reCAPTCHAPartners powering the Gemini Enterprise agent ecosystemThought leadershipThe ultimate prompting guide for Veo 3.1How you can secure your AI supply chainHow Google Does It: Building AI agents for cybersecurity and defenseCustomer storiesIntroducing an agentic commerce solution for merchants from PayPal and Google CloudHow the Max Planck Institute is sharing expert skills through multimodal agentsThe oracles of DeFi: How DZ Bank builds trustworthy data feeds for decentralized applications

NovemberWhether it was Gemini 3, Nana Banana Pro, or our seventh-generation Ironwood TPUs, this was the month that we gave enterprise customers access to all our latest and greatest AI tech. We also did a deep dive on how we built the largest-ever Kubernetes cluster, clocking in at a massive 130,000 nodes, and we announced a new collaboration with AWS to improve connectivity between clouds.Meanwhile, we updated our findings on the adversarial misuse of AI by threat actors and on the ROI of AI for security, and executives vibed out on our piece about vibe coding. Then, just in time for the holidays, we took a look at how Mattel is using AI tools to revamp its toys, and Waze showed how it uses Memorystore to keep the holiday traffic flowing.Bringing Gemini 3 to EnterpriseHow Google Does It: Building the largest known Kubernetes cluster, with 130,000 nodesAnnouncing Nano Banana Pro for every builder and businessAnnouncing Ironwood TPUs General Availability and new Axion VMs to power the age of inferenceAWS and Google Cloud collaborate to simplify multicloud networkingThought leadershipRecent advances in how threat actors use AI toolsBeyond the hype: Analyzing new data on ROI of AI in securityHow vibe coding can help leaders move fasterCustomer storiesMattel’s game changer: How AI is turning customer feedback into real-time product updatesWaze keeps traffic flowing with 1M+ real-time reads per second on Memorystore

DecemberThe year is winding down, but we still have lots to say. Early returns show that you were interested in how to mitigate the React2Shell vulnerability, support for MCP across Google services, and the early access launch of AlphaEvolve. And let’s not forget Gemini 3 Flash, which is turning heads with its high-level reasoning, plus amazing speed and a flexible cost profile.What does this all mean for you and your future? It’s important to contextualize these technology developments, especially AI. For example, the DORA team put together a guide on how high-performing platform teams can integrate AI capabilities into their workflows, we discussed what it looks like to have an AI-ready workforce, and our Office of the CISO colleagues put out their 2026 cybersecurity predictions. More to the point (guard), you could do like Golden State Warrior Stephen Curry and turn to Gemini to analyze your game, to prepare for the year ahead. We’ll be watching on Christmas Day to see how Steph is faring with Gemini’s advice.Responding to React2Shell (CVE-2025-55182): Secure your React and Next.js workloadsAnnouncing Model Context Protocol (MCP) support for Google servicesAlphaEvolve on Google Cloud: AI for agentic discovery and optimizationIntroducing Gemini 3 Flash: Intelligence and speed for enterprisesThought leadershipFrom adoption to impact: Putting the DORA AI Capabilities Model to workIs AI fluency the ingredient or the result of an AI-ready workforce?Our 2026 Cybersecurity Forecast reportCustomer storiesWhat Stephen Curry learned about his game from a custom Gemini agent

The Curry sibling rivalry is going strong

And that’s a wrap on 2025! Thanks for reading, and see you next year!
Quelle: Google Cloud Platform

Supporting Viksit Bharat: Announcing our newest AI investments in India

India’s developer community, vibrant startup ecosystem, and leading enterprises are embracing AI with incredible speed. To meet this moment for India, we are investing in powerful, locally-available tools in India that can help foster a diverse ecosystem, and ensure our platform delivers the controls you need for compliance and AI sovereignty.
Today, we’re announcing a significant expansion of our local AI hardware capacity for customers in India. This increase in local compute, powered by Google’s AI Hypercomputer architecture with the latest Trillium TPUs, will help more businesses and public sector organizations train and serve their most advanced Gemini models in India. 
By unblocking new opportunities for high-performance, low-latency AI applications we can help customers meet India’s data residency and sovereignty requirements.
Enabling models and control: AI tools built for India’s context
While infrastructure is the foundation for digital sovereignty, it also requires control over the data and the models built on it. We’re committed to bringing our latest AI advancements to India faster than ever, with the controls you need.
Our new services would enable you to build, tune, and deploy models that understand India’s unique business logic and rich cultural context.

Next-generation models, here in India: Earlier this year, Google Cloud made Gemini available to regulated Indian customers by deploying Gemini 2.5 Flash with local machine-learning processing support. Now, we’re opening early testing for our latest and most advanced Gemini models to Indian customers. We’re also committing to launching the most powerful Gemini models in India with full data residency support. This is a first for Google Cloud, and a direct response to help meet the needs of our Indian customers.

More AI capabilities, available locally: We’re providing additional consumption models and pre-built AI-powered applications tailored for local context by launching a suite of new capabilities with data residency support in India:

Batch support for Gemini 2.5 Flash: Now generally available, this allows organizations to run high-volume, non-real-time AI tasks at a lower cost, all in India.

Document AI: Now in preview, we’re providing local support to help Indian businesses automate document processing.

More local context in your AI: Grounding on Google Maps is a new capability to ground model responses in real time from Google Maps, ensuring AI applications can provide accurate, location-aware answers.

A sovereign AI ecosystem: Building for India, with India
The most durable and decisive factor for long-term digital sovereignty lies in cultivating the “human element” — the skilled talent and innovation ecosystem. A sovereign AI future depends on building a strong local ecosystem.
Our strategy is to support India’s ecosystem-led approach by investing in the researchers, developers, and startups who are building for India’s specific needs.
Collaboration with IIT Madras: Google Cloud and Google DeepMind are thrilled to collaborate with IIT Madras to support the launch of Indic Arena. Run independently by the renowned AI4Bharat center at IIT Madras, this platform will allow users from all over India to anonymously evaluate and rank AI models on tasks unique to India’s rich multilingual landscape. To support this initiative, we are providing cloud credits to power this critical, community-driven resource.
“At AI4Bharat, our mission is to build AI for India’s specific needs. A critical part of this is having a neutral, standardized benchmark to understand how models are performing across our many languages,” said Mitesh Khapra, associate professor, IIT Madras. “Indic Arena will be that platform. We are delighted to have Google Cloud’s support to provide the initial compute power to bring this independent, public-facing project to life for the entire Indian AI community.”
We encourage all developers, researchers, and organizations in India to explore the Indic Arena platform and contribute to building a more inclusive AI future.
We invite the entire Indian ecosystem, from startups and universities to government bodies and enterprises, to take advantage of this new, dedicated capacity for Gemini in Vertex AI and our sovereign-ready infrastructure to build the next generation of AI that is built by Indians, for Indians.
Quelle: Google Cloud Platform

How scientists can leverage AI agents using Gemini Enterprise, Gemini Code Assist, and Gemini CLI

Scientific inquiry has always been a journey of curiosity, meticulous effort, and groundbreaking discoveries. Today, that journey is being redefined, fueled by the incredible capabilities of AI. It’s moving beyond simply processing data to actively participating in every stage of discovery, and Google Cloud is at the forefront of this transformation, building the tools and platforms that make it possible. 
The sheer volume of data generated by modern research is immense, often too vast for human analysis alone. This is where AI steps in, not just as a tool, but as a collaborative force. We’re seeing powerful new models and AI agents assist with everything from identifying relevant literature and generating novel hypotheses to designing experiments, running simulations, and making sense of complex results. This collaboration doesn’t replace human intellect; it amplifies it, allowing researchers to explore more avenues, more quickly, and with greater precision. 
At Google Cloud, we’re bringing together high-performance computing (HPC) and advanced AI on a single, integrated platform. This means you can seamlessly move from running massive-scale simulations to applying sophisticated machine learning models, all in one environment. 
So, how can you leverage these capabilities to get to insights faster? The journey begins at the foundation of scientific inquiry: the hypothesis.
AI-enhanced scientific inquiry
Every great discovery starts with a powerful hypothesis. With millions of research papers published annually, identifying novel opportunities is a monumental task. To overcome this information overload, scientists can now turn to AI as a powerful research partner.
Our Deep Research agent tackles the first step: performing a comprehensive analysis of published literature to produce detailed reports on a given topic that would otherwise take months to compile. Building on that foundation, our Idea Generation agent then deploys an ensemble of AI collaborators to brainstorm, evaluate, propose, debate, and rank novel hypotheses. This powerful combination, available in Gemini Enterprise, transforms the initial phase of scientific inquiry, empowering researchers to augment their expertise and find connections they might otherwise miss.
Go from hypothesis to results, faster
Once a hypothesis is formed, the work of translating it into executable code begins. This is where AI coding assistants, such as Gemini Code Assist, excel. They automate the tedious tasks of writing analysis scripts and simulation models by generating code from natural language and providing real-time suggestions, dramatically speeding up the core development process. 
But modern research is more than just a single script; it’s a complete workflow of data, environments, and results managed from the command line. For this, Gemini CLI brings that same conversational power directly to your terminal. It acts as the ultimate workflow accelerator, allowing you to instantly synthesize research and generate hypotheses with simple commands, then seamlessly transition to experimentation by generating sophisticated analysis scripts, and debugging errors on the fly, all without ever breaking your focus. Gemini CLI can further accelerate your path to impact by transforming raw results into publication-ready text, generating the code for figures and tables, and refining your work for submission. 
This capability extends to automating the entire research environment. Beyond single commands, Gemini CLI can manage complex, multi-step processes like cloning a scientific application, installing its dependencies, and then building and testing it—all with a simple prompt, maximizing your productivity.
The new era of discovery: Your expertise, AI agents, and Google Cloud
The new era of scientific discovery is here. By embedding AI into every stage of the scientific process – from sparking the initial idea to accelerating the final analysis – Google Cloud provides a single, unified platform for discovery. This new era of AI-enhanced scientific inquiry is built on a robust, intelligent infrastructure that combines the strengths of HPC simulation and AI. This includes purpose-built solutions like our H4D VMs optimized for scientific simulations, alongside the latest A4 and A4X VMs, powered by the latest NVIDIA GPUs, and Google Cloud Managed Lustre, a parallel file system that eliminates storage bottlenecks and allows your HPC and AI workloads to create and analyze massive datasets simultaneously. We provide the power to streamline the entire process so you can focus on scientific creativity – and changing the world! 
Join the Google Cloud Advanced Computing Community to connect with other researchers, share best practices, and stay up to date on the latest advancements in AI for scientific and technical computing, or contact sales to get started today.
Quelle: Google Cloud Platform

Gemeinsam gegen Geldwäsche: Wie EuroDaT den sicheren Austausch sensibler Finanzdaten ermöglicht

Ein Beitrag von Dr. Alexander Alldridge, Geschäftsführer von EuroDaTGeldwäschebekämpfung ist Teamarbeit. Banken, Regierungen und Technologiepartner müssen eng zusammenarbeiten, um kriminelle Netzwerke effektiv aufzudecken. Diese Herausforderung ist im streng regulierten Finanzsektor besonders komplex: Wie funktioniert Datenabgleich, wenn die Daten, um die es geht, hochsensibel sind? In diesem Blogbeitrag erklärt Dr. Alexander Alldridge, Geschäftsführer von EuroDaT, welche Rolle ein Datentreuhänder dabei spielen kann – und wie EuroDaT mit Lösungen von Google Cloud eine skalierbare, DSGVO-konforme Infrastruktur für genau diesen Zweck aufgebaut hat.
Wenn eine Bank eine verdächtige Buchung bemerkt, beginnt ein sensibler Abstimmungsprozess. Um mögliche Geldflüsse nachzuverfolgen, bittet sie andere Banken um Informationen zu bestimmten Transaktionen oder Konten. Aktuell geschieht das meist telefonisch – nicht, weil es keine digitalen Alternativen gäbe, sondern weil die Weitergabe sensibler Finanzdaten wie IBANs oder Kontobewegungen nur unter sehr engen rechtlichen Vorgaben erlaubt ist.Das Hin und Her per Telefon ist nicht nur mühsam, sondern auch fehleranfällig. Deutlich schneller und sicherer wäre ein digitaler Datenabgleich, der nur berechtigten Stellen Zugriff auf genau die Informationen gibt, die sie im konkreten Verdachtsfall benötigen.Hier bei EuroDaT, einer Tochtergesellschaft des Landes Hessen, bieten wir genau das: Als Europas erster transaktionsbasierter Datentreuhänder ermöglichen wir einen kontrollierten, anlassbezogenen Austausch sensibler Finanzdaten, der vertrauliche Informationen schützt und alle gesetzlichen Vorgaben erfüllt.safeAML: Ein neuer Weg für den Datenaustausch im FinanzsektorMit safeAML haben wir in Zusammenarbeit mit der Commerzbank, der Deutschen Bank und N26 ein System entwickelt, das den Informationsaustausch zwischen Finanzinstituten digitalisiert. Statt aufwendig andere Institute abzutelefonieren, kann künftig jede Bank selbst die relevanten Daten von anderen Banken hinzuziehen, um auffällige Transaktionen besser einordnen zu können.Der Datenaustausch läuft dabei kontrolliert und datenschutzkonform ab: Die Daten werden pseudonymisiert verarbeitet und so weitergegeben, dass nur die anfragende Bank sie am Ende wieder zuordnen kann. Wir bei EuroDaT haben als Datentreuhänder zu keinem Zeitpunkt Zugriff auf personenbezogene Inhalte.

safeAML Anwendung

Höchste Sicherheits- und Compliance-Standards mit Google CloudsafeAML ist eine Cloud-native Anwendung, wird also vollständig in der Cloud entwickelt und betrieben. Dafür braucht es eine Infrastruktur, die nicht nur technisch leistungsfähig ist, sondern auch die strengen Vorgaben im Finanzsektor erfüllt – von der DSGVO bis zu branchenspezifischen Sicherheits- und Cyber-Resilienz-Anforderungen. Google Cloud bietet dafür eine starke Basis, weil das Google Cloud-Team technisch und vertraglich schon früh die passenden Grundlagen für solche sensiblen Anwendungsfälle gelegt hat. Für uns war das ein entscheidender Vorteil gegenüber anderen Anbietern.Unsere gesamte Infrastruktur ist auf Google Kubernetes Engine (GKE) aufgebaut. Darüber richten wir sichere, isolierte Umgebungen ein, in denen jede Anfrage nachvollziehbar und getrennt von anderen verarbeitet werden kann. Alle technischen Ressourcen, darunter auch unsere Virtual Private Clouds (VPCs), sind in der Google-Cloud-Umgebung über Infrastruktur als Code definiert. Das bedeutet: Die gesamte Infrastruktur von EuroDaT wird automatisiert und wiederholbar aufgebaut, inklusive der Regeln dafür, welche Daten wohin fließen dürfen.Diese transparente, einfach reproduzierbare Architektur hilft uns auch dabei, die strengen Compliance-Anforderungen im Finanzsektor zu erfüllen: Wir können jederzeit belegen, dass sicherheitsrelevante Vorgaben automatisch umgesetzt und überprüft werden.
Banken nutzen safeAML für schnellere VerdachtsprüfungsafeAML ist inzwischen bei den ersten deutschen Banken testweise im Einsatz, um verdächtige Transaktionen schneller und besser einordnen zu können. Anstatt wie gewohnt zum Telefon greifen zu müssen, können Ermittler*innen jetzt gezielt ergänzende Informationen von anderen Instituten einholen, ohne dabei sensible Daten offenzulegen.Das beschleunigt nicht nur die Prüfung, sondern reduziert auch Fehlalarme, die bisher viel Zeit und Kapazitäten gebunden haben. Die Meldung, ob ein Geldwäscheverdacht vorliegt, bleibt dabei weiterhin eine menschliche Einzelfallentscheidung, wie es das deutsche Recht verlangt.Dass Banken über safeAML erstmals kontrolliert Daten austauschen können, ist bereits ein großer Schritt für die Geldwäschebekämpfung in Deutschland. Wir stehen aber noch am Anfang: Jetzt geht es darum, mehr Banken einzubinden, die Vernetzung national und international auszuweiten und den Prozess so unkompliziert wie möglich zu machen. Denn je mehr Institute mitmachen, desto besser können wir ein vollständiges Bild verdächtiger Geldflüsse zeichnen. Die neue Datenbasis kann künftig auch dabei helfen, Verdachtsfälle besser einzuordnen und fundierter zu bewerten.
Nachhaltiger Datenschutz: Sicherer Austausch von ESG-DatenUnsere Lösung ist aber nicht auf den Finanzbereich beschränkt. Als Datentreuhänder können wir das Grundprinzip, sensible Daten nur gezielt und kontrolliert zwischen dazu berechtigten Parteien zugänglich zu machen, auch auf viele andere Bereiche übertragen. Wir arbeiten dabei immer mit Partnern zusammen, die ihre Anwendungsideen auf EuroDaT umsetzen, und bleiben als Datentreuhänder selbst neutral.

Leistungsangebot EuroDaT

Ein aktuelles Beispiel sind ESG-Daten: Nicht nur große Firmen, sondern auch kleine und mittlere Unternehmen stehen zunehmend unter Druck, Nachhaltigkeitskennzahlen offenzulegen – sei es wegen neuer gesetzlicher Vorgaben oder weil Geschäftspartner wie Banken und Versicherer sie einfordern.Gerade für kleinere Firmen ist es schwierig, diesen Anforderungen gerecht zu werden. Sie haben oft nicht die nötigen Strukturen oder Ressourcen, um ESG-Daten standardisiert bereitzustellen, und möchten sensible Informationen wie Verbrauchsdaten verständlicherweise auch nicht einfach öffentlich machen.Hier kommt EuroDaT ins Spiel: Wir sorgen als vertrauenswürdige Zwischenstelle dafür, dass Nachhaltigkeitsdaten sicher weitergegeben werden, ohne dass Unternehmen die Kontrolle darüber verlieren. Mit dem Deutschen Nachhaltigkeitskodex (DNK) führen wir aktuell Gespräche zu einer Lösung, die kleinen Firmen das Übermitteln von ESG-Daten an Banken, Versicherungen und Investor*innen über EuroDaT als Datentreuhänder erleichtern kann.
Forschung im Gesundheitssektor: Sensible Daten, sichere ErkenntnisseAuch im Gesundheitssektor sehen wir großes Potenzial für unsere Technologie. Hier geht es natürlich um besonders sensible Daten, die nur unter strengen Auflagen verarbeitet werden dürfen. Trotzdem gibt es viele Fälle, in denen Gesundheitsdaten zusammengeführt werden müssen – etwa für die Grundlagenforschung, die Ausgestaltung klinischer Studien und politische Entscheidungen.Im Auftrag der Bundesregierung hat die Unternehmensberatung d-fine jetzt gezeigt, wie Gesundheitsdaten mithilfe von EuroDaT genutzt werden können – etwa zur Analyse der Auswirkungen von Post-COVID auf die Erwerbstätigkeit. Dafür müssen diese Daten mit ebenfalls hochsensiblen Erwerbsdaten zusammengeführt werden, was durch EuroDaT möglich wird: Als Datentreuhänder stellen wir sicher, dass die Daten vertraulich bleiben und dennoch sinnvoll genutzt werden können.Datensouveränität als Schlüssel zur digitalen ZusammenarbeitWenn Daten nicht ohne Weiteres geteilt werden dürfen, hat das meist gute Gründe. Gerade im Finanzwesen oder im Gesundheitssektor sind Datenschutz und Vertraulichkeit nicht verhandelbar. Umso wichtiger ist, dass der Austausch dieser Daten, wenn er tatsächlich notwendig wird, rechtlich sicher und kontrolliert stattfinden kann.Als Datentreuhänder sorgen wir deshalb nicht nur für sicheren Datenaustausch in sensiblen Branchen, sondern stärken dabei auch die Datensouveränität aller Beteiligten. Gemeinsam mit Google Cloud verankern wir Datenschutz fest im Kern der digitalen Zusammenarbeit zwischen Unternehmen, Behörden und Forschungseinrichtungen.
Quelle: Google Cloud Platform

Gemeinsam gegen Geldwäsche: Wie EuroDaT den sicheren Austausch sensibler Finanzdaten ermöglicht

Ein Beitrag von Dr. Alexander Alldridge, Geschäftsführer von EuroDaTGeldwäschebekämpfung ist Teamarbeit. Banken, Regierungen und Technologiepartner müssen eng zusammenarbeiten, um kriminelle Netzwerke effektiv aufzudecken. Diese Herausforderung ist im streng regulierten Finanzsektor besonders komplex: Wie funktioniert Datenabgleich, wenn die Daten, um die es geht, hochsensibel sind? In diesem Blogbeitrag erklärt Dr. Alexander Alldridge, Geschäftsführer von EuroDaT, welche Rolle ein Datentreuhänder dabei spielen kann – und wie EuroDaT mit Lösungen von Google Cloud eine skalierbare, DSGVO-konforme Infrastruktur für genau diesen Zweck aufgebaut hat.
Wenn eine Bank eine verdächtige Buchung bemerkt, beginnt ein sensibler Abstimmungsprozess. Um mögliche Geldflüsse nachzuverfolgen, bittet sie andere Banken um Informationen zu bestimmten Transaktionen oder Konten. Aktuell geschieht das meist telefonisch – nicht, weil es keine digitalen Alternativen gäbe, sondern weil die Weitergabe sensibler Finanzdaten wie IBANs oder Kontobewegungen nur unter sehr engen rechtlichen Vorgaben erlaubt ist.Das Hin und Her per Telefon ist nicht nur mühsam, sondern auch fehleranfällig. Deutlich schneller und sicherer wäre ein digitaler Datenabgleich, der nur berechtigten Stellen Zugriff auf genau die Informationen gibt, die sie im konkreten Verdachtsfall benötigen.Hier bei EuroDaT, einer Tochtergesellschaft des Landes Hessen, bieten wir genau das: Als Europas erster transaktionsbasierter Datentreuhänder ermöglichen wir einen kontrollierten, anlassbezogenen Austausch sensibler Finanzdaten, der vertrauliche Informationen schützt und alle gesetzlichen Vorgaben erfüllt.safeAML: Ein neuer Weg für den Datenaustausch im FinanzsektorMit safeAML haben wir in Zusammenarbeit mit der Commerzbank, der Deutschen Bank und N26 ein System entwickelt, das den Informationsaustausch zwischen Finanzinstituten digitalisiert. Statt aufwendig andere Institute abzutelefonieren, kann künftig jede Bank selbst die relevanten Daten von anderen Banken hinzuziehen, um auffällige Transaktionen besser einordnen zu können.Der Datenaustausch läuft dabei kontrolliert und datenschutzkonform ab: Die Daten werden pseudonymisiert verarbeitet und so weitergegeben, dass nur die anfragende Bank sie am Ende wieder zuordnen kann. Wir bei EuroDaT haben als Datentreuhänder zu keinem Zeitpunkt Zugriff auf personenbezogene Inhalte.

safeAML Anwendung

Höchste Sicherheits- und Compliance-Standards mit Google CloudsafeAML ist eine Cloud-native Anwendung, wird also vollständig in der Cloud entwickelt und betrieben. Dafür braucht es eine Infrastruktur, die nicht nur technisch leistungsfähig ist, sondern auch die strengen Vorgaben im Finanzsektor erfüllt – von der DSGVO bis zu branchenspezifischen Sicherheits- und Cyber-Resilienz-Anforderungen. Google Cloud bietet dafür eine starke Basis, weil das Google Cloud-Team technisch und vertraglich schon früh die passenden Grundlagen für solche sensiblen Anwendungsfälle gelegt hat. Für uns war das ein entscheidender Vorteil gegenüber anderen Anbietern.Unsere gesamte Infrastruktur ist auf Google Kubernetes Engine (GKE) aufgebaut. Darüber richten wir sichere, isolierte Umgebungen ein, in denen jede Anfrage nachvollziehbar und getrennt von anderen verarbeitet werden kann. Alle technischen Ressourcen, darunter auch unsere Virtual Private Clouds (VPCs), sind in der Google-Cloud-Umgebung über Infrastruktur als Code definiert. Das bedeutet: Die gesamte Infrastruktur von EuroDaT wird automatisiert und wiederholbar aufgebaut, inklusive der Regeln dafür, welche Daten wohin fließen dürfen.Diese transparente, einfach reproduzierbare Architektur hilft uns auch dabei, die strengen Compliance-Anforderungen im Finanzsektor zu erfüllen: Wir können jederzeit belegen, dass sicherheitsrelevante Vorgaben automatisch umgesetzt und überprüft werden.
Banken nutzen safeAML für schnellere VerdachtsprüfungsafeAML ist inzwischen bei den ersten deutschen Banken testweise im Einsatz, um verdächtige Transaktionen schneller und besser einordnen zu können. Anstatt wie gewohnt zum Telefon greifen zu müssen, können Ermittler*innen jetzt gezielt ergänzende Informationen von anderen Instituten einholen, ohne dabei sensible Daten offenzulegen.Das beschleunigt nicht nur die Prüfung, sondern reduziert auch Fehlalarme, die bisher viel Zeit und Kapazitäten gebunden haben. Die Meldung, ob ein Geldwäscheverdacht vorliegt, bleibt dabei weiterhin eine menschliche Einzelfallentscheidung, wie es das deutsche Recht verlangt.Dass Banken über safeAML erstmals kontrolliert Daten austauschen können, ist bereits ein großer Schritt für die Geldwäschebekämpfung in Deutschland. Wir stehen aber noch am Anfang: Jetzt geht es darum, mehr Banken einzubinden, die Vernetzung national und international auszuweiten und den Prozess so unkompliziert wie möglich zu machen. Denn je mehr Institute mitmachen, desto besser können wir ein vollständiges Bild verdächtiger Geldflüsse zeichnen. Die neue Datenbasis kann künftig auch dabei helfen, Verdachtsfälle besser einzuordnen und fundierter zu bewerten.
Nachhaltiger Datenschutz: Sicherer Austausch von ESG-DatenUnsere Lösung ist aber nicht auf den Finanzbereich beschränkt. Als Datentreuhänder können wir das Grundprinzip, sensible Daten nur gezielt und kontrolliert zwischen dazu berechtigten Parteien zugänglich zu machen, auch auf viele andere Bereiche übertragen. Wir arbeiten dabei immer mit Partnern zusammen, die ihre Anwendungsideen auf EuroDaT umsetzen, und bleiben als Datentreuhänder selbst neutral.

Leistungsangebot EuroDaT

Ein aktuelles Beispiel sind ESG-Daten: Nicht nur große Firmen, sondern auch kleine und mittlere Unternehmen stehen zunehmend unter Druck, Nachhaltigkeitskennzahlen offenzulegen – sei es wegen neuer gesetzlicher Vorgaben oder weil Geschäftspartner wie Banken und Versicherer sie einfordern.Gerade für kleinere Firmen ist es schwierig, diesen Anforderungen gerecht zu werden. Sie haben oft nicht die nötigen Strukturen oder Ressourcen, um ESG-Daten standardisiert bereitzustellen, und möchten sensible Informationen wie Verbrauchsdaten verständlicherweise auch nicht einfach öffentlich machen.Hier kommt EuroDaT ins Spiel: Wir sorgen als vertrauenswürdige Zwischenstelle dafür, dass Nachhaltigkeitsdaten sicher weitergegeben werden, ohne dass Unternehmen die Kontrolle darüber verlieren. Mit dem Deutschen Nachhaltigkeitskodex (DNK) führen wir aktuell Gespräche zu einer Lösung, die kleinen Firmen das Übermitteln von ESG-Daten an Banken, Versicherungen und Investor*innen über EuroDaT als Datentreuhänder erleichtern kann.
Forschung im Gesundheitssektor: Sensible Daten, sichere ErkenntnisseAuch im Gesundheitssektor sehen wir großes Potenzial für unsere Technologie. Hier geht es natürlich um besonders sensible Daten, die nur unter strengen Auflagen verarbeitet werden dürfen. Trotzdem gibt es viele Fälle, in denen Gesundheitsdaten zusammengeführt werden müssen – etwa für die Grundlagenforschung, die Ausgestaltung klinischer Studien und politische Entscheidungen.Im Auftrag der Bundesregierung hat die Unternehmensberatung d-fine jetzt gezeigt, wie Gesundheitsdaten mithilfe von EuroDaT genutzt werden können – etwa zur Analyse der Auswirkungen von Post-COVID auf die Erwerbstätigkeit. Dafür müssen diese Daten mit ebenfalls hochsensiblen Erwerbsdaten zusammengeführt werden, was durch EuroDaT möglich wird: Als Datentreuhänder stellen wir sicher, dass die Daten vertraulich bleiben und dennoch sinnvoll genutzt werden können.Datensouveränität als Schlüssel zur digitalen ZusammenarbeitWenn Daten nicht ohne Weiteres geteilt werden dürfen, hat das meist gute Gründe. Gerade im Finanzwesen oder im Gesundheitssektor sind Datenschutz und Vertraulichkeit nicht verhandelbar. Umso wichtiger ist, dass der Austausch dieser Daten, wenn er tatsächlich notwendig wird, rechtlich sicher und kontrolliert stattfinden kann.Als Datentreuhänder sorgen wir deshalb nicht nur für sicheren Datenaustausch in sensiblen Branchen, sondern stärken dabei auch die Datensouveränität aller Beteiligten. Gemeinsam mit Google Cloud verankern wir Datenschutz fest im Kern der digitalen Zusammenarbeit zwischen Unternehmen, Behörden und Forschungseinrichtungen.
Quelle: Google Cloud Platform

Top 25 blogs of 2025… so far

Six months into 2025, we’ve already published hundreds of posts here on the Google Cloud blog. We asked ourselves, why wait until the busy end of the year to review your favorites? With everything from new AI models, product launches, emerging cyber threats, company news, certifications and customer stories, here is a mid-year recap that will get you up to speed on the latest from Google Cloud and the rapidly emerging cloud and AI landscape. 
25. How Google Does It: Making threat detection high-quality, scalable, and modern
Published January 7, 2025
Google and Alphabet run the largest Linux fleet in the world, with nearly every flavor of operating system available, and see a steady stream of malicious system and network activity. Learn how our threat detection and response team detects, analyzes, and responds to threats on a vast scale.   
Read the blog. 
24. Cloud Run GPUs are now generally available
Published June 2, 2025
More and more organizations are turning to Cloud Run, Google Cloud’s serverless runtime, for its simplicity, flexibility, and scalability. And now, with the general availability of NVIDIA GPUs on the platform, developers can choose Cloud Run for applications that require powerful graphics processing, like machine learning models.
Read the blog. 
23. BigQuery emerges as autonomous data-to-AI platform
Published April 10, 2025
This is not your grandfather’s data warehouse. BigQuery is now an AI-native, multimodal, and agentic data-to-AI platform. The blog post provides an overview of the many new features and capabilities that went into this new designation, including new data preparation, data analysis, code generation and management and troubleshooting capabilities. 
Read the blog. 
22. Announcing Gen AI Toolbox for Databases. Get started today
Published February 6, 2025
Tired of building custom plumbing to connect your AI apps to your databases? This article announces the public beta of the Gen AI Toolbox for Databases, an open-source server built with LangChain that provides a secure, scalable, and manageable way to connect your generative AI applications to your data.
Read the blog. 
21. Ghost in the router: China-nexus espionage actor UNC3886 targets Juniper Networks
Published March 11, 2025
After discovering in 2024 that threat actors deployed custom backdoors to Juniper Networks’ Junos OS routers, Mandiant worked with Juniper to investigate this activity and observed that the affected routers were running end-of-life hardware and software. Learn more about the threat and how to remediate it in your environment. 
Read the blog.
20. What’s new with AI Hypercomputer?
Published April 9, 2025
It’s a platform, it’s a system, it’s AI Hypercomputer, Google Cloud’s fully managed supercomputing system for running AI and HPC workloads. As discussed at Google Cloud Next 2025, AI Hypercomputer supports all the latest and greatest compute, networking and storage infrastructure, and its software layer helps AI practitioners and engineers move faster with open and popular ML frameworks. Finally, there’s a full suite of workload management and observability tools to help you manage the thing.
Read the blog. 
19. Ipsos research shows why cloud certification matters — get certified with Google Cloud

Published February 25, 2025Google Cloud partnered with Ipsos, the global research firm, to study the impact of cloud certifications on career advancement and achievement. For example, 8 out of 10 survey respondents said earning a recognized certificate helped them land a job faster and 75% believe they secured a higher salary through their certification.Read the blog.

18. Connect globally with Cloud WAN for the AI Era
Published April 9, 2025
With 202 points of presence (PoPs), powered by over 2 million miles of fiber, 33 subsea cables, and backed by a 99.99% reliability SLA, Google’s backbone network is, how do we put it? Vast. And with Cloud WAN, enterprises can now use it for their own wide area network (WAN) architectures. 
Read the blog. 
17. Expanding generative media for enterprise on Vertex AI

Published April 9, 2025At Google Cloud Next 25, we announced powerful new creative controls for our generative media models on Vertex AI. Now you can edit video with in-painting and out-painting, use camera controls for dynamic shots, and even create custom voices for AI-powered narration with as little as 10 seconds of audio.Read the blog.

16. Suspected China-nexus threat actor actively exploiting critical Ivanti Connect Secure vulnerability
Published April 3, 2025
Threat actors continue to target edge devices globally, leveraging deep device knowledge and using both zero-day and now n-day flaws. This activity aligns with the broader strategy that the Google Threat Intelligence Group has observed among suspected China-nexus espionage groups, who invest significantly in exploits and custom malware for critical edge infrastructure.
Read the blog. 
15. Defending against UNC3944: Cybercrime hardening guidance from the frontlines 
Published May 6, 2025
Who is UNC3944? A financially-motivated threat actor characterized by its persistent use of social engineering and brazen communications with victims. Mandiant provides guidance and strategies for hardening systems and defenses against the cybercrime group, offering practical steps to protect against their specific attack methods.
Read the blog. 
14. MCP Toolbox for Databases (formerly Gen AI Toolbox for Databases)
Published April 22, 2025
Ready to build AI agents that can actually use your data? This article announces that our MCP Toolbox for Databases now supports the Model Context Protocol (MCP), making it easier than ever to connect your generative AI agents to enterprise data. With new support for the Agent Development Kit (ADK) and LangGraph, you can build powerful, stateful agents with intuitive code and connect them to your databases securely.
Read the blog.
13. Formula E’s AI equation: A new Driver Agent for the next era of racing

Published March 25, 2025As motorsport has grown in popularity, the ability of fans from diverse backgrounds to enter the cockpit has not always kept up. Formula E sought to level the course for aspiring drivers by creating an AI-powered Driver Agent; connected to a Formula E simulator, the agent provides drivers and coaches with real-time feedback on technique and tactics, help them improve faster than a flying lap.Read the blog.

12. Google Agentspace enables the agent-driven enterprise
Published April 9, 2025
Do you want to search all your company’s information in a few clicks, or generate ideas with built-in agents that already know your company’s style? Google Agentspace now includes a no-code agent designer, a gallery for discovering agents, and two new expert agents for deep research and idea generation, all integrated directly into Chrome.
Read the blog.
11. Announcing Veo 3, Imagen 4, and Lyria 2 on Vertex AI

Published May 20, 2025The next generation of creating for enterprise is here. We expanded Vertex AI to include our most powerful generative AI media models: Imagen 4 for stunningly realistic images with crisp text, Veo 3 for breathtaking video with synchronized audio, and Lyria 2 for composing high-fidelity, original music.Read the blog.

10. Adversarial misuse of generative AI
Published January 19, 2025
In the security realm, large language models (LLMs) open a world of new possibilities, from sifting through complex telemetry to secure coding, vulnerability discovery, and streamlining operations. However, some of these same AI capabilities are also available to attackers, leading to understandable anxieties about the potential for AI to be misused for malicious purposes.
Read the blog.
9. Ivanti Connect Secure VPN targeted in new zero-day exploitation
Published January 8, 2025
Ivanti kicked off the year by disclosing two new vulnerabilities impacting its Ivanti Connect Secure (ICS) VPN appliances. Mandiant identified UNC5221, a suspected China-nexus espionage actor that previously exploited two other Ivanti vulnerabilities as early as December 2023, as the threat actor targeting the new zero-days. Successfully exploiting one of the vulnerabilities could result in downstream compromise of a victim network.
Read the blog. 
8. Google announces agreement to acquire Wiz
Published March 18, 2025
Google Cloud shares a vision with Wiz to improve security by making it easier and faster for organizations of all types and sizes to protect themselves, end-to-end, across all major clouds, and this post announces Google’s agreement to acquire the cloud security startup.
Read the blog.
7. Veo 3 available for everyone in preview on Vertex AI
Published June 26, 2025
You dream it, Veo creates it. This post announces Veo 3, our most powerful text-to-video model yet, is now open for everyone to try in public preview on Vertex AI. Create stunning, near-cinematic videos with synchronized sound, and join the next wave of creative storytelling, now available to Google Cloud customers and partners.
Read the blog. 
6. Vertex AI offers new ways to build and manage multi-agent systems

Published April 9, 2025This article announces ways to build multi-agentic systems, an evolution of traditional AI agents. To get there, we launched a new suite of tools in Vertex AI to help developers build and deploy them, including an open-source Agent Development Kit (ADK) and a managed Agent Engine. We also introduce the Agent2Agent (A2A) protocol, a new open standard to allow agents built by different companies to communicate and collaborate.Read the blog.

5. Techniques for improving text-to-SQL
Published May 16, 2025
Even though it’s been around for a long time, not all developers speak fluent SQL. English, on the other hand, is pretty well-known. In this technical deep dive for developers working with natural language processing and databases, get the insights and techniques you need to enhance the accuracy and performance of your text-to-SQL conversions.
Read the blog.
4. Firebase Studio lets you build full-stack AI apps with Gemini
Published April 9, 2025
For over a decade, developers the world over have relied on Firebase’s backend cloud computing services and application development platforms to power their web applications. And with the new Firebase Studio, they can now use it to develop full-stack AI applications, integrating with the Gemini AI model.
Read the blog.  
3. Multiple Russia-aligned threat actors targeting Signal Messenger
Published February 19, 2025
As part of the ongoing Russian-Ukrainian conflict, Signal Messenger accounts are of great interest to Russia’s intelligence services for their potential to deliver sensitive government and military communications. Google Threat Intelligence Group has observed increasing efforts from several Russia state-aligned threat actors to compromise Signal Messenger accounts used by individuals of interest to Russia’s intelligence services.
Read the blog.
2. New Google Cloud certification in generative AI
One of the top questions we hear is “how do I get ahead”? This isn’t just another certification in a sea of technical qualifications. The Generative AI Leader certification is specifically focused on generative AI, and designed for visionary professionals like you — the managers, administrators, strategic leaders and more who understand that AI’s impact stretches far beyond code.
Read the blog.
1. 601 real-world gen AI use cases from the world’s leading organizations

Published April 9, 2025Since Next 2024, we’ve been gathering examples of how our customers are putting generative AI to use everyday across their operations and offerings. We nearly doubled the number of entries for Next 2025, and clearly they’re still resonating, as this has been our most popular story of the year. What use cases are most exciting you? Pop over to our LinkedIn page and let us know.Read the blog.

Thank you for being a part of the Google Cloud blog community! We look forward to bringing you lots more blogs for you to devour in the second half of the year.
Quelle: Google Cloud Platform

News you can use: What we announced in AI this month

2025 is off to a racing start. From announcing strides in the new Gemini 2.0 model family to retailers accelerating with Cloud AI, we spent January investing in our partner ecosystem, open-source, and ways to make AI more useful. We’ve heard from people everywhere, from developers to CMOs, about the pressure to adapt the latest in AI with efficiency and speed – and the delicate balance of being both conservative and forward-thinking. We’re here to help. Each month, we’ll post a retrospective that recaps Google Cloud’s latest announcements in AI – and importantly, how to make the most of these innovations. 
Top announcements: Bringing AI to you 
This month, we announced agent evaluation in Vertex AI. A surprise to nobody, AI agents are top of mind for many industries looking to deploy their AI and boost productivity. But closing the gap between impressive model demos and real-world performance is crucial for successfully deploying generative AI. That’s why we announced Vertex AI’s RAG Engine, a fully managed service that helps you build and deploy RAG implementations with your data and methods. Together, these new innovations can help you build reliable, trustworthy models.
From an infrastructure perspective, we announced new updates to AI Hypercomputer. We wanted to make it easier for you to run large multi-node workloads on GPUs by launching A3 Ultra VMs and Hypercompute Cluster, our new highly scalable clustering system. This builds on multiple advancements in AI infrastructure, including Trillium, our sixth-generation TPU.

aside_block
<ListValue: [StructValue([('title', '$300 in free credit to try Google Cloud AI and ML'), ('body', <wagtail.rich_text.RichText object at 0x3eec5a0d4610>), ('btn_text', 'Start building for free'), ('href', 'http://console.cloud.google.com/freetrial?redirectPath=/vertex-ai/'), ('image', None)])]>

What’s new in partners and open-source 
This month, we invested in our relationship with our partners. We shared how Gemini-powered content creation in Partner Marketing Studio will help partners co-market faster. These features are designed to streamline marketing efforts across our entire ecosystem, empowering our partners to unlock new levels of success, efficiency, and impact. 
At the same time, we shared several important announcements in the world of open-source. We announced Mistral AI’s Mistral Large 24.11 and Codestral 25.01 models on Vertex AI. These models will help developers write code and build faster – from high-complexity tasks to reasoning tasks, like creative writing. To help you get started, we provided sample code and documentation.
And, most recently, we announced the public beta of Gen AI Toolbox for Databases in partnership with LangChain, the leading orchestration framework for developers building LLM applications. Toolbox is an open-source server that empowers application developers to connect production-grade, agent-based generative AI applications to databases. You can get started here.
Industry news: Google Cloud at the National Retail Federation (NRF) 
The National Retail Federation kicked off the year with their annual NRF conference, where Google Cloud showed how AI agents and AI-powered search are already helping retailers operate more efficiently, create personalized shopping experiences, and use AI to get the latest products and experiences to their customers. Check our new AI tools to help retailers build gen AI search and agents. 
As an example, Google Cloud worked with NVIDIA to empower retailers to boost their customer engagements in exciting new ways, deliver more hyper-personalized recommendations, and build their own AI applications and agents. Now with NVIDIA’s AI Enterprise software available on Google Cloud, retailers can handle more data and more complex AI tasks without their systems getting bogged down.
News you can use 
This month, we shared several ways to better implement fast-moving AI, from a comprehensive guide on Supervised Fine Tuning (SFT), to how developers can help their LLMs deliver more accurate, relevant, and contextually aware responses, minimizing hallucinations and building trust in AI applications by optimizing their RAG retrieval.
We also published new documentation to use open models in Vertex AI Studio. Model selection isn’t limited to Google’s Gemini anymore. Now, choose models from Anthropic, Meta, and more when writing or comparing prompts.
Hear from our leaders
We closed out the month with The Prompt, our monthly column that brings observations from the field of AI. This month, we heard from Warren Barkley, AI product leader, who shares some best practices and essential guidance to help organizations successfully move AI pilots to production. Here’s a snippet:
More than 60% of enterprises are now actively using gen AI in production, helping to boost productivity and business growth, bolster security, and improve user experiences. In the last year alone, we witnessed a staggering 36x increase in Gemini API usage and a nearly 5x increase of Imagen API usage on Vertex AI — clear evidence that our customers are making the move towards bringing gen AI to their real-world applications.
Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.
Quelle: Google Cloud Platform

What’s new with Google Cloud – 2024

Week of Dec 16 – Dec 20Windows Server 2025 is now available on Google Compute Engine. We are excited to announce the general availability of Windows Server 2025 on Google Compute Engine. You can now run Windows Server 2025 Data Center, and Windows Server 2025 Data Center Core editions as well as Windows SQL Server 2022 on Windows Server 2025 with pay-as-you-go licenses. Customers that don’t already have a Microsoft Enterprise Agreement can use the Google Compute Engine’s provided image to take advantage of Google Cloud’s relationships with Microsoft for pay-as-you-go licenses that scale with your workload and offers premium support.Google Agentspace is here: unlock enterprise expertise for employees with agents that bring together Gemini’s advanced reasoning, Google-quality search, and enterprise data, regardless of where it’s hosted. Google Agentspace makes your employees highly productive by helping them accomplish complex tasks that require planning, research, content generation, and actions – all with a single prompt.Week of Dec 9 – Dec 13Best of N: Generating High-Quality Grounded Answers with Multiple Drafts. We are excited to announce that Check Grounding API has released a new helpfulness score feature. Building on top of our existing groundedness score, we now enable users to implement Best of N to improve RAG response quality without requiring extensive model retraining!A3 Ultra VMs powered by NVIDIA H200 Tensor Core GPUs and Hypercompute Clusters are in preview. A3 Ultra VMs offer a significant leap in performance over previous generations. Coupled with Hypercompute Cluster, which the infrastructure and workload provisioning, and ongoing operations of AI supercomputers up to tens of thousands of accelerators is seamless. This delivers a easy to manage, secure and high-performance cloud experience for AI workloads. Learn more about the offering!Week of Nov 25 – Nov 29The Cyber Threat Intelligence Program Design Playbook, Now Available. Mandiant Academy published a new on-demand course, Cyber Threat Intelligence (CTI) Program Design Playbook. Comprised of three 2-hour courses, this learning track explains Mandiant’s approach to design and build, efficiently operate, and enhance a CTI program. This is the inaugural track from the Academy’s updated approach to on-demand learning with succinct, operations-oriented lessons designed to provide the world’s cybersecurity professionals the answers and resources they need to succeed.Week of Nov 11 – Nov 15Subsea cable connectivity is coming to Tuvalu for the first time with the addition of the Tuvalu Vaka cable. Building on the Bulikula subsea cable system announced last year, this new network infrastructure is a collaboration among several partners including Australia, Japan, New Zealand, Taiwan, Tuvalu, Tuvalu Telecommunications Corporation and the United States, and will help reduce the digital divide in the Pacific.Week of Nov 4-8We are excited to announce the reCAPTCHA Password Leak Detection Container App, a new tool that makes it easier than ever to protect your users from account takeovers. This container app simplifies the integration of reCAPTCHA’s powerful password leak detection, allowing you to instantly detect compromised credentials and proactively prompt users to change their password before their account is compromised. With pre-built libraries and a streamlined process, you can significantly reduce integration time and enhance your website’s security with ease.Week of Oct 21-25We’re excited to announce GA support for scanning: Rocky Linux, Alma, SUSE (SLES), Red Hat (UBI), Chainguard, Wolfi & Google Distroless. These operating systems are now supported in both Artifact Registry scanning, as well as On Demand Scanning. When the Container Scanning API is enabled, any container with these new operating systems or distroless images will automatically be scanned for vulnerabilities when pushed to Artifact Registry. We’ve also upgraded our On Demand scan to include; NPM, Python, Ruby, Rust, .Net & PHP language packages. See all supported package types.Term Extension Now Available for Compute Engine Committed Use Discounts: You can now extend the term length of your Compute Engine resource-based Committed Use Discounts (“CUDs”) beyond the preset 1-year and 3-year options. CUDs offer significant cost savings for predictable workloads. You can now choose a CUDs term length beyond the original commitment end date that perfectly aligns with your workload needs, from one year and one day up to 6 years. Learn moreWeek of Oct 14-18Announcing Google Cloud Marketplace private offer enhancements that enable additional payment flexibility for enterprises, including when transacting generative AI models.Week of Oct 7-11We are excited to announce the launch of new Google Cloud Cortex Framework data integration and analytics solution content for BigQuery and Looker with Oracle EBS data. To learn more read our announcement blog.Google Cloud is partnering with leading AI and cybersecurity startups to accelerate their growth and innovation, through the ISV Startup Springboard program, announced this week at the Google Cloud Startup Summit. Learn more and register interest.Privileged Access Manager (PAM) is now Generally Available. The GA release offers new capabilities in addition to recently released public preview and includes features such as Pub/Sub integration for custom alerting and monitoring, alerts on IAM grant modifications outside of PAM, and integration with VPC Service Controls to tackle data exfiltration. Learn more.Week of Sept 23-27We are excited to announce that registration is open for the App Dev & Infrastructure Summit on October 30 (AMER) and October 31 (EMEA). Google Technology Fellows – our luminary technical leaders – and industry experts will share strategies and learnings on how to improve efficiency, reduce costs, and speed up AI innovation for your cloud and application infrastructure at this global digital event. Register here.Week of Sept 16-20Starting this week, Google Cloud customers with eligible support plans can access assistance for the Cluster Toolkit through the Cloud Console. Cluster Toolkit, formerly known as Cloud HPC Toolkit, is open-source software offered by Google Cloud which simplifies the process for you to deploy HPC, AI and ML workloads on Google Cloud. The Cloud Support team will handle filed cases, ensuring that users receive timely and effective support for their Cluster Toolkit implementations. Select ‘Cluster Toolkit’ as the sub-category under ‘Compute Engine’ when creating a support ticket in your Cloud Console to get in touch about any Cluster Toolkit issues.Backup and DR service is excited to announce the public preview of backup vaults and simplified VM backup offering. Backup vaults provide secure backups for cyber resilience through immutable and indelible backups for VMs and databases, delivering security against accidental or malicious data deletion. Simplified Compute Engine VM backup with a fully-managed experience, directly integrated into the cloud console makes backing of VMs as easy as 1-2-3. The solution also enables backup admins to empower application developers to self-protect their VMs while retaining centralized governance and oversight. Read the full blog to learn more and try out the new features.Week of Sept 2-6We’re excited to share that Topaz will be extended to Taiwan. Announced in 2022, the transpacific subsea cable system was the first to connect Canada and Japan. Now, with the extension of Topaz to Taiwan, we’ll provide the region with increased reliability and resilience for network operators, for Google, and for users.Week of Aug 26-30We are excited to announce the general availability of Instant snapshots for Google Compute Engine Persistent Disks, which provide near-instantaneous, high-frequency, point-in-time checkpoints of a disk that can be rapidly restored as needed. Read the full blog to try it out.In response to customer and partner requests for pollen data in Japan, we are excited to announce that data for Japanese Cedar and Cypress trees-the 2 main sources of pollen allergens in Japan-have been added to our Pollen API from Google Maps Platform.Week of Aug 19-23We are excited to announce we’re adding support for NVIDIA L4 GPUs to Cloud Run, in preview. Developers love Cloud Run for its simplicity, fast autoscaling, scale-to-zero capabilities, and pay-per-use pricing. Those same benefits come into play for real-time inference apps serving open gen AI models. Check out this launch blog. Also watch demos from this launch event webinar Run AI on Cloud run.We are excited to announce that Google Cloud Functions is now Cloud Run functions — event-driven programming in one unified serverless platform. This goes beyond a simple name change. We’ve unified the Cloud Functions infrastructure with Cloud Run, and developers of Cloud Functions (2nd gen) get immediate access to all new Cloud Run features, including NVIDIA GPUs. Read the launch blog and watch demos from this launch event webinar Run AI on Cloud run.Week of Aug 5-9Google’s Workforce Identity federation now enables Microsoft Entra ID users to access Google BigQuery from Microsoft Power BI with Single-Sign-On. No users or groups need to be provisioned in Google Cloud as Workforce identity Federation leverages a syncless federation capability using attribute based access control to authorize access to Google BigQuery using Microsoft Entra user attributes such as user group membership. You can refer to our documentation to learn more.We are excited to announce the preview of SQL support in Bigtable to bring Google’s pioneering NoSQL database to a broader developer audience. Bigtable leverages GoogleSQL ─the same SQL dialect used by BigQuery─ making it easier to use Bigtable as low-latency analytics serving layer in combination with BigQuery’s newly announced continuous queries but does so with extensions to support its signature data model so you can use SQL without giving up on all the flexibility that comes with a NoSQL database. It also simplifies migrations from open source databases such as Apache Cassandra. With over 100 new functions from JSON processing capabilities, kNN for GenAI and HLL for real-time analytics, SQL opens the door to many new possibilities with Bigtable. Learn more in our detailed blog post.We are excited to announce the public preview of BigQuery continuous queries, a groundbreaking new feature that empowers users to run continuously processing SQL statements that can process, analyze, and transform data as new events arrive in BigQuery, ensuring insights are always up to date. Native integration with the Google Cloud ecosystem unlocks the ability of Vertex AI and Gemini to perform machine learning inference on incoming data in real time. As well as streaming replication of continuous query results to Pub/Sub topics, Bigtable instances, or other BigQuery tables. Read the full blog and try it out!AlloyDB’s AutoPilot capabilities- Automatic memory management, Adaptive AutoVacuum, Automatic storage tiering ,Automatic data columnarization and query rewrite- makes management super efficient and easy. AlloyDB eliminates the drudgery of maintaining a PostgreSQL database by using, behind the scenes , advanced self-tuning machine learning algorithm. In this blog we will look into a real world example of AlloyDB Adaptive AutoVacuum in work and how AlloyDB Cluster Storage Space is ReleasedGoogle Cloud Identity Platform, our consumer identity solution, now supports Passkeys. With Passkeys, developers can authenticate their app’s end users securely, protecting them from account takeover attacks like phishing and leaked credentials. To join the private preview, contact your Google account team.Week of July 15-19Google Cloud is excited to launch the Modern SecOps Masterclass, now available on Coursera. This course equips security professionals with cutting-edge skills to modernize their Security Operations Centers (SOCs) using our Autonomic Security Operations framework and Continuous Detection, Continuous Response (CD/CR) methodology. Read the full blog and enroll now.Learn how to potentially achieve a strong consistency in Cloud Bigtable for your next big data solution. Bigtable offers high throughput at low latency. It is ideal for storing large amounts of data in a key-value store while supporting high read and write throughput at low latency for fast access. Bigtable provides eventual consistency as well as strong consistency. This blog talks about achieving strong data consistency in a multi-cluster Bigtable instance. Read the full blog.Week of June 24-28Introducing Google Cloud Marketplace Channel Private Offers, enabling customers, ISV partners, and channel partners to efficiently transact private offers via reseller-initiated sales of third-party solutions listed on the Google Cloud Marketplace. This differentiated program also empowers channel partners to manage the customer relationship from billing, collections to revenue recognition. Read the full blog.A blog on benchmark study (collaborated with Yahoo) by comparing the cost and performance of Apache Flink and Google Cloud Dataflow for two specific streaming data processing use cases. The goal of the study was to determine the most cost-effective platform for these use cases by establishing a fair comparison methodology and controlling variables such as throughput and workload. The results indicate that, with some optimization on Dataflow can perform on-par with Apache Flink. Read the full blog.A Blog on Secure Gateways: Mutual TLS for Ingress Gateway Secure Gateways: Mutual TLS for Ingress Gateway,” discusses the implementation of mutual TLS (mTLS) for enhanced security in ingress gateways. It explains how mTLS ensures both client and server authentication through certificates, going beyond the traditional server-only verification. The article explores the setup process and the benefits of using mTLS, emphasizing its role in establishing secure communication channels in modern cloud architectures. Read the full blog.A Blog on Wildcard certificates with Ingress Gateway “Wildcard certificates with Ingress Gateway” provides a guide on how to use wildcard certificates to secure multiple services behind a single Istio Ingress Gateway. This simplifies certificate management and improves the user experience by allowing seamless connections across different services within the same domain. The article demonstrates the configuration process step-by-step and explains how wildcard certificates are matched to incoming requests. Read the full blogWeek of June 17-21Learn how to leverage BigQuery vector search to analyze your logs and asset metadata stored in BigQuery. Using vector search, you can find semantically similar logs which can be helpful in several use cases such as outlier detection, triage and investigation. This how-to blog walks you through the setup from processing logs, generating vector embeddings, to analyzing vector search results. It includes sample SQL queries which can be adapted for your own logs and use case. Read the full blog.Nuvem, first announced last year, is a transatlantic subsea cable system that will connect Portugal, Bermuda, and the United States. We are now working with the Regional Government of Azores to enable extending the system to the Azores as well. Named after the Portuguese word for “cloud,” Nuvem will improve network resiliency across the Atlantic, helping meet growing demand for digital services and further establishing its landing locations as digital hubs.Week of June 10-14General Availability of A3 Mega, a new instance type in the A3 VM family. A3 Mega is powered by the NVIDIA H100 Tensor Core GPU, delivers a 2.4x improvement in large scale training performance over multiple A3 instances.2x the GPU-to-GPU networking bandwidth over A3 Instances.Enhanced GPUDirect-TCPXO networking offloads GPUDirect memory access from the CPU, providing direct access through through the NIC (Network Interface Card) to GPU memory, based on Titanium TOPs, which improves performance of multi-node distributed training workloads.Simplify your Network: The Cloud Networking Product Management and Engineering team will be traveling across US cities in June/July and Sept. Learn how Cross-Cloud Network can transform your infrastructure. The workshop will address Cross-Cloud Networking for hybrid and multicloud enterprises with distributed applications, internet-facing content and applications, security, and AI-assisted network operations with Gemini Cloud Assist. Join us at one of the following Google office locations and meet the experts who will share the latest innovations, use cases, and demos. Register here.Learn how you can leverage the cloud deployment archetypes (zonal, regional, multi-regional, global, hybrid, & multicloud) to architect cloud topologies that meet your workload’s requirements for reliability, cost, performance, & operational simplicity. Read the full blog.Week of May 20-24Maximize performance and optimize spend with Compute Engine’s latest General Purpose VMs, N4 and C4. N4’s flexible configurations and price-performance gains help optimize costs, while C4 provides top-tier performance for demanding applications. With N4 and C4, you get tailored solutions for all your general-purpose workloads, so you can lower the total cost of running your business without compromising on performance or workload-specific requirements. Learn more here.Week of Apr 22 – April 26Simplify your connectivity to Google by using a Verified Peering Provider to connect to Google, instead of using Direct Peering. Verified Peering Providers handle all of the complex connectivity allowing you to focus on your core business. Learn more here.Week of Apr 15- Apr 19New training in AI, data analytics and cybersecurity, designed to expand onramps to tech careers through colleges and employers. Learn more.Week of Apr 1- Apr 5Security Command Center (SCC) Enterprise is now generally available (GA). It is the industry’s first cloud risk management solution that converges cloud security and enterprise security operations into a single platform, supercharged by Mandiant expertise and AI. Learn more in our announcement blog.Identify common container runtime attacks, analyzes suspicious code, and use natural language processing to pinpoint malicious scripts with GKE threat detection, powered by Security Command Center. Now in public preview.Get a fully managed compliance service that automatically delivers end-to-end coverage for GKE, scanning for issues against the most important benchmarks with GKE compliance, now in public preview. Near-real-time insights are available in a centralized dashboard, with compliance reports automatically produced for you.Streamline your GCE backup strategy! With tag-based backups in Google Backup and DR, protection is automated – new VMs with the right tags are protected immediately, saving you time and increasing reliability. Read more on the blog here. Differential privacy enforcement with privacy budgeting is now available in BigQuery data clean rooms so organizations can prevent data from being reidentified when it is shared.Week of Mar 18- Mar 22Google Kubernetes Engine (GKE) and NVIDIA NeMo framework are used to train large language models (LLMs). Due to the increasing demand for efficient and scalable training of LLMs, the need for GPUs at a large scale with high speed networking is rapidly growing. GKE offers a comprehensive set of features that make it suitable for enterprise-level training and inference. This blog post shows how generative AI models can be adapted to your use cases by demonstrating how to train models on Google Kubernetes Engine (GKE) using the NVIDIA NeMo framework.Cloud Run now supports volume mounts! Mount a Cloud Storage bucket or NFS file share as a volume to easily serve static assets, access app configuration data, or access an AI/ML model. Learn more in our blog post.Week of Mar 11- Mar 15Datastream adds support for SQL Server sources, now in preview. With existing support for MySQL, PostgreSQL, and Oracle, support for SQL Server sources extends the reach of Datastream and empowers you to replicate data from a range of relational sources to several Google Cloud services, such as BigQuery, Cloud Storage, AlloyDB, and Spanner. Read more in the blog here.Week of Feb 5- Feb 9Check out this new blog and learn more about the Integrated Commerce Network (ICN) delivered by Kin + Carta and built on Google Cloud. The ICN features 3 of our premier digital commerce partners for an integrated end-to-end solution including Bloomreach, commercetools and Quantum Metric.Week of Jan 29- Feb 2IDC finds 318% ROI from migrating to Google Cloud IaaS: Check out the latest IDC research study to learn how organizations worldwide are benefitting by adopting Google Cloud Infrastructure as a Service.Week of Jan 15-19Check out the latest generative AI training available from Google Cloud : Take a look at our top ten trainings in Duet AI to help boost your productivity in 2024.Week of Jan 1-5The year in Google Cloud: Top news of 2023: A look back at the biggest stories of 2023 from Google Cloud, covering generative AI, DevOps, containers, data and databases, security, and more.
Quelle: Google Cloud Platform