Interactions API: A unified foundation for models and agents
Today, we’re introducing the Interactions API, a unified interface for interacting with our models, like Gemini 3 Pro, and agents like Gemini Deep Research. It’s available for developers in public beta through the Gemini API in Google AI Studio.
The Interactions API introduces a native interface specifically designed to handle complex context management when building agentic applications with interleaved messages, thoughts, tool calls and their state.
Alongside our suite of Gemini models, the Interactions API provides access to our first built-in agent: Gemini Deep Research (Preview), a state-of-the-art agent capable of executing long-horizon research tasks and synthesizing findings into comprehensive reports.
This is just the start. We will expand built-in agents and introduce the ability to build and bring your own agents. This will enable you to connect Gemini models, Google’s built-in agents, and your custom agents using one API.
The Interactions API at a glance
Interactions API offers a single RESTful endpoint (/interactions) for interacting with models and specialized agents.
Interact with models by specifying the "model" parameter.
Interact with agents by specifying the "agent" parameter.
Currently supports deep-research-pro-preview-12-2025.
The Interactions API extends the core capabilities of generateContent with the features required for modern agentic applications, including:
- Optional server-side state: The ability to offload history management to the server. This simplifies your client code, minimizes context management errors and may result in reduced costs via increased chance of cache hits.
- Interpretable and composable data model: A clean schema designed for complex agentic histories. You can debug, manipulate, stream and reason over interleaved messages, thinking, tools and their results.
- Background execution: The ability to offload long-running inference loops to the server without maintaining client-side connections.
- Remote MCP tool support: Models can directly call Model Context Protocol (MCP) servers as tools.
Our motivations for a new API
Models are becoming systems and over time, might even become agents themselves. When we designed generateContent, the primary use case was stateless request-response text generation. This works perfectly for chatbots and completion tasks.
Since then, the landscape has shifted. With the arrival of new model capabilities like “thinking” and advanced tool use, we heard lots of feedback around providing a native interface designed to support these complex interaction patterns. As we expand the Gemini ecosystem from models to fully managed agents like Gemini Deep Research, trying to force these capabilities into generateContent would have resulted in an overly complex and fragile API.
While the Interactions API supports most generateContent features and offers a more robust developer experience, it is currently in public beta and thus is still subject to breaking changes. For standard production workloads, generateContent remains the primary path and will continue to be developed and maintained.
Getting started
You can start building with the Interactions API public beta today using your Gemini API key from Google AI Studio following the API documentation. You can also check out the OpenAPI spec here.
We want to ensure this API solves the real friction points you face when moving agents to production, while still letting you seamlessly spin up new lightweight experiences, so we encourage you to test it and share your feedback.
We are committed to bringing these capabilities to the broader open source ecosystem so you can use them with your favorite tools. As a first step, the Agent Development Kit (ADK) and Agent2Agent (A2A) protocol now support the Interactions API. You can expect broader support across other tools in the coming months. Interactions API and Gemini Deep Research will be coming soon to Vertex AI.