@_philschmid: Gemini Interactions API Update As we move beyond simple prompts, strict "user" and "model" roles felt limiting. That's …
Summary
Google is updating the Gemini Interactions API to replace strict user/model roles with a flexible step-based system (outputs + roles → steps), introducing agentic steps like user_input, thought, function_call, tool_call, and model_output. The update also consolidates response_format controls and requires SDK upgrades (Python/JS ≥2.0.0) or a new API header to opt-in.
Similar Articles
@_philschmid: Install Skill: ``` npx skills add google-gemini/gemini-skills --skill gemini-interactions-api ``` Prompt your agent: "/…
Google released a library of skills for the Gemini API that improves agent accuracy in generating correct API code, with evaluations showing 87% accuracy with Gemini 3 Flash and 96% with Gemini 3.1 Pro.
Introducing the Gemini 2.5 Computer Use model
Google releases Gemini 2.5 Computer Use model via the Gemini API, enabling developers to build AI agents that can interact with user interfaces through clicking, typing, and scrolling. The model outperforms alternatives on web and mobile control benchmarks with lower latency and is available in preview on Google AI Studio and Vertex AI.
Introducing Gemini 2.0: our new AI model for the agentic era
Google DeepMind introduces Gemini 2.0, a new agentic AI model with native image and audio output, enhanced tool use, and multimodal capabilities designed for the next era of AI agents. Gemini 2.0 Flash is now available to developers with wider availability planned for early 2025.
Start building with Gemini 3
Google has launched Gemini 3 Pro, a new AI model designed to outperform previous versions in coding, agentic workflows, and multimodal reasoning. The model is available via the Gemini API, Google AI Studio, and the new Google Antigravity development platform.
Improved Gemini audio models for powerful voice experiences
Google has updated Gemini 2.5 Flash Native Audio to improve live voice agent capabilities, including sharper function calling, better instruction following, and smoother conversation context retrieval. The update also introduces live speech translation in the Google Translate app beta, preserving intonation across 70+ languages.