💻 Introducing Gemini 2.5 Computer Use, available today in preview via the API. The model builds on Gemini 2.5 Pro’s vision and reasoning capabilities to power agent interactions with UIs. It completes tasks with lower latency, and outperforms alternatives on web and mobile control benchmarks. Gemini 2.5 Computer Use is optimized for web tasks such as clicking, typing, and scrolling, and shows promise for mobile UI control tasks. These improved capabilities mean AI models can excel with less direct interaction. 🟦 Blog for more details: https://goo.gle/3Wtwghy 🟦 Dev docs for a quickstart: https://goo.gle/4mZOa6x 🟦 Experiment in Google AI Studio: ai.studio
Google AI for Developers
Technology, Information and Internet
AI for every developer. So what will you build?
About us
Our goal is to equip developers with the most advanced models to build new applications, helpful tools to write better and faster code, and make it easy to integrate across platforms and devices.
- Website
-
https://goo.gle/ai-devs
External link for Google AI for Developers
- Industry
- Technology, Information and Internet
- Company size
- 10,001+ employees
Updates
-
Build powerful AI agents with Gemini CLI extensions. Integrate with partners like Pinecone, Elastic, Neo4j, HashiCorp, MongoDB, GitLab, and more directly from your terminal. Develop agentic workflows to automate MLOps tasks, manage data pipelines, and streamline the entire process of building with generative AI. Learn more → https://goo.gle/4mSymCs
-
-
Join the Google Web AI Summit 2025 in Sunnyvale, CA on October 17 to hear from top experts on building AI applications in the browser: https://lnkd.in/gbTPiqmN Check out the agenda: https://goo.gle/4gWHfcK Register with code: GEMMAFORWEBAI
-
-
Join the Google Chrome Built-in AI Challenge 2025 to create new web applications or Chrome Extensions using a suite of AI APIs and powerful built-in models, including Gemini Nano. 🦖 Submit by October 31: https://goo.gle/4nZfH8U
-
Build multimodal apps with the Gemini Live API and Pipecat in the special voice and real-time AI hackathon on October 11 👇
Join us for a voice and realtime, multimodal AI hackathon at Y Combinator on October 11th. Build an application using Gemini and Pipecat. See some new APIs. Show off interesting things you're doing in your startup or side project. Hang out with engineers from Google DeepMind and Google Cloud, plus YC companies Daily, Boundary, Coval, Langfuse, and Tavus. Eat Outta Sight pizza. Space is limited. Apply here: https://lnkd.in/gPNhK5Mr
-
-
Give your apps the power of multi-spectral imagery using the native multimodal capabilities of Gemini models. Learn how: https://goo.gle/3ITyjIC Multi-spectral data unlocks imagery across the electromagnetic spectrum, allowing for the detection of water in floodplains, burn areas after a wildfire, vegetation to assess crop health, and more. Get started with our Colab Notebook: https://lnkd.in/eZaAzwbA Dive into the research: https://lnkd.in/emQ3q3um
-
-
🖼️ Nano Banana is now generally available and ready for production. Learn how you can build dynamic user experiences with a wider range of aspect ratios, the ability to specify image-only output, and more creative control in your app: https://goo.gle/4o0p73Z Our SOTA image gen and editing model features: → 10 supported aspect ratios → Multi-image blending → Consistent character likeness → Targeted edits using natural language Start building via the Gemini API and in Google AI Studio: https://goo.gle/4gU2uvR
-