New generation

Gemini: Google AI can complete tasks independently

Google Gemini Quelle Primakov / Shutterstock.com
Image source: Primakov / Shutterstock.com

A new generation of Google’s Gemini AI system will in future be able to perform certain tasks independently as an assistant. Google announced this on the anniversary of the first Gemini premiere in Mountain View.

It is conceivable that the “AI agent” could, for example, find the components for a hobby project in online stores and place them in the shopping cart independently. However, the actual ordering process would still be carried out by the user.

Ad

“AI should become more useful”

Google CEO Sundar Pichai said that the first generation Gemini 1.0 was about organizing and understanding information. “Gemini 2.0 is about being much more useful.”

The new AI assistants are part of “Project Mariner”, which was implemented with the new generation of AI. Gemini 2.0 has been developed for the age of agents, said Google manager Tulsee Doshi. The system is able to use intelligent tools and can directly access Google products such as search. It can even execute program code. “These capabilities result in agents that can think, remember, plan and even take action on behalf of users.”

Click like a human

“Mariner” therefore behaves in exactly the same way as human users would in a browser, emphasized Doshi. “It can click, type and scroll, just like you as a user”. However, the system also identifies a number of tasks that the agent should not perform on behalf of a user. “A good example of this is that Mariner does not complete a purchase without first asking the user whether they really want to do so.” “Mariner” will first be tested with trustworthy test persons before it is made available to a wider public.

Ad

New AI glasses

In “Project Astra”, Google is now driving forward a research project for exploring the environment based on Gemini 2.0, which was first presented at the Google I/O developer conference last spring. In future, users will not only receive useful information on a smartphone, but also on smart glasses. They are similar to the Ray-Ban glasses from the Facebook group Meta, which can recognize buildings or works of art, for example, or help with cooking.

For software developers who offer their own solutions based on Google AI, the Gemini Flash system variant is particularly relevant. It can run not only on large computers in the data center, but also locally on personal computers or certain smartphone models. Google presented Gemini Flash 2.0 on Wednesday, which, according to the company, offers improved performance with similarly fast response times.

dpa

Ad

Weitere Artikel