
Gemini Live: Enhanced AI Features and Accessibility
Google's autonomous AI assistant Gemini Live has improved speech understanding and enhanced access to services like Maps.
The 'game changing phone', as Google's Head of Product Rick Osterloh called the new Pixel 10 at the 'Made by Google' event, clearly includes more artificial intelligence. Gemini is a deeply integrated part of the Google smartphone. However, the AI model called Gemini is also accessible via headphones, smartwatches, soon in the car and for the TV, as well as being usable for the Home setup. The autonomous Gemini Live app is also getting an update. This is the AI assistant, which is set to become even more helpful and fun.
Gemini Live can see what the camera lens sees. At the same time, you can talk to the AI and ask questions. A new feature is that Gemini can highlight individual objects in the image. For example, if you ask in front of the spice rack which spice could replace another, Gemini circles the one that it thinks would fit. In the blog post, Google gives the example of missing cumin, which can be replaced by coriander.
However, this should also be helpful if you are looking for the right tool or don't know which pair of shoes goes better with an outfit. Gemini, the style advisor. The function is called Visual Guidance. It is immediately available for Pixel 10 devices – as soon as they hit the shelves –, other Android devices will follow in the coming weeks. It is not entirely clear whether this will also apply to Germany immediately.
However, if Google's event is anything to go by, Gemini Live will soon make it easy to become a Formula 1 driver or professional basketball player. However, a video in which racing driver Lando Norris and basketball player Giannis Antetokounmpo swap roles ends with both of them giving up.
Gemini Live now also has access to Google Calendar, Keep and Tasks – so you can enter appointments in another app via Gemini, make shopping lists and create tasks and reminders. Messages, the smartphone itself and the watch will soon be added, as will Google Maps.
As with Gemini Live, all functions can be accessed using natural language. Gemini itself, i.e., the AI model, is set to receive an update in the coming weeks, which will significantly improve its linguistic capabilities. This includes the ability to ask Gemini to speak more slowly or in a funny dialect or with a dramatic tone.
Attentive AI enthusiasts will now be thinking that this is perfect for AI glasses. Obviously, Google is not finished yet. Even at the beginning of the event, Osterloh says that they are working on it. So it can't be long now.