Google's new AI can now find lost specs

Google appears keen to stress its tools are as capable of this kind of so-called multimodal understanding as its rival

New Update
Google's new AI can now find lost specs
Advertisment

AI systems able to interpret information in images via a phone camera, in videos and sounds and in spoken language have been shown off by Google.

Advertisment

In one demo a prototype AI powered assistant running on a phone was able to answer the age-old question "where did I put my glasses".

It comes a day after rival OpenAI's launch of its latest AI system, GPT-4o, which included an eye-catching presentation in which it read human expressions via a phone camera, and chatted and flirted fluently.

Google appears keen to stress its tools are as capable of this kind of so-called multimodal understanding as its rival.

The firm showcased multimodal features in Gemini Nano, an AI assistant that runs "on device" on its Pixel phone, and in the Gemini App.

It also demonstrated a prototype scam alert feature it was testing for Gemini Nano that could listen to a phone call and warn that it was a scam, without any information about the call leaving the phone.

The new AI powered demos were revealed at Google I/O, the firm's annual presentation for software developers.

Latest Stories