SAN FRANCISCO, May 14 (Reuters) - Alphabet is
expected to announce on Tuesday how it is building on artificial
intelligence across its businesses, a day after OpenAI
demonstrated a new AI model that can work with sound and images
in addition to text.
The annual I/O developer event is typically when Alphabet
showcases its latest new technology and whatever it unveils will
help investors evaluate its progress against Microsoft ( MSFT )
backed OpenAI. This event serves as a critical indicator of
Alphabet's competitive position in tech as it introduces AI into
products such as search, used by billions of people around the
world.
Google is expected to launch the next version of its Android
operating system and a host of updates to its most advanced AI
technology called Gemini.
A day ahead of the event, Google teased in a video a piece
of software which appeared to demonstrate AI software that could
draw conclusions about the live feed from the camera. The phone
operator asked several questions and the app answered in a
virtual-assistant-like voice.
The software looked to have similar capabilities as one of
the features demonstrated by OpenAI in its launch of its latest
AI model, GPT-4o.
OpenAI on Monday showcased its new multimodal AI model
called GPT-4o, which enables ChatGPT to respond via voice in
real-time and be interrupted, both hallmarks of realistic voice
conversations that AI voice assistants like Google Assistant
have found challenging. The company said it would be offering it
to ChatGPT users for free, with no advertisements.
The conference is held at the Shoreline Amphitheater in
Mountain View, California. CEO Sundar Pichai and other
executives are expected to begin presenting at 10:00 a.m.
Pacific Time.