NEW YORK: Google is expected to announce on Tuesday how it is building on artificial intelligence (AI) across its businesses, a day after OpenAI demonstrated a new AI model that can work with sound and images in addition to text.
Google is gearing up to announce its latest innovations across its businesses at the annual I/O developer event.
The I/O developer event serves as a platform for Alphabet, Google’s parent company, to showcase its cutting-edge technology, offering investors valuable insights into its competitive standing, particularly in comparison to rivals like Microsoft-backed OpenAI. With AI increasingly integrated into products such as Google’s ubiquitous search engine, the event holds significant weight in assessing Alphabet’s progress in this rapidly evolving field.
Among the expected highlights of the event is the launch of the next iteration of Google’s Android operating system, alongside a slew of updates to its advanced AI technology known as Gemini. Google’s teaser video ahead of the event hinted at the unveiling of AI software capable of drawing conclusions from live camera feeds, showcasing virtual-assistant-like interactions that mirror real-world conversation.
Interestingly, the capabilities demonstrated in Google’s teaser bear resemblance to features showcased by OpenAI in its recent introduction of GPT-4o, a cutting-edge multimodal AI model. OpenAI’s latest offering enables ChatGPT to respond in real-time via voice and seamlessly handle interruptions, mimicking the fluidity of human conversation—a feat that traditional AI voice assistants have struggled to achieve.
OpenAI’s decision to offer GPT-4o to ChatGPT users for free, without advertisements, underscores the company’s commitment to advancing the accessibility and functionality of AI technology. The move also highlights the intensifying competition in the AI landscape, with major players like Google and OpenAI vying to push the boundaries of what AI can accomplish.