NEW YORK: Google is expected to announce Tuesday how it leverages artificial intelligence (AI) across its business, a day after OpenAI demonstrated a new AI model that can work with sound and images in addition of text.
Google is preparing to announce its latest innovations in its business at the annual I/O developer event.
The I/O developer event serves as a platform for Alphabet, Google’s parent company, to showcase its cutting-edge technology, providing investors with valuable insights into its competitive position, particularly compared to competitors like OpenAI, supported by Microsoft. As AI is increasingly integrated into products such as Google’s ubiquitous search engine, the event holds significant importance in assessing Alphabet’s progress in this rapidly evolving area.
Among the expected highlights of the event is the launch of the next iteration of Google’s Android operating system, as well as a series of updates to its advanced AI technology known as Gemini. Google’s teaser video ahead of the event hinted at the unveiling of AI software capable of drawing conclusions from live camera feeds, showcasing virtual assistant-like interactions that mirror a real-world conversation.
Interestingly, the capabilities demonstrated in Google’s teaser resemble features shown by OpenAI in its recent introduction of GPT-4o, a cutting-edge multimodal AI model. OpenAI’s latest offering allows ChatGPT to respond in real-time via voice and seamlessly handle interruptions, mimicking the fluidity of human conversation, a feat that traditional AI voice assistants struggle to achieve.
OpenAI’s decision to offer GPT-4o to ChatGPT users for free, without advertising, highlights the company’s commitment to advancing the accessibility and functionality of AI technology. The move also highlights intensifying competition in the AI landscape, with major players like Google and OpenAI vying to push the boundaries of what AI can accomplish.