Bits With Brains
Curated AI News for Decision-Makers
What Every Senior Decision-Maker Needs to Understand About AI and its Impact
The AI Revolution: Google and OpenAI's Relentless Quest to One-Up Each Other
5/20/24
Editorial team at Bits with Brains
Rapid advancements in artificial intelligence (AI) are transforming industries and reshaping the way we live and work, virtually in real-time. At the forefront are two very competitive frontier labs, Google and OpenAI, each with their unique strategies for driving AI innovation.
OpenAI and Google are two of the most important frontrunners in generative AI. OpenAI recently introduced GPT-4o, a powerful multimodal model that can process text, images, audio and video, offering significant improvements in conversational AI, reasoning capabilities, and real-time interactions compared to previous models like GPT-4. Meanwhile, Google unveiled advancements at its I/O 2024 event, including Gemini 1.5 for enhancing AI capabilities across its products, and Project Astra, a real-time AI agent that can interact with the environment through a mobile camera. The two companies are taking different approaches, with OpenAI focusing on cutting-edge research and democratizing access, while Google aims to seamlessly integrate AI into its existing ecosystem.
OpenAI's Spring Event: GPT-4o (omni) and Multimodal AI
OpenAI's Spring Event on May 13th introduced GPT-4o, a very powerful multimodal model that can handle audio, video, images, and text. This model offers a significant leap in conversational AI, offering faster, more efficient, and context-aware interactions. The event showcased GPT-4o's ability to solve complex problems, generate images, and create 3D objects, highlighting the model's versatility and potential for various applications.
Here are the key features of GPT-4o and how it compares to its predecessors like GPT-4:
Multimodal capabilities: GPT-4o is a multimodal model that can handle text, images, audio, and video inputs and outputs simultaneously. This is a significant advancement over previous models which were primarily focused on text and had multimodal capabilities added subsequently.
Improved speed and efficiency: GPT-4o is much faster than GPT-4, with up to 2x faster response times. It is also 50% cheaper for developers to implement and has 5x higher rate limits compared to GPT-4 Turbo.
Enhanced language support: GPT-4o offers improved performance across 50+ languages. It especially shows better results on non-English languages compared to GPT-4.
Advanced reasoning capabilities: On benchmarks like MMLU (general knowledge questions), GPT-4o sets new records, achieving 88.7% on zero-shot Chain of Thought prompts and 87.2% on 5-shot prompts, demonstrating strong reasoning skills.
State-of-the-art vision and audio understanding: GPT-4o outperforms previous models like Whisper-v3 on speech recognition, translation, and visual understanding tasks across various benchmarks.
Real-time interactions: With response times as low as 232ms on average, GPT-4o enables more natural real-time conversations, nearing human level interaction speeds.
Expanded access: OpenAI is making GPT-4o available to all ChatGPT users, including the free tier, though with some usage limits compared to paid plans. This democratizes access to advanced AI capabilities.
GPT-4o is a significant advancement over models like GPT-4 and GPT-3.5, with its multimodal skills, improved speed and efficiency, better language support, and top-tier performance on reasoning and perception benchmarks. Its real-time interaction abilities and expanded availability to users set the stage for more advanced and natural human-AI interactions.
Google I/O 2024: Gemini AI and Project Astra
Google I/O 2024, held the day after, on May 14 2024, showcased more AI advancements. The introduction of Gemini 1.5 with a one million token context window (around 750,000 words) and its lightweight variant, Gemini 1.5 Flash, highlighted Google's focus on enhancing AI capabilities across its suite of products, including Gmail, Google Photos, and Google Drive. These models are designed to improve user experience and productivity by streamlining workflows and making complex tasks more manageable, similar to Microsoft’s use of CoPilot.
One of the most exciting announcements at Google I/O was Project Astra, a real-time AI agent capable of interacting with the environment through a mobile phone camera. This technology opens new possibilities for AI applications, particularly in fields such as augmented reality, real-time data analysis, and interactive learning. The ability to process and respond to visual data in what was effectively real-time is a significant step towards more intuitive and intelligent AI systems that can assist users in a variety of contexts.
Here are some key features of Google's Project Astra:
Multimodal AI assistant: Project Astra is a multimodal AI agent that can answer real-time questions fed to it through text, video, images, and speech. It can process visual information immediately and work from a massive database.
Sees and understands the world: Astra uses the camera and microphone to understand the user's environment and context. It knows what things are and where the user left them.
Responds naturally in conversation: Astra can engage in natural back-and-forth conversation with the user at a conversational pace, making interactions feel more lifelike. It has a wide range of intonations thanks to Google's speech models.
Remembers and recalls information: Astra encodes video frames and speech into a timeline that it can refer back to later. This allows it to remember objects and their locations during a session.
Identifies objects and provides details: In demos, Astra could identify specific parts of objects like speakers, explain code on a computer screen, and even interpret hand-drawn images.
Integration with devices: While still a prototype, the goal is for Astra's capabilities to eventually be accessible through smartphones and smart glasses, acting as an ever-present AI assistant.
Project Astra relies on Google's Gemini foundation model to enable its multimodal reasoning and interaction capabilities. It represents Google's vision for a highly capable, contextually aware AI assistant that can see, hear, understand, converse, and remember - with the aim of being a useful companion in people's daily lives. While still in the research and development phase, it is very promising.
Google also announced AI-driven search enhancements. These allow for multi-step reasoning and provide comprehensive answers to complex queries directly within the search results – a counter to the success of Perplexity AI.
Google's strategy for AI revolves around seamless integration into its existing ecosystem of products. The company aims to enhance user experience and productivity by leveraging AI tools and features powered by the Gemini models. From real-time email summarization in Gmail to context-aware photo searches in Google Photos and advanced data analysis in Google Sheets, Google is making AI an integral part of everyday digital life within its ecosystem.
The Rise of AI Agents and Multimodal Capabilities
Both Google and OpenAI are investing heavily in the development of AI agents with multimodal capabilities. These agents, such as Google's Project Astra and OpenAI's GPT-4o, can interact with their environment in real-time, processing inputs from various sources like text, images, and video. This capability enables more dynamic and responsive AI applications, ranging from real-time object recognition to interactive storytelling.
The integration of these AI agents into everyday tools and devices is a significant shift towards more intuitive and intelligent AI systems that can assist users in a variety of contexts.
Both Google and OpenAI are at the forefront of AI innovation, but their strategies differ. Google's primary goal is to integrate AI seamlessly into its existing ecosystem, enhancing user experience and productivity through tools like Gemini and Project Astra. The company aims to make AI a natural part of everyday life, accessible to users across its suite of products.
On the other hand, OpenAI seems to be more aggressively pushing the boundaries of AI capabilities with models like GPT-4o, which offer multimodal processing and advanced conversational abilities. The company's focus is on developing cutting-edge technologies that redefine what AI can do, democratizing access to these powerful tools through initiatives like making GPT-4o available to all ChatGPT users.
Both Google and OpenAI's strategies have their respective strengths, and together they are helping drive the AI industry forward, offering diverse solutions to meet diverse needs. As these companies, and others, continue to innovate and compete, we can expect to see continuing rapid advancements in AI technology, ultimately benefiting businesses and consumers alike.
For C-level decision-makers in the public and private sectors, understanding the strategies and advancements of these AI frontier labs is crucial for making competent decisions about implementing AI in their organizations.
Sources:
[1] https://economictimes.indiatimes.com/topic/project-astra-ai-assistant
[2] https://www.techtarget.com/searchenterpriseai/tip/GPT-35-vs-GPT-4-Biggest-differences-to-consider
[3] https://openai.com/index/gpt-4/
[4] https://openai.com/index/gpt-4-research/
[5] https://www.digitaltrends.com/computing/chatgpt-4-claims-40-percent-better-at-facts/
[6] https://cloud.google.com/blog/topics/startups/ai-startups-at-next24
[8] https://blog.google/technology/ai/
[9] https://openai.com/index/new-models-and-developer-products-announced-at-devday/
[10] https://www.datastax.com/products/real-time-ai
[11] https://www.wired.com/story/what-is-chatgpt-plus-gpt4-openai/
[12] https://www.datastax.com/products/datastax-astra
[13] https://www.youtube.com/watch?v=--khbXchTeE
[14] https://cloud.google.com/products/ai
[15] https://azure.microsoft.com/en-us/blog/introducing-gpt4-in-azure-openai-service/
[16] https://explodingtopics.com/blog/ai-statistics
[18] https://community.openai.com/t/how-to-get-access-to-gpt-4-vision-preview/475454
[20] https://blog.google/products/chrome/google-chrome-generative-ai-features-january-2024/
Sources