Back to AI Timeline
2024Model

Gemini Launches (Google)

Google launched Gemini, its most capable multimodal AI model family, designed to natively understand and reason across text, code, images, audio, and video. Gemini Ultra matched or exceeded GPT-4 on key benchmarks, signaling Google's return to the forefront of the AI race. The model was integrated across Google products from Search to Workspace.

In December 2023 (with broader availability in early 2024), Google launched Gemini, its most ambitious AI model family, designed from the ground up to be natively multimodal -- understanding and reasoning across text, code, images, audio, and video simultaneously. Gemini Ultra, the largest variant, matched or exceeded GPT-4 on key benchmarks, signaling that Google had closed the gap in the AI race it had seemed to be losing.

The Model Family

Gemini came in three sizes: Ultra (the most capable, for complex tasks), Pro (balanced performance for a wide range of tasks), and Nano (efficient enough to run on mobile devices). This tiered approach allowed Google to deploy Gemini across its entire product ecosystem, from cloud services requiring maximum capability to smartphones requiring on-device inference. Each variant was designed to excel at its intended deployment scenario.

Native Multimodality

Unlike models that bolted on image understanding as an afterthought, Gemini was trained from the start to process multiple modalities simultaneously. It could reason about video clips, understand audio conversations, analyze images, and process text -- all within the same model and sometimes within the same query. Google demonstrated the model solving physics problems by watching a video, understanding handwritten notes, and generating code from architectural diagrams.

The Benchmark Results

Gemini Ultra achieved state-of-the-art results on 30 of 32 academic benchmarks tested. Most notably, it scored 90.0 percent on MMLU (Massive Multitask Language Understanding), claiming to be the first model to outperform human experts on this comprehensive knowledge benchmark. It also showed strong results on multimodal benchmarks that tested combined reasoning across text and images.

The Demo Controversy

Google's initial demo video for Gemini was criticized for being misleadingly edited. The video appeared to show real-time multimodal interaction, but it was actually composed of carefully selected responses to still images with text prompts. Bloomberg and other outlets reported the discrepancy, leading to accusations of overhyping. Google acknowledged that the demo was illustrative rather than literal, but the controversy highlighted the tension between marketing and transparency in AI.

Integration Across Google

Gemini was integrated into Google's vast product ecosystem. Google Search received Gemini-powered overviews. Google Workspace products (Docs, Gmail, Sheets) gained Gemini-based assistance. Android phones received Gemini Nano for on-device AI features. Google Cloud offered Gemini Pro and Ultra through its Vertex AI platform. The breadth of integration demonstrated Google's advantage in having established products that could immediately benefit from improved AI.

Google's AI Reorganization

The launch of Gemini was accompanied by a major reorganization at Google. The Google Brain and DeepMind teams were merged into a single unit called Google DeepMind, led by Demis Hassabis. This consolidation brought together Google's two premier AI research groups, combining Brain's expertise in large-scale model training with DeepMind's strength in reinforcement learning and scientific applications.

Competitive Significance

Gemini's launch marked Google's return as a serious competitor in the foundation model space after appearing to fall behind during the ChatGPT era. The company's advantages -- vast computational resources, massive proprietary datasets, a large AI research team, and established products for distribution -- made it a formidable competitor. The AI race was no longer a two-horse contest between OpenAI and everyone else.

Key Figures

Demis HassabisSundar PichaiJeff DeanOriol Vinyals

Lasting Impact

Gemini marked Google's return to the forefront of the AI race with a natively multimodal model family integrated across its product ecosystem. It demonstrated that the competition for AI leadership would be fierce and multi-dimensional.

Related Events

2023Model
GPT-4 Launches

OpenAI released GPT-4, a multimodal model capable of processing both text and images with significantly improved reasoning abilities compared to its predecessors. It scored in the top percentiles on professional exams including the bar exam and medical licensing tests. GPT-4 set a new benchmark for what large language models could achieve.

2022Milestone
ChatGPT Launches

OpenAI released ChatGPT on November 30, 2022, and it became the fastest-growing consumer application in history, reaching 100 million users in just two months. Built on GPT-3.5 with reinforcement learning from human feedback, it made conversational AI accessible to the general public. ChatGPT fundamentally shifted public perception of AI capabilities and triggered an industry-wide race.

2024Model
Claude 3 Family

Anthropic released the Claude 3 model family consisting of Haiku, Sonnet, and Opus, offering different capability and speed trade-offs. Claude 3 Opus matched leading models on benchmarks while maintaining Anthropic's emphasis on safety and reliability. The tiered approach gave users flexibility to choose the right balance of performance and cost for their needs.