2026 — Unified Models & Robotics
Mar 12, 2026
LTX 2.3 generates synchronised video and audio in a single pass
Lightricks released LTX 2.3, a 22-billion-parameter diffusion transformer model that generates synchronised video and audio in a single forward pass. The model supports resolutions up to 4K at 50 frames per second, marking a significant leap in real-time media generation quality.
Info
Mar 5, 2026
GPT-5.4 launches with 1 million token context window
OpenAI released GPT-5.4, its most capable frontier model, available in Standard, Thinking, and Pro variants. The model features a context window of up to 1 million tokens (the largest from OpenAI), a reported 33% reduction in factual errors compared to GPT-5.2, and improved capabilities across coding, reasoning, and agentic workflows.
Info
2025 — Robotics & Reasoning
Aug 20, 2025
Atlas Humanoid with Neural Large Behavior Models
Boston Dynamics demonstrated a newly redesigned Atlas humanoid robot powered by neural Large Behavior Models from Toyota Research Institute. The robot performed complex multi-task sequences with self-correction, learning control policies without hand-coded routines. Boston Dynamics has deployed over 500 robots with revenue exceeding $130 million.
Info
Jan 20, 2025
DeepSeek R1 Open-Source Reasoning Model
DeepSeek released R1, an open-source reasoning model that demonstrates competitive performance with proprietary frontier models. The release includes both full weights and distilled versions, making advanced reasoning capabilities accessible to the open-source community.
Warning
2024 — Agent & Vision Capabilities
Oct 29, 2024
Claude 3.5 Sonnet with Computer Use
Anthropic released Claude 3.5 Sonnet with native computer interaction capabilities, allowing the model to see, understand, and control a computer screen. This enables autonomous execution of multi-step digital workflows without relying on separate tool APIs.
Warning
Sep 12, 2024
OpenAI o1 Reasoning Model Launch
OpenAI introduced o1, a model trained to spend more time thinking through problems before responding. It achieves state-of-the-art performance on mathematical, coding, and scientific reasoning tasks by using reinforcement learning to develop internal reasoning processes.
Warning
Jul 23, 2024
Meta Llama 3.1 405B Open-Source Release
Meta released Llama 3.1 405B, a 405-billion parameter open-source model that rivals closed proprietary models on performance benchmarks. The full weights were made freely available for research and commercial use.
Warning
May 13, 2024
GPT-4o Multimodal Model
OpenAI released GPT-4o, a model optimized to handle text, vision, and audio seamlessly in a unified way. The model shows significant performance improvements over GPT-4 and can process audio and images natively without intermediate conversions.
Warning
May 8, 2024
AlphaFold 3 Predicts Protein-Ligand Complexes
DeepMind released AlphaFold 3, expanding beyond protein structure prediction to accurately predict protein-DNA, protein-RNA, and protein-ligand interactions. The model achieved 50% accuracy improvement over AlphaFold 2 and contributed to structural understanding underlying the 2024 Nobel Prize in Chemistry.
Info
2023 — Context & Reasoning
Feb 15, 2024
Google Gemini 1.5 Pro with 1M Token Context
Google introduced Gemini 1.5 Pro, a model capable of processing a context window of up to 1 million tokens. This enables the model to work with entire books, lengthy video transcripts, and massive code repositories in a single prompt.
Info
Dec 11, 2023
Mixtral 8x7B Mixture-of-Experts Model
Mistral AI released Mixtral 8x7B, a sparse mixture-of-experts model that achieves performance comparable to much larger models while maintaining efficiency. The model uses 8 expert networks, activating only 2 per token for computational efficiency.
Warning
Nov 14, 2023
GraphCast Achieves Superior Weather Forecasting
DeepMind released GraphCast, a graph neural network model that predicts weather globally at 0.25-degree resolution in under 1 minute. The model outperformed the European Centre for Medium-Range Weather Forecasts (ECMWF) on 90% of evaluated meteorological variables, producing predictions that took traditional systems 10 minutes to calculate.
Info
Nov 6, 2023
GPT-4 Turbo with 128K Context
OpenAI released GPT-4 Turbo with a 128,000 token context window, 4x the original GPT-4 context. The model also features reduced hallucination rates and lower API costs compared to previous versions.
Warning
Sep 6, 2023
Technology Innovation Institute Releases Falcon 180B
The Technology Innovation Institute (TII) released Falcon 180B, a 180-billion parameter open-source language model trained on 3.5 trillion tokens. At release, it was the largest openly available language model, surpassing Llama 2 on multiple benchmarks including MMLU, LAMBADA, and HellaSwag.
Info
Jul 18, 2023
Meta Llama 2 Open-Source Release
Meta released Llama 2, a family of open-source language models ranging from 7B to 70B parameters. Made freely available for research and commercial use, with models trained on 2 trillion tokens of public data.
Warning
Jul 11, 2023
Claude 2 Language Model Release
Anthropic released Claude 2, a significantly improved version with longer context (100K tokens), better performance on complex reasoning tasks, and improved safety properties. The model set new benchmarks for instruction following and factuality.
Warning
2022 — Multimodal & Generative
Mar 14, 2023
GPT-4 Launch
OpenAI released GPT-4, a multimodal model accepting both text and image inputs. It demonstrated significant improvements in reasoning, safety, and reliability compared to GPT-3.5, with performance surpassing human experts on many professional benchmarks.
Info
Nov 30, 2022
ChatGPT Public Launch
OpenAI released ChatGPT to the public, a conversational interface powered by GPT-3.5. It reached 1 million users in 5 days and 100 million in 2 months, becoming the fastest-growing application in history.
Info
Sep 21, 2022
OpenAI Open-Sources Whisper Speech Recognition
OpenAI released Whisper, an open-source automatic speech recognition (ASR) model trained on 680,000 hours of multilingual and multitask supervised data collected from the web. The model achieves 50% fewer errors than current specialized models and supports transcription in 96 languages.
Warning
Aug 22, 2022
Stable Diffusion Public Release
Stability AI released Stable Diffusion, an open-source text-to-image generation model. Available under an open license, it could run on consumer hardware and sparked a wave of creative applications and fine-tuned variants.
Warning
Jul 11, 2022
BigScience Releases BLOOM 176B Multilingual Model
The BigScience collaborative initiative released BLOOM, a 176-billion parameter open-source language model trained across 46 natural languages and 13 programming languages. At the time of release, it was the largest openly available language model in existence.
Info
Apr 6, 2022
DALL-E 2 Image Generation Model
OpenAI released DALL-E 2, a significantly improved text-to-image model with better understanding of natural language prompts and higher image quality. The model demonstrated zero-shot generalization to novel concepts and creative variations.
Warning
2021 — Protein & Scale
Dec 1, 2020
AlphaFold 2 Solves Protein Folding
DeepMind's AlphaFold 2 solved the protein folding problem, predicting 3D protein structures to near-experimental accuracy at the CASP14 competition. The breakthrough came from combining attention mechanisms with evolutionary biology insights. The achievement later contributed to the 2024 Nobel Prize in Chemistry awarded to David Baker, Demis Hassabis, and John Jumper.
Info
2020 — Scaling & Few-Shot
Jun 11, 2020
GPT-3 Language Model Breakthrough
OpenAI published GPT-3, a 175-billion parameter language model demonstrating few-shot learning across diverse tasks without task-specific fine-tuning. The model showed emergent abilities like chain-of-thought reasoning and simple code generation.
Info
2017 — Transformers
Jun 12, 2017
"Attention Is All You Need" Transformer Paper
Google researchers published "Attention Is All You Need," introducing the Transformer architecture built entirely on attention mechanisms. This paper became one of the most cited in machine learning, fundamentally changing how neural networks are designed.
Info
2016 — Deep Reasoning
Mar 9, 2016
AlphaGo Defeats Lee Sedol
DeepMind's AlphaGo defeated world champion Lee Sedol in a 5-game match of Go, winning 4-1. Using deep neural networks combined with tree search, AlphaGo exhibited intuitive play and strategic understanding previously thought impossible for machines.
Info
2014 — Generative Models
Jun 10, 2014
Generative Adversarial Networks (GANs) Introduced
Ian Goodfellow and collaborators introduced Generative Adversarial Networks, a framework where two neural networks compete—one generating data and one discriminating real from fake. This sparked a revolution in generative modeling and unsupervised learning.
Info
2012 — Deep Learning Revolution
Sep 30, 2012
AlexNet Wins ImageNet Competition
A deep convolutional neural network called AlexNet won the ImageNet Large Scale Visual Recognition Challenge with a top-5 error rate of 15.4%, far exceeding traditional computer vision approaches at 26.2%. The win sparked the deep learning revolution in vision.
Info