Foundations — 1936 – 1969
1936
Alan Turing Defines Computation
British mathematician Alan Turing published "On Computable Numbers," introducing the concept of a universal machine that could simulate any computation. This theoretical framework laid the mathematical bedrock for every computer and every AI system that would follow.
1950
Turing Proposes the Imitation Game
Turing published "Computing Machinery and Intelligence," posing the question "Can machines think?" and proposing a practical test (now called the Turing Test) to evaluate machine intelligence. He predicted that by the year 2000 machines would fool human judges 30% of the time.
1956
The Dartmouth Workshop — AI Is Born
John McCarthy, Marvin Minsky, Nathaniel Rochester, and Claude Shannon organised the Dartmouth Summer Research Project on Artificial Intelligence. The workshop coined the term "artificial intelligence" and established AI as a distinct academic discipline. McCarthy later founded the Stanford AI Laboratory.
1957
Frank Rosenblatt Builds the Perceptron
At Cornell, Frank Rosenblatt built the Mark I Perceptron, the first machine that could learn from examples. Funded by the US Navy, it was a hardware neural network that learned to classify simple visual patterns using adaptive weights.
1958
McCarthy Invents LISP
John McCarthy at MIT created LISP (List Processing), a programming language built on lambda calculus with features like recursion, dynamic typing, and garbage collection. It became the standard language for AI research for the next three decades.
1966
ELIZA — The First Chatbot
Joseph Weizenbaum at MIT created ELIZA, a simple pattern-matching program that simulated a psychotherapist. Despite using no real understanding, many users became emotionally attached to it, revealing humanity's readiness to attribute intelligence to machines.
1969
Minsky & Papert Publish "Perceptrons"
Marvin Minsky and Seymour Papert published "Perceptrons," a mathematical analysis proving that single-layer perceptrons could not learn certain functions (like XOR). The book was widely interpreted as a death blow to neural network research.
Knowledge Systems & AI Winters — 1970 – 1992
1973
The Lighthill Report Triggers the First AI Winter
British mathematician James Lighthill published a devastating government-commissioned report concluding that AI had failed to deliver on its promises. The UK government cut nearly all AI funding, and the report influenced funders worldwide.
1980
John Searle's Chinese Room Argument
Philosopher John Searle at UC Berkeley proposed the Chinese Room thought experiment, arguing that a computer manipulating symbols according to rules does not truly "understand" anything. The argument challenged whether symbol-processing AI could ever achieve genuine intelligence.
1982
Japan's Fifth Generation Computer Project
Japan's Ministry of International Trade and Industry (MITI) launched the Fifth Generation Computer Systems project, a $400 million national initiative to build intelligent computers using logic programming and parallel processing. It aimed to achieve conversational AI and expert reasoning within a decade.
1986
Backpropagation Revives Neural Networks
David Rumelhart, Geoffrey Hinton, and Ronald Williams published a clear, practical method for training multi-layer neural networks using backpropagation of errors. Though the algorithm had been discovered earlier, this paper demonstrated it could learn useful internal representations.
1988
Hans Moravec's Paradox
Roboticist Hans Moravec at Carnegie Mellon articulated what became known as Moravec's Paradox: high-level reasoning (chess, logic) is computationally cheap for machines, but low-level sensorimotor skills (walking, catching a ball) are enormously hard.
1989
Yann LeCun's Convolutional Neural Networks
Yann LeCun at Bell Labs demonstrated that convolutional neural networks (CNNs) trained with backpropagation could recognise handwritten digits with high accuracy. The system was deployed commercially to read ZIP codes on US mail.
1991
Hochreiter Identifies the Vanishing Gradient Problem
Sepp Hochreiter's diploma thesis formally identified the vanishing gradient problem — the mathematical reason deep neural networks were failing to learn. Gradients shrank exponentially through layers, making training beyond a few layers impractical.
Learning Machines — 1997 – 2012
1997
Deep Blue Defeats Garry Kasparov
IBM's Deep Blue defeated world chess champion Garry Kasparov in a six-game match. The system used brute-force search evaluating 200 million positions per second, combined with hand-tuned evaluation functions and an opening book crafted by grandmasters.
1997
LSTM Networks Solve Long-Range Dependencies
Sepp Hochreiter and Jürgen Schmidhuber published Long Short-Term Memory (LSTM), a recurrent neural network architecture with gated memory cells that could learn to store, retrieve, and forget information over long sequences.
2004
Canada Bets on Neural Networks When Nobody Else Would
The Canadian Institute for Advanced Research (CIFAR) launched its Neural Computation and Adaptive Perception programme, providing sustained funding to Geoffrey Hinton, Yoshua Bengio, and Yann LeCun when neural network research was deeply unfashionable.
2006
Hinton Cracks Deep Learning with Pretraining
Geoffrey Hinton at the University of Toronto published a breakthrough method for training deep neural networks using layer-by-layer unsupervised pretraining followed by fine-tuning. For the first time, networks with many layers could be trained effectively.
2009
Fei-Fei Li Creates ImageNet
Stanford professor Fei-Fei Li and her team published ImageNet, a dataset of 14 million hand-labelled images across 20,000+ categories. The associated annual competition (ILSVRC) became the benchmark that drove computer vision forward.
2011
Google Brain Is Founded
Jeff Dean and Andrew Ng launched Google Brain, a deep learning research project within Google. Using 16,000 CPU cores across 1,000 machines, the team trained a neural network that learned to detect cats in YouTube videos without being told what a cat was.
2012
AlexNet Wins ImageNet — Deep Learning's "Big Bang"
Alex Krizhevsky, Ilya Sutskever, and Geoffrey Hinton at the University of Toronto trained AlexNet, a deep convolutional neural network, on GPUs. It won the ImageNet competition with a top-5 error rate of 15.3% — crushing the runner-up at 26.2%.
The Deep Learning Explosion — 2013 – 2019
2014
Google Acquires DeepMind for $500M
Google acquired London-based DeepMind Technologies for approximately $500 million. Founded by Demis Hassabis, Shane Legg, and Mustafa Suleiman, DeepMind's explicit mission was to "solve intelligence."
2014
Goodfellow Invents Generative Adversarial Networks
Ian Goodfellow at the University of Montreal invented GANs — a framework where two neural networks compete, one generating data and one judging it. The idea reportedly came to him during a conversation at a bar.
2016
AlphaGo Defeats Lee Sedol
DeepMind's AlphaGo defeated world Go champion Lee Sedol 4-1 in Seoul. Go has more possible positions than atoms in the universe, making brute-force search impossible. AlphaGo combined deep neural networks with Monte Carlo tree search to develop intuitive, human-like play.
2017
"Attention Is All You Need" — The Transformer
Eight researchers at Google published the transformer architecture, replacing recurrence entirely with self-attention mechanisms. The paper's deceptively simple title belied its revolutionary impact.
2017
China Publishes Its "New Generation AI Development Plan"
China's State Council released a national strategy aiming to make China the world leader in AI by 2030, with a domestic AI industry worth $150 billion. The plan committed massive government funding and integrated AI into education at all levels.
2017
AlphaZero Learns Chess, Go, and Shogi from Scratch
DeepMind's AlphaZero mastered chess, Go, and shogi (Japanese chess) in hours, starting from only the rules — no human games, no human knowledge, no opening books.
2018
BERT and GPT — Two Paths to Language Understanding
In 2018, Google released BERT (bidirectional pretraining) and OpenAI released GPT-1 (autoregressive pretraining), two competing approaches to making language models understand context.
2019
UAE's Technology Innovation Institute Launches Major AI Push
The United Arab Emirates established the Technology Innovation Institute (TII) and began building what would become the Falcon series of large language models.
2019
GPT-2 — "Too Dangerous to Release"
OpenAI trained GPT-2, a 1.5-billion parameter language model that generated remarkably coherent text. They initially withheld the full model, citing concerns about misuse — the first major public debate about whether AI capabilities should be freely shared.
The Scaling Era — 2020 – 2023
June 2020
GPT-3 Demonstrates Emergent Abilities
OpenAI released GPT-3, a 175-billion parameter model that could perform tasks it was never explicitly trained for — translation, code generation, arithmetic — simply from a few examples in its prompt.
December 2020
AlphaFold 2 Solves Protein Folding
DeepMind's AlphaFold 2 solved the 50-year-old protein folding problem, predicting 3D structures of proteins to near-experimental accuracy. It later predicted the structure of nearly every known protein — over 200 million structures.
2021
Israel Emerges as a Dense AI Hub
AI21 Labs released Jurassic-1, a 178-billion parameter model rivalling GPT-3, built in Tel Aviv. Israel — with more AI startups per capita than any other nation — demonstrated disproportionate impact in AI development.
2022
Chinchilla Scaling Laws Rewrite the Rules
DeepMind researchers published findings showing that most large language models were significantly undertrained. Their "Chinchilla" model, with 70 billion parameters trained on 1.4 trillion tokens, outperformed the 280-billion parameter Gopher.
November 2022
ChatGPT Goes Viral — AI Enters Mainstream Consciousness
OpenAI released ChatGPT, a conversational interface to GPT-3.5. It reached 1 million users in 5 days and 100 million in 2 months — the fastest-growing application in history.
March 2023
GPT-4 — Multimodal, Near-Expert Performance
OpenAI released GPT-4, a multimodal model that could process text and images. It passed the bar exam in the 90th percentile, scored in the top percentiles on AP exams, and demonstrated reasoning abilities that led some researchers to publish papers about "sparks of AGI."
2023
Mistral AI — Europe's Frontier Lab
Three former Google DeepMind and Meta researchers founded Mistral AI in Paris. Within months, they released Mixtral 8x7B, an open-source mixture-of-experts model that rivalled GPT-3.5.
The AGI Horizon — 2024 – Present
January 2025
DeepSeek R1 — China's Open-Source Reasoning Shock
Chinese lab DeepSeek released R1, an open-source reasoning model that matched proprietary frontier models at a fraction of the training cost. The model demonstrated strong chain-of-thought reasoning and was released with full weights.
September 2024
OpenAI o1 — Inference-Time Reasoning
OpenAI released o1, a model that spends more time "thinking" before answering. Using reinforcement learning to develop internal reasoning chains, o1 achieved state-of-the-art results on mathematics, coding, and scientific reasoning benchmarks.
October 2024
Claude Gets Computer Use — AI Agents Arrive
Anthropic released Claude 3.5 Sonnet with the ability to see, understand, and control a computer screen. For the first time, an AI could autonomously navigate software, fill forms, write documents, and execute multi-step workflows across applications.
2024
India Becomes the World's Largest AI Talent Pipeline
India surpassed the US and China in the number of AI and machine learning engineers, with its IIT system and tech ecosystem producing more AI practitioners than any other country.
March 2025
Claude Opus 4 and Sonnet 4 — Sustained Reasoning at Scale
Anthropic released Claude Opus 4 and Sonnet 4, models capable of sustained, multi-hour agentic work — coding, research, and analysis across complex, multi-step tasks.
March 2026
GPT-5.4 — 1 Million Token Context, Unified Capabilities
OpenAI released GPT-5.4 with a 1-million token context window, unifying previously separate reasoning, coding, and general capabilities into a single model. The system demonstrated 33% fewer factual errors and improved agentic workflow completion.
The Question That Remains
What's Still Missing?
Despite extraordinary progress, researchers continue to debate what's still needed for AGI. Open challenges include: genuine causal reasoning (not just pattern matching), persistent memory and learning from experience, embodied intelligence and physical world understanding, robust common sense, goal-setting and autonomous motivation.