A short history of AI

Download as pdf or txt
Download as pdf or txt
You are on page 1of 12

12/26/24, 7:46 PM A short history of AI

Menu Try for free My Economist

Schools brief | Artificial intelligence

A short history of AI
In the first of six weekly briefs, we ask how
AI overcame decades of underdelivering

https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai 1/12
12/26/24, 7:46 PM A short history of AI

image: mike haddad

https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai 2/12
12/26/24, 7:46 PM A short history of AI

Jul 16th 2024 Share

https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai 3/12
12/26/24, 7:46 PM A short history of AI

Listen to this story.

0:00 / 0:00

O ver the summer of 1956 a small but illustrious group gathered at


Dartmouth College in New Hampshire; it included Claude Shannon, the
begetter of information theory, and Herb Simon, the only person ever to win both
the Nobel Memorial Prize in Economic Sciences awarded by the Royal Swedish
Academy of Sciences and the Turing Award awarded by the Association for
Computing Machinery. They had been called together by a young researcher,
John McCarthy, who wanted to discuss “how to make machines use language,
form abstractions and concepts” and “solve kinds of problems now reserved for
humans”. It was the first academic gathering devoted to what McCarthy dubbed
“artificial intelligence”. And it set a template for the field’s next 60-odd years in
coming up with no advances on a par with its ambitions.

In this series on artificial intelligence


1. A short history of ai*
2. Running out of data
3. Controlling the chip supply
4. Fashionable models
Schools
5. Future applications
brief
6. Risks and regulations
July 20th
2024
The Dartmouth meeting did not mark the beginning of scientific inquiry into
https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai 4/12
12/26/24, 7:46 PM A short history of AI

→ A short which could think like people. Alan Turing, for whom the Turing prize
machines
history of
is named,
AI wondered about it; so did John von Neumann, an inspiration to
McCarthy. By 1956 there were already a number of approaches to the issue;
historians think one of the reasons McCarthy coined the term artificial
intelligence, later ai, for his project was that it was broad enough to encompass
them all, keeping open the question of which might be best. Some researchers
favoured systems based on combining facts about the world with axioms like
those of geometry and symbolic logic so as to infer appropriate responses; others
preferred building systems in which the probability of one thing depended on the
constantly updated probabilities of many others.

→ Break on through to the other side


Global corporate Notable machine-
AI investment learning models 51
$337bn By sector
Industry
Others 23
15
Academia
2013 2023 2003 2015 2023

AI conference
attendance, ’000

1984 86 88 1990 92 94 96 98 2000 02 04 06 08 10 12 14 16 18


AAAI IJCAI NeurIPS ICRA ICML CVPR IROS
So rce Stanford Uni ersit AI Inde report 2024

The following decades saw much intellectual ferment and argument on the topic,
but by the 1980s there was wide agreement on the way forward: “expert systems”
https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai 5/12
y
12/26/24, 7:46 PM
9 g y
A short history of AI
p y
which used symbolic logic to capture and apply the best of human know-how.
The Japanese government, in particular, threw its weight behind the idea of such
systems and the hardware they might need. But for the most part such systems
proved too inflexible to cope with the messiness of the real world. By the late
1980s ai had fallen into disrepute, a byword for overpromising and
underdelivering. Those researchers still in the field started to shun the term.

It was from one of those pockets of perseverance that today’s boom was born. As
the rudiments of the way in which brain cells—a type of neuron—work were
pieced together in the 1940s, computer scientists began to wonder if machines
could be wired up the same way. In a biological brain there are connections
between neurons which allow activity in one to trigger or suppress activity in
another; what one neuron does depends on what the other neurons connected to
it are doing. A first attempt to model this in the lab (by Marvin Minsky, a
Dartmouth attendee) used hardware to model networks of neurons. Since then,
layers of interconnected neurons have been simulated in software.
These artificial neural networks are not programmed using explicit rules; instead,
they “learn” by being exposed to lots of examples. During this training the
strength of the connections between the neurons (known as “weights”) are
repeatedly adjusted so that, eventually, a given input produces an appropriate
output. Minsky himself abandoned the idea, but others took it forward. By the
early 1990s neural networks had been trained to do things like help sort the post
by recognising handwritten numbers. Researchers thought adding more layers of
neurons might allow more sophisticated achievements. But it also made the
systems run much more slowly.
A new sort of computer hardware provided a way around the problem. Its
potential was dramatically demonstrated in 2009, when researchers at Stanford
University increased the speed at which a neural net could run 70-fold, using a
gaming pc in their dorm room. This was possible because, as well as the “central
processing unit” (cpu) found in all pcs, this one also had a “graphics processing
unit” (gpu) to create game worlds on screen. And the gpu was designed in a way
suited to running the neural-network code.

C li th t h d d ith
https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai ffi i t t i i l ith t 6/12
12/26/24, 7:46 PM A short history of AI
Coupling that hardware speed-up with more efficient training algorithms meant
that networks with millions of connections could be trained in a reasonable time;
neural networks could handle bigger inputs and, crucially, be given more layers.
These “deeper” networks turned out to be far more capable.
The power of this new approach, which had come to be known as “deep learning”,
became apparent in the ImageNet Challenge of 2012. Image-recognition systems
competing in the challenge were provided with a database of more than a million
labelled image files. For any given word, such as “dog” or “cat”, the database
contained several hundred photos. Image-recognition systems would be trained,
using these examples, to “map” input, in the form of images, onto output in the
form of one-word descriptions. The systems were then challenged to produce
such descriptions when fed previously unseen test images. In 2012 a team led by
Geoff Hinton, then at the University of Toronto, used deep learning to achieve an
accuracy of 85%. It was instantly recognised as a breakthrough.
By 2015 almost everyone in the image-recognition field was using deep learning,
and the winning accuracy at the ImageNet Challenge had reached 96%—better
than the average human score. Deep learning was also being applied to a host of
other “problems…reserved for humans” which could be reduced to the mapping
of one type of thing onto another: speech recognition (mapping sound to text),
face-recognition (mapping faces to names) and translation.
In all these applications the huge amounts of data that could be accessed through
the internet were vital to success; what was more, the number of people using the
internet spoke to the possibility of large markets. And the bigger (ie, deeper) the
networks were made, and the more training data they were given, the more their
performance improved.
Deep learning was soon being deployed in all kinds of new products and services.
Voice-driven devices such as Amazon’s Alexa appeared. Online transcription
services became useful. Web browsers offered automatic translations. Saying
such things were enabled by ai started to sound cool, rather than embarrassing,
though it was also a bit redundant; nearly every technology referred to as ai then
and now actually relies on deep learning under the bonnet.

Chatgpt and its rivals really do seem to “use language


https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai 7/12
12/26/24, 7:46 PM A short history of AI

and form abstractions”


In 2017 a qualitative change was added to the quantitative benefits being
provided by more computing power and more data: a new way of arranging
connections between neurons called the transformer. Transformers enable neural
networks to keep track of patterns in their input, even if the elements of the
pattern are far apart, in a way that allows them to bestow “attention” on particular
features in the data.
Transformers gave networks a better grasp of context, which suited them to a
technique called “self-supervised learning”. In essence, some words are randomly
blanked out during training, and the model teaches itself to fill in the most likely
candidate. Because the training data do not have to be labelled in advance, such
models can be trained using billions of words of raw text taken from the internet.
Mind your language model
Transformer-based large language models (llms) began attracting wider
attention in 2019, when a model called gpt-2 was released by Openai, a startup
(gpt stands for generative pre-trained transformer). Such llms turned out to be
capable of “emergent” behaviour for which they had not been explicitly trained.
Soaking up huge amounts of language did not just make them surprisingly adept
at linguistic tasks like summarisation or translation, but also at things—like
simple arithmetic and the writing of software—which were implicit in the
training data. Less happily it also meant they reproduced biases in the data fed to
them, which meant many of the prevailing prejudices of human society emerged
in their output.

In November 2022 a larger Openai model, gpt-3.5, was presented to the public in
the form of a chatbot. Anyone with a web browser could enter a prompt and get a
response. No consumer product has ever taken off quicker. Within weeks
Chatgpt was generating everything from college essays to computer code. ai had
made another great leap forward.
Where the first cohort of ai-powered products was based on recognition, this
second one is based on generation. Deep-learning models such as Stable
Diffusion and dall-e, which also made their debuts around that time, used a
https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai 8/12
12/26/24, 7:46 PM A short history of AI

technique called diffusion to turn text prompts into images. Other models can
produce surprisingly realistic video, speech or music.
The leap is not just technological. Making things makes a difference. Chatgpt
and rivals such as Gemini (from Google) and Claude (from Anthropic, founded by
researchers previously at Openai) produce outputs from calculations just as other
Explore more
deep-learning systems do. But the fact that they respond to requests with
novelties makes them feel
Artificial intelligence very unlike software which recognises faces, takes
OpenAI
dictation or translates menus. They really do seem to “use language” and “form
abstractions”, just as McCarthy had hoped.
This article appeared in the Schools brief section of the print edition under the headline “A short history of AI ”
This series of briefs will look at how these models work, how much further their
powers can grow, what new uses they will be put to, as well as what they will not,
or should not, be used for. ■ From the July 20th 2024
edition
Discover stories from this section
and more in the list of contents

⇒ Explore the edition

Share Reuse this content

https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai 9/12
12/26/24, 7:46 PM A short history of AI

More from Schools brief

AI needs regulation, but what kind, and how much?


Different countries are taking different approaches to regulating artificial intelligence

LLMs will transform medicine, media and


more
But not without a helping (human) hand

How AI models are getting smarter


Deep neural networks are learning diffusion and other tricks

https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai 10/12
12/26/24, 7:46 PM A short history of AI

The race is on to control the global supply chain for AI chips


The focus is no longer just on faster chips, but on more chips clustered together

AI firms will soon exhaust most of the internet’s data


Can they create more?

Finding living planets


Life evolves on planets. And planets with life evolve

Subscribe Reuse our content

Economist Enterprise Help and contact us

Keep updated

Published since September 1843 to take part in “a severe contest between


intelligence, which presses forward, and an unworthy, timid ignorance
obstructing our progress.”

The Economist The Economist Group

About The Economist Group Working here


Advertise Economist Intelligence Economist Education Courses
Press centre Economist Impact Executive Jobs
SecureDrop Economist Impact Events

To enhance your experience and ensure our website runs smoothly, we use cookies and similar
technologies.

Manage Cookies
https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai 11/12
12/26/24, 7:46 PM A short history of AI

Terms of Use Privacy Cookie Policy Accessibility Modern Slavery Statement Sitemap Your Data Rights

Copyright © The Economist Newspaper Limited 2024. All rights reserved.

https://www.economist.com/schools-brief/2024/07/16/a-short-history-of-ai 12/12

You might also like