What Defines Artificial Intelligence? The Complete WIRED …

Artificial intelligence is overhypedthere, we said it. Its also incredibly important.

Superintelligent algorithms arent about to take all the jobs or wipe out humanity. But software has gotten significantly smarter of late. Its why you can talk to your friends as an animated poop on the iPhone X using Apples Animoji, or ask your smart speaker to order more paper towels.

Tech companies heavy investments in AI are already changing our lives and gadgets, and laying the groundwork for a more AI-centric future.

The current boom in all things AI was catalyzed by breakthroughs in an area known as machine learning. It involves training computers to perform tasks based on examples, rather than by relying on programming by a human. A technique called deep learning has made this approach much more powerful. Just ask Lee Sedol, holder of 18 international titles at the complex game of Go. He got creamed by software called AlphaGo in 2016.

For most of us, the most obvious results of the improved powers of AI are neat new gadgets and experiences such as smart speakers, or being able to unlock your iPhone with your face. But AI is also poised to reinvent other areas of life. One is health care. Hospitals in India are testing software that checks images of a persons retina for signs of diabetic retinopathy, a condition frequently diagnosed too late to prevent vision loss. Machine learning is vital to projects in autonomous driving, where it allows a vehicle to make sense of its surroundings.

Theres evidence that AI can make us happier and healthier. But theres also reason for caution. Incidents in which algorithms picked up or amplified societal biases around race or gender show that an AI-enhanced future wont automatically be a better one.

The Beginnings of Artificial Intelligence

Artificial intelligence as we know it began as a vacation project. Dartmouth professor John McCarthy coined the term in the summer of 1956, when he invited a small group to spend a few weeks musing on how to make machines do things like use language. He had high hopes of a breakthrough toward human-level machines. We think that a significant advance can be made, he wrote with his co-organizers, if a carefully selected group of scientists work on it together for a summer.

Moments that Shaped AI

1956

The Dartmouth Summer Research Project on Artificial Intelligence coins the name of a new field concerned with making software smart like humans.

1965

Joseph Weizenbaum at MIT creates Eliza, the first chatbot, which poses as a psychotherapist.

1975

Meta-Dendral, a program developed at Stanford to interpret chemical analyses, makes the first discoveries by a computer to be published in a refereed journal.

1987

A Mercedes van fitted with two cameras and a bunch of computers drives itself 20 kilometers along a German highway at more than 55 mph, in an academic project led by engineer Ernst Dickmanns.

1997

IBMs computer Deep Blue defeats chess world champion Garry Kasparov.

2004

The Pentagon stages the Darpa Grand Challenge, a race for robot cars in the Mojave Desert that catalyzes the autonomous-car industry.

2012

Researchers in a niche field called deep learning spur new corporate interest in AI by showing their ideas can make speech and image recognition much more accurate.

2016

AlphaGo, created by Google unit DeepMind, defeats a world champion player of the board game Go.

Those hopes were not met, and McCarthy later conceded that he had been overly optimistic. But the workshop helped researchers dreaming of intelligent machines coalesce into a proper academic field.

Early work often focused on solving fairly abstract problems in math and logic. But it wasnt long before AI started to show promising results on more human tasks. In the late 1950s Arthur Samuel created programs that learned to play checkers. In 1962 one scored a win over a master at the game. In 1967 a program called Dendral showed it could replicate the way chemists interpreted mass-spectrometry data on the makeup of chemical samples.

As the field of AI developed, so did different strategies for making smarter machines. Some researchers tried to distill human knowledge into code or come up with rules for tasks like understanding language. Others were inspired by the importance of learning to human and animal intelligence. They built systems that could get better at a task over time, perhaps by simulating evolution or by learning from example data. The field hit milestone after milestone, as computers mastered more tasks that could previously be done only by people.

Deep learning, the rocket fuel of the current AI boom, is a revival of one of the oldest ideas in AI. The technique involves passing data through webs of math loosely inspired by how brain cells work, known as artificial neural networks. As a network processes training data, connections between the parts of the network adjust, building up an ability to interpret future data.

Artificial neural networks became an established idea in AI not long after the Dartmouth workshop. The room-filling Perceptron Mark 1 from 1958, for example, learned to distinguish different geometric shapes, and got written up in The New York Times as the Embryo of Computer Designed to Read and Grow Wiser. But neural networks tumbled from favor after an influential 1969 book co-authored by MITs Marvin Minsky suggested they couldnt be very powerful.

Not everyone was convinced, and some researchers kept the technique alive over the decades. They were vindicated in 2012, when a series of experiments showed that neural networks fueled with large piles of data and powerful computer chips could give machines new powers of perception.

View original post here:
What Defines Artificial Intelligence? The Complete WIRED ...

Related Posts
This entry was posted in $1$s. Bookmark the permalink.