The danger of AI is weirder than you think Janelle Shane

So, artificial intelligence

is known for disrupting
all kinds of industries.

What about ice cream?

What kind of mind-blowing
new flavors could we generate

with the power of an advanced
artificial intelligence?

So I teamed up with a group of coders
from Kealing Middle School

to find out the answer to this question.

They collected over 1,600
existing ice cream flavors,

and together, we fed them to an algorithm
to see what it would generate.

And here are some of the flavors
that the AI came up with.

[Pumpkin Trash Break]

(Laughter)

[Peanut Butter Slime]

[Strawberry Cream Disease]

(Laughter)

These flavors are not delicious,
as we might have hoped they would be.

So the question is: What happened?

What went wrong?

Is the AI trying to kill us?

Or is it trying to do what we asked,
and there was a problem?

In movies, when something
goes wrong with AI,

it’s usually because the AI has decided

that it doesn’t want to obey
the humans anymore,

and it’s got its own goals,
thank you very much.

In real life, though,
the AI that we actually have

is not nearly smart enough for that.

It has the approximate computing power

of an earthworm,

or maybe at most a single honeybee,

and actually, probably maybe less.

Like, we’re constantly learning
new things about brains

that make it clear how much our AIs
don’t measure up to real brains.

So today’s AI can do a task
like identify a pedestrian in a picture,

but it doesn’t have a concept
of what the pedestrian is

beyond that it’s a collection
of lines and textures and things.

It doesn’t know what a human actually is.

So will today’s AI
do what we ask it to do?

It will if it can,

but it might not do what we actually want.

So let’s say that you
were trying to get an AI

to take this collection of robot parts

and assemble them into some kind of robot
to get from Point A to Point B.

Now, if you were going to try
and solve this problem

by writing a traditional-style
computer program,

you would give the program
step-by-step instructions

on how to take these parts,

how to assemble them
into a robot with legs

and then how to use those legs
to walk to Point B.

But when you’re using AI
to solve the problem,

it goes differently.

You don’t tell it
how to solve the problem,

you just give it the goal,

and it has to figure out for itself
via trial and error

how to reach that goal.

And it turns out that the way AI tends
to solve this particular problem

is by doing this:

it assembles itself into a tower
and then falls over

and lands at Point B.

And technically, this solves the problem.

Technically, it got to Point B.

The danger of AI is not that
it’s going to rebel against us,

it’s that it’s going to do
exactly what we ask it to do.

So then the trick
of working with AI becomes:

How do we set up the problem
so that it actually does what we want?

So this little robot here
is being controlled by an AI.

The AI came up with a design
for the robot legs

and then figured out how to use them
to get past all these obstacles.

But when David Ha set up this experiment,

he had to set it up
with very, very strict limits

on how big the AI
was allowed to make the legs,

because otherwise …

(Laughter)

And technically, it got
to the end of that obstacle course.

So you see how hard it is to get AI
to do something as simple as just walk.

So seeing the AI do this,
you may say, OK, no fair,

you can’t just be
a tall tower and fall over,

you have to actually, like,
use legs to walk.

And it turns out,
that doesn’t always work, either.

This AI’s job was to move fast.

They didn’t tell it that it had
to run facing forward

or that it couldn’t use its arms.

So this is what you get
when you train AI to move fast,

you get things like somersaulting
and silly walks.

It’s really common.

So is twitching along the floor in a heap.

(Laughter)

So in my opinion, you know what
should have been a whole lot weirder

is the “Terminator” robots.

Hacking “The Matrix” is another thing
that AI will do if you give it a chance.

So if you train an AI in a simulation,

it will learn how to do things like
hack into the simulation’s math errors

and harvest them for energy.

Or it will figure out how to move faster
by glitching repeatedly into the floor.

When you’re working with AI,

it’s less like working with another human

and a lot more like working
with some kind of weird force of nature.

And it’s really easy to accidentally
give AI the wrong problem to solve,

and often we don’t realize that
until something has actually gone wrong.

So here’s an experiment I did,

where I wanted the AI
to copy paint colors,

to invent new paint colors,

given the list like the ones
here on the left.

And here’s what the AI
actually came up with.

[Sindis Poop, Turdly, Suffer, Gray Pubic]

(Laughter)

So technically,

it did what I asked it to.

I thought I was asking it for,
like, nice paint color names,

but what I was actually asking it to do

was just imitate the kinds
of letter combinations

that it had seen in the original.

And I didn’t tell it anything
about what words mean,

or that there are maybe some words

that it should avoid using
in these paint colors.

So its entire world
is the data that I gave it.

Like with the ice cream flavors,
it doesn’t know about anything else.

So it is through the data

that we often accidentally tell AI
to do the wrong thing.

This is a fish called a tench.

And there was a group of researchers

who trained an AI to identify
this tench in pictures.

But then when they asked it

what part of the picture it was actually
using to identify the fish,

here’s what it highlighted.

Yes, those are human fingers.

Why would it be looking for human fingers

if it’s trying to identify a fish?

Well, it turns out that the tench
is a trophy fish,

and so in a lot of pictures
that the AI had seen of this fish

during training,

the fish looked like this.

(Laughter)

And it didn’t know that the fingers
aren’t part of the fish.

So you see why it is so hard
to design an AI

that actually can understand
what it’s looking at.

And this is why designing
the image recognition

in self-driving cars is so hard,

and why so many self-driving car failures

are because the AI got confused.

I want to talk about an example from 2016.

There was a fatal accident when somebody
was using Tesla’s autopilot AI,

but instead of using it on the highway
like it was designed for,

they used it on city streets.

And what happened was,

a truck drove out in front of the car
and the car failed to brake.

Now, the AI definitely was trained
to recognize trucks in pictures.

But what it looks like happened is

the AI was trained to recognize
trucks on highway driving,

where you would expect
to see trucks from behind.

Trucks on the side is not supposed
to happen on a highway,

and so when the AI saw this truck,

it looks like the AI recognized it
as most likely to be a road sign

and therefore, safe to drive underneath.

Here’s an AI misstep
from a different field.

Amazon recently had to give up
on a résumé-sorting algorithm

that they were working on

when they discovered that the algorithm
had learned to discriminate against women.

What happened is they had trained it
on example résumés

of people who they had hired in the past.

And from these examples, the AI learned
to avoid the résumés of people

who had gone to women’s colleges

or who had the word “women”
somewhere in their resume,

as in, “women’s soccer team”
or “Society of Women Engineers.”

The AI didn’t know that it wasn’t supposed
to copy this particular thing

that it had seen the humans do.

And technically, it did
what they asked it to do.

They just accidentally asked it
to do the wrong thing.

And this happens all the time with AI.

AI can be really destructive
and not know it.

So the AIs that recommend
new content in Facebook, in YouTube,

they’re optimized to increase
the number of clicks and views.

And unfortunately, one way
that they have found of doing this

is to recommend the content
of conspiracy theories or bigotry.

The AIs themselves don’t have any concept
of what this content actually is,

and they don’t have any concept
of what the consequences might be

of recommending this content.

So, when we’re working with AI,

it’s up to us to avoid problems.

And avoiding things going wrong,

that may come down to
the age-old problem of communication,

where we as humans have to learn
how to communicate with AI.

We have to learn what AI
is capable of doing and what it’s not,

and to understand that,
with its tiny little worm brain,

AI doesn’t really understand
what we’re trying to ask it to do.

So in other words, we have
to be prepared to work with AI

that’s not the super-competent,
all-knowing AI of science fiction.

We have to be prepared to work with an AI

that’s the one that we actually have
in the present day.

And present-day AI is plenty weird enough.

Thank you.

(Applause)

因此,

人工智能以颠覆
各种行业而闻名。

冰淇淋呢? 借助先进的人工智能,我们可以创造出

什么样的令人兴奋的
新口味

因此,我与 Kealing 中学的一群程序员合作

,找出了这个问题的答案。

他们收集了 1,600 多种
现有的冰淇淋口味

,然后我们一起将它们输入一个算法
,看看它会产生什么。

以下
是人工智能提出的一些口味。

[Pumpkin Trash Break]

(笑声)

[花生酱史莱姆]

[草莓奶油病]

(笑声)

这些味道并不
像我们希望的那样美味。

所以问题是:发生了什么?

什么地方出了错?

人工智能是想杀死我们吗?

还是它试图按照我们的要求去做
,但出现了问题?

在电影中,当
人工智能出现问题时

,通常是因为人工智能已经决定

不再服从人类

,它有自己的目标,
非常感谢。

然而,在现实生活中
,我们实际拥有的人工智能

还不够聪明。

它具有

近似蚯蚓的计算能力,

或者最多可能是一只蜜蜂

,实际上,可能更少。

就像,我们不断地学习
关于大脑的新事物

,这清楚地表明我们的人工智能
在多大程度上无法与真正的大脑相媲美。

所以今天的人工智能可以完成
像识别图片中的行人这样的任务,

但它没有
关于行人是什么的概念,

除了它
是线条、纹理和事物的集合。

它不知道人类究竟是什么。

那么今天的人工智能会
按照我们的要求去做吗?

如果可以的话,它会,

但它可能不会做我们真正想要的。

因此,假设
您试图让 AI

收集这些机器人零件

并将它们组装成某种机器人
,以便从 A 点到达 B 点。

现在,如果您要尝试

通过编写一个 传统风格的
计算机程序,

你会
一步一步地给程序

说明如何获取这些部件,

如何将它们组装
成一个有腿的机器人

,然后如何使用这些腿
走到 B 点。

但是当你' 重新使用人工智能
来解决问题,

它会有所不同。

你不告诉它
如何解决问题,

你只是给它一个目标

,它必须通过反复试验自己弄清楚

如何达到这个目标。

事实证明,人工智能
倾向于解决这个特定问题的方式

是这样做:

它将自己组装成一座塔
,然后翻倒

并降落在 B 点。从

技术上讲,这解决了问题。

从技术上讲,它达到了 B 点。

人工智能的危险不在于
它会反抗我们,

而在于它会
完全按照我们的要求去做。

那么
使用 AI 的技巧就变成了:

我们如何设置
问题,让它真正做我们想要的?

所以这里的这个小机器人
是由人工智能控制的。

AI 提出了机器人腿的设计

,然后想出了如何使用它们
来克服所有这些障碍。

但是当大卫哈设置这个实验时,

他必须
非常非常严格地限制

允许人工智能制造腿的大小,

因为否则……

(笑声)从

技术上讲,它已经
到了最后 那个障碍课程。

所以你会看到让
人工智能做一些像走路这样简单的事情是多么困难。

所以看到人工智能这样做,
你可能会说,好吧,不公平,

你不能只是
一个高塔然后摔倒,

你实际上必须,比如,
用腿走路。

事实证明,
这也并不总是有效。

这个人工智能的工作是快速移动。

他们没有告诉它它必须
面向前方奔跑

或者它不能使用它的手臂。

所以
当你训练人工智能快速移动时,

你会得到这样的结果,你会得到翻筋斗
和愚蠢的走路之类的东西。

这真的很常见。

地板上成堆的抽搐也是如此。

(笑声)

所以在我看来,你知道

“终结者”机器人应该更奇怪。

如果你给它一个机会,人工智能会做的另一件事是黑客攻击“黑客帝国”。

因此,如果您在模拟中训练 AI,

它将学习如何做一些事情,例如
破解模拟的数学错误

并收集它们以获取能量。

或者它会通过反复撞到地板来弄清楚如何更快地移动

当你使用人工智能时,

它不像是和另一个人一起工作,

而更像是
和某种奇怪的自然力量一起工作。

不小心
给人工智能解决了错误的问题真的很容易,

而且我们通常
直到出现问题时才意识到这一点。

所以这是我做的一个实验

,我希望
人工智能复制油漆颜色

,发明新的油漆颜色,

给定的列表就像左边的列表一样

这就是人工智能
真正想出的。

[Sindis Poop, Turdly, Suffer, Grey Pubic]

(笑声

)从技术上讲,

它按照我的要求做了。

我以为我在要求它,
比如漂亮的油漆颜色名称,

但我实际上要求它做

的只是模仿

它在原版中看到的各种字母组合。

而且我没有告诉它任何
单词的含义,

或者它可能

应该避免
在这些油漆颜色中使用某些单词。

所以它的整个世界
就是我给它的数据。

就像冰淇淋口味一样,
它不知道其他任何东西。

因此,我们经常通过数据

不小心告诉
AI 做错事。

这是一种叫做丁鱼的鱼。

还有一组研究

人员训练了一个人工智能来识别
图片中的这种鲱鱼。

但是当他们问

它实际使用图片的哪个部分
来识别鱼时,

它突出显示了这部分内容。

是的,那是人类的手指。

如果它试图识别一条鱼,它为什么要寻找人类的手指?

好吧,事实证明,丁鲷
是一种奖杯鱼

,所以在
AI 在训练期间看到的很多关于这条鱼的图片中

,这条鱼看起来是这样的。

(笑声

) 它不知道手指
不是鱼的一部分。

所以你明白为什么
设计一个

能够真正理解
它所看到的东西的人工智能是如此困难。

这就是为什么

在自动驾驶汽车中设计图像识别如此困难,

以及为什么如此多的自动驾驶汽车失败

是因为人工智能被混淆了。

我想谈谈 2016

年的一个例子。有人
在使用特斯拉的自动驾驶 AI 时发生了一起致命事故,

但他们没有
像设计的那样在高速公路上

使用它,而是在城市街道上使用它。

发生的事情是,

一辆卡车从汽车前面开出
,汽车没有刹车。

现在,人工智能肯定被训练
来识别图片中的卡车。

但看起来发生的事情

是,人工智能被训练来识别
高速公路行驶中的卡车

,你会期望
从后面看到卡车。

侧面的卡车不
应该出现在高速公路上

,所以当 AI 看到这辆卡车时,

看起来 AI 认为
它最有可能是一个路标

,因此可以安全地在下面行驶。

这是
来自不同领域的人工智能失误。

亚马逊最近不得不放弃他们正在
研究的一种简历排序算法

当时他们发现该算法
已经学会了歧视女性。

发生的事情是,他们已经
根据

他们过去雇用的人的简历样本对其进行了培训。

从这些例子中,人工智能学会
了避免

那些上过女子大学

或在简历中某处有“女性”一词
的人的简历,

例如“女子足球队”
或“女性工程师协会”。

人工智能不知道它不
应该复制

它看到人类所做的这种特殊事情。

从技术上讲,它做
了他们要求它做的事情。

他们只是不小心要求
它做错事。

这种情况一直发生在人工智能中。

人工智能可以是真正具有破坏性的
并且不知道它。

因此
,在 Facebook 和 YouTube 中推荐新内容的 AI

进行了优化,以
增加点击次数和观看次数。

不幸的是
,他们发现这样做的一种方法

是推荐
阴谋论或偏执的内容。

AI 本身
对这个内容实际上是

什么没有任何概念,并且他们

对推荐这个内容可能产生的后果没有任何概念。

因此,当我们使用 AI 时,我们有责任

避免出现问题。

避免事情出错,

这可能归结
为古老的沟通问题

,我们人类必须学习
如何与人工智能沟通。

我们必须了解
人工智能能做什么和不能做什么,

并且要理解
,人工智能凭借其微小的蠕虫大脑

并不能真正
理解我们试图要求它做什么。

所以换句话说,我们
必须准备好与

不是科幻小说中的超级能力、
无所不知的人工智能合作。

我们必须准备好

使用我们今天实际拥有
的人工智能。

现在的人工智能已经够奇怪了。

谢谢你。

(掌声)