Augmented Intelligence

[Music]

30 years ago when i was an undergrad

ai suddenly became really popular there

was this wave of enthusiasm that swept

across the campus flipped across the

business world

it seemed like it was only a matter of

time before we were all going to be made

redundant

then aia faltered and it faded

and just went away now it’s back same

enthusiasm same promises

i wonder how well it’s going to do this

time around

how far it’s going to be able to

actually go

all aa systems have got two parts to

them the first thing an ai system has to

do is figure out where it’s at that

means it takes an input

this is data from sensors cameras and it

builds a model of the world

and based on that model of the world it

has to go ahead and decide

what kind of actions it should take and

decisions it has to make

and hopefully these are good but we look

at the

modeling portion it’s really an

important part of the ai system because

the model allows the ai system to make

predictions if the model is

poor then the predictions will be poor

and the ai system is going to end up

making

bad decisions you end up making mistakes

so here’s a couple of questions to think

about with regards to the ass system the

first one is

do a systems how do they do when they

encounter the unknown

and the second one is do ai systems have

any sense

of being lost or when they’re lost so

i’ve got two examples that help explain

and illustrate these answer these

questions the first example

will help us understand what i mean by

an ai system encountering the unknown

the second example will help us see how

easy it is for an ai system to become

lost so my first example i’m assuming

that i’m going to

build a drone delivery system in new

york city so clearly

one of the most important parts of this

is knowing for every location you can be

in new york city

how high you have to fly to make sure

you don’t run into buildings

so i’ve got a map of lower manhattan

this starts top of it is at the bottom

part of central park it goes down to the

southern tip

and i’ve got a couple of points

highlighted here just landmarks

give you some reference first one is new

york public library it’s by bryant park

the second one is the empire state

building a couple of blocks south from

the new york public library

and then a couple of miles down on the

southern tip is the one world trade

center

now this map is actually 3d data it’s a

3d rendering

of all the buildings in new york it was

done by new york city back in 2014

available on the web it’s it’s a pretty

cool set of data

i used this in order to create my

training data because that’s the first

thing i have to have in order to

build an ai system and to build a

predictive model

now i took 50 000 randomly selected

points

locations and then calculated the

heights but i focused all of these

around new york public library so you

can see that kind of concentric

almost target-like set of circles that

shows you where the data is centered

so most of the data is around the new

york public library a little bit around

the empire state building and there’s no

data down by one world trade center

and what i mean by known and unknown for

an ai system

an ai system is in known territory when

it’s got training data that’s close to

the location

it’s trying to forecast or predict it’s

an unknown territory when that training

data is far away

so known as up by bryant park new york

public library

unknown is down by the one world trade

center

now clearly if i was trying to build a

real drone system

delivery system and ai system for that i

would collect data

all across new york but this is just to

illustrate what happens when we counter

known and unknown

the next thing i need is a model so i’m

going to calibrate a

neural net and i’m going to actually do

this 10 different times

and so what i do is i bring my data into

the computer i push the button

and it goes back and calibrates the no

net and it trains there’s a calcula

bunch of calculations that happen

in the background then it gives me a

formula that can make predictions i can

give any location it will predict the

height of the building

and it’s kind of interesting and really

dirty little secret about

neural nets is every time you push the

button every time you do that regardless

if you have the same data

you’ll get a different formula you’ll

get different predictions so if i do

that 10 different times

i get 10 different predictions and i did

that 10 times so i have 10 neural nets

and i want to show the difference in how

well they do when known territory

versus unknown territory so we’ll start

off by looking

at the new york public library

you can see the red dot there that’s the

top of the new york public library

it’s 190 feet off the ground and if i

take the average of these

10 different predictions it’s not bad

it’s 164 feet

but if i actually plot all 10 of the

different predictions

you’ll see that we have the highest one

at 196 feet the lowest 147 feet

and that range is actually not too bad

given the amount of data i have and i

could arbitrarily shrink that range by

just collecting more data

nearby now

what happens when we move down to a

place that’s unknown territory

one world trade center here the one

world trade center

is 1792 feet above the ground

the average of my 10 neural nets 102

feet

right that’s the little circle with the

blue cross through it

that’s not very good but it gets even

worse

because if you plot all 10 of the

predictions

you’ll see that one of them well you’ve

heard the phrase

off by a mile one of them’s essentially

not only underground but it’s

essentially off by a mile

so first take away from this talk is

that when an ai system encounters the

unknown as far away from its training

data

it’s going to make mistakes mistakes

that could be catastrophic because its

predictions are going to be

catastrophically wrong

now you should say well this is kind of

an easy example in some ways because

it’s easy to know how far away we are

from the trading data

because i only have two numbers that are

coming in the location is just

latitude and longitude right it’s street

address so it’s easy to say that

yes one world trade center is miles away

from where all my training data is

but what happens when your input gets a

lot more complicated that takes us to

our second example

so when you look at images which no nuts

are usually

commonly used in order to do

classifications what you’ve got is

you’ve got

rows and rows and rows of dots and each

one of these dots have got different

numbers

that tell you the color of those dots

and so if you have a high resolution

image you could literally have

millions of inputs this means that the

the model that we calibrate

the formula we come up with is

increased in complexity dramatically and

now he’s increasing complexity

that increasing complexity makes it so

that these models actually are very

fragile

they’re very sensitive to small changes

in data that you might present to it so

let me let me give an example or two

this is our cycle second example

the catch phrase in the academic

literature is adversarial attacks

for these type of examples and what

happens in an adversarial attack

is you imagine that someone’s built a

neural net they’ve

trained it with a bunch of images to do

classification

and then they want to see what small

changes you can do to an image in order

to fool the neural net

so assuming some adversaries trying to

attack this ai system

so the first people are really one of

the first people that really did some of

this work

were a group of researchers in japan and

they put together

a model with a bunch of animals images

and they built a classifier and then

they showed this image of a horse to it

so this image was not in the training

data set

and they asked the classifier to tell

what it was and said it was a horse

then what they went and did is they

actually just changed a single

pixel they said one pixel and they

change it from white to black

and they put this in the classifier and

asked what it was and the classifier

came back and said

that it was a frog so this should be a

little troubling

to you but it shouldn’t be as troubling

as the next example because the next

example

well the next example should be very

troubling to you if you ever think

sometime in the future you might like a

let an ai system take over driving your

car for you so this one focused on

street signs

and classifying street signs like a

bunch of street signs and images of

street signs built a neural net

classifier then it showed it

this image of a stop sign and it was

able to classify it as a stop sign

but then it took the image and just put

a few pieces of tape on it

randomly and asked the neuro net to go

ahead and classify this

and when it came back with the

classification it said it wasn’t a stop

sign it was a go

sign said you go up to 45 miles an hour

now what’s going on in these examples

what’s going on is

their middle the original images are in

some sense

close to the images

that were trained it’s kind of like

being up by bryant park in the drone

example

and the other images we make little

small changes are suddenly far away

and part of the problem with doing image

analysis is really hard to know how far

away two images are because there’s so

many different

parts of the input that you could change

and so

the second takeaway from this talk is

that when

ai systems encounter something new they

don’t know they’ve encountered something

new

ai systems just don’t have a good way of

figuring out when they’re lost

and this is a real problem it’s a

problem we need to think about it’s a

problem that ai’s community

has to find some way of solving so one

obvious solution is to collect more data

but it’s not just

that easy you can’t just go ahead and

say well i’m going to collect more data

because if you collect more data where

you already have data so you collect

more data around bryant park around the

new york public library that doesn’t

help you that much

when you go and encounter something new

like the one world trade center you have

to know when you’ve encountered

something new and then collect the data

what typically happens well what

typically happens is you just do your

best

you build an ai system you collect as

much data as you can and then you run

the ai system

and how do you know that you’ve

encountered something new well there’s a

crash that happens

and i’m not a big fan of this approach

when the ai systems crash

usually that’s bad sometimes it’s

catastrophically bad

and i think we can do a better job i

think we can find a better approach to

that but i have to say that is really

kind of the default approach

that is being used for almost every eye

system out there

so what’s an alternative approach well

alternative approach is to try to mimic

the brain in some way so it’s really

interesting our brain is constantly

making models of the world but what i

find fascinating is it’s actually

doing multiple models so there’s one

model that’s being created by our

brainstem it’s kind of a quick and dirty

model of our world

and there’s another version that’s being

created by our cortex which is more

thoughtful and abstract

and then in the center of our brain

there’s a portion a part of the brain

called the hippocampus and one of its

jobs is to constantly compare these two

models

and see if they line up if they line up

you just kind of carry on

if they don’t line up however you get

this sudden feeling that

that something’s wrong there’s an

anomaly there’s something you need to

focus on

in psychology we call this the orienting

reflex

right so there’s an anomaly and your

brain is a very quick calculation

danger or not if it’s danger you go into

fight-or-flight mode

adrenaline kicks in and

you go into survival mode but if it’s

not dangerous it’s interesting

you still go into a different mode you

go into the explorer mode

because you want to go over and find out

why it is that your two models didn’t

match up you go in there

and collect more data and update your

models

and i think this could offer a really

powerful way

of thinking about how to solve the

problem of helping ai systems

understand when they’re lost

so you could call it maybe the ai

orienting reflex

you run multiple models if the models

agree then you know in your known

territory you carry on

but if the models disagree then you know

you’ve encountered something new

you know that there’s unknown territory

you need to be cautious you need to

actually

slow down right and collect more data

so do ai systems do a good job

when they encounter the unknown no

do a systems know when they’re lost so

far

no is this a problem that we have to

solve absolutely

if we can’t solve this problem there’s a

very real chance the ai

systems will fade just like they did 30

years ago

it’s not just that these ai systems

don’t know when they’re lost

you know that does place a serious limit

on

what we can do with ai systems but

there’s a more fundamental

limit on what ai systems can do

and you’ll notice so far in this talk

i’ve not actually

given a name to ai for most people ai

means artificial intelligence

for me i think that’s misleading i would

prefer to call these systems

augmented intelligence systems and the

reason is that when you think of

artificial intelligence you think of

human cognition values goals

consciousness that’s not what these

systems are what these systems are

at the end of the day really are tools

they’re tools that are powerful they’re

tools that augment our ability

as humans to use our intelligence to

solve problems

and if you think about this difference

it actually will help you understand

the fundamental boundary that ai systems

encounter

and i want to illustrate this by talking

about the problem

or the concept of singularity so the

idea or

maybe it’s better stated the fear of

singularity is eventually

machines are going to get so smart that

they’re going to take over

right from an artificial intelligence

perspective

that’s kind of roughly like saying that

the supercomputer at tesla that does all

the

neural nets and does updates to all the

cars suddenly develops a mind of its own

suddenly

it comes up its own objectives its own

values independent of any human

and who knows what it’ll do maybe it

will go ahead and

and tell the teslas to kick their

drivers out and converge on sturgis

north dakota for a big rally some kind

of big electronic car dance party

from an augmented perspective

intelligence perspective

the singularity risk is different

because

when you have an augmented intelligence

system what it does is it concentrates

more and more control

in the hands of fewer and fewer people

so the risk is something along the lines

of

some teenager in their basement it could

be anywhere it could be russia

china the u.s hacking into that tesla

computer

that supercomputer and taking over and

telling the tesla computer

supercomputer where to send you know to

kick out the

to have the cars kick out their drivers

and go to the big dance party

and that’s the risk i think that in

terms of singularity we should really be

worried about

as a society it’s not the artificial

intelligence risk of the robots taking

over

now is ai here to stay maybe

hopefully if it fades i have no doubt

that it will

most likely come back and the problems

that we’ll solve will be breathtaking

but you shouldn’t be fooled there will

always be a fundamental

limit to ai systems there’ll always be a

boundary and this fundamental

boundary is the ai systems are tools

that means they reflect the values

of the people who created them and who

use them they don’t create their own

values

ultimately the task of determining

what’s valuable what’s moral what’s

ethical

that’s a human task it’s something that

we can never outsource

thank you

[音乐]

30年前还是本科生的时候,

人工智能突然火了

,这股热潮

席卷校园,翻遍

商界

,似乎

我们都被造就只是时间问题了

冗余,

然后 aia 步履蹒跚,它消失了

,然后就消失了,现在它又恢复了同样的

热情,同样的

承诺 人工智能系统

要做的就是弄清楚它在哪里,这

意味着它需要一个输入,

这是来自传感器相机的数据,它

建立了一个世界模型,

并基于这个世界模型,它

必须继续并决定

哪种类型 它应该采取的行动和

必须做出的决定

,希望这些都是好的,但我们看看

建模部分,它确实是

人工智能系统的重要组成部分,因为

该模型允许人工智能系统做出

预测,如果 模型

很差,那么预测也会很差

,人工智能系统最终会做出

错误的决定,你最终会犯错误,

所以这里有几个关于屁股系统需要考虑的问题,

第一个是

如何做一个系统 当他们遇到未知时他们会这样做吗?

第二个是人工智能系统是否

有迷路的感觉或者当他们迷路时所以

我有两个例子可以帮助解释

和说明这些

问题的答案第一个例子

将帮助我们 理解我所说的

遇到未知的人工智能系统的意思

第二个例子将帮助我们

了解人工智能系统

丢失的容易程度所以我的第一个例子我

假设我要

建立一个无人机交付系统在新的

很明显,纽约市

最重要的部分之一

就是知道你

在纽约市的每个位置

你必须飞多高才能确保

你不会撞到建筑物

所以我有一张曼哈顿下城的地图

这从顶部开始 它位于中央公园的底部

,一直延伸到

南端

,我在

这里突出显示了几个点,只是地标

给你一些参考第一个是

纽约公共图书馆,它是布莱恩特公园

的第二个是 帝国

大厦在纽约公共图书馆以南几个街区

,然后在南端几英里

处是一个世界贸易

中心,

现在这张地图实际上是 3d 数据它是

纽约所有建筑物的 3d 渲染它

早在 2014 年由纽约市完成,

可在网络上获得,这是一组非常

酷的数据,

我使用它来创建我的

训练数据,因为这是

我必须拥有的第一件事,以便

构建人工智能系统并构建 现在是一个

预测模型,

我随机选择了 50 000

个点

位置,然后计算了

高度,但我将所有这些

都集中在纽约公共图书馆周围,这样你

就可以看到那种同心的

几乎像目标一样的圆集

向您显示数据的中心位置,

因此大部分数据都在

纽约公共图书馆周围,

在帝国大厦周围一点点,没有

数据被一个世界贸易中心记录下来,

我所说的人工智能系统的已知和未知是什么意思

它的训练数据接近

它试图预测的位置时,人工智能系统处于已知区域,或者

当训练数据距离很远时,它是一个未知区域,

因此被布莱恩特公园纽约

公共图书馆称为

未知区域。 世界贸易

中心

现在很清楚,如果我试图建立一个

真正的无人机系统

交付系统和人工智能系统,我

会收集

整个纽约的数据,但这只是为了

说明当我们对抗已知和未知时会发生

什么我需要的下一件事是 一个模型,所以我

要校准一个

神经网络,我要实际

做 10 次不同的时间

,所以我要做的就是把我的数据

带入计算机我按下按钮

,它会返回并 c 校准无

网并训练在后台

发生的一系列计算

然后它给了我

一个可以进行预测的公式我可以

给出任何位置它将预测

建筑物

的高度它有点有趣而且非常

肮脏 关于

神经网络的秘密是每次你按下

按钮时,无论

你是否有相同的数据,

你都会得到不同的公式,你会

得到不同的预测,所以如果我这样

做 10 次不同的时间,

我会得到 10 个不同的预测和 我这样做

了 10 次,所以我有 10 个神经网络

,我想

展示它们在已知领域

与未知领域时的表现差异,所以我们将从

纽约公共图书馆开始,

你可以看到那里的红点 那

是纽约公共图书馆的顶部,

它离地面 190 英尺,如果我

取这

10 个不同预测的平均值,

它是 164 英尺还不错,

但如果我真的绘制了所有 10 个

差异 租金预测

你会看到我们有最高

的 196 英尺,最低的 147 英尺

考虑到我拥有的数据量,这个范围实际上并不算太差,我

可以通过在附近收集更多数据来任意缩小这个范围

现在

会发生什么 当我们向下移动到一个

未知领域的地方时,

这里的一个世界贸易中心 一个

世界贸易中心

离地面 1792 英尺

我的 10 个神经网络的平均值 102

英尺

右边是带有蓝色十字的小圆圈

,不是很好 但它变得

更糟,

因为如果你绘制所有 10 个

预测,

你会发现其中一个很好,你已经

听过这个短语

,其中一个基本上

不仅在地下,而且

基本上偏离了一英里,

所以首先采取 远离这个话题的是

,当人工智能系统遇到

远离其训练数据的未知数时,

它会犯错误

,这可能是灾难性的,因为它的

预测将是

灾难性的 strophically wrong

now you should say well 这是

一个简单的例子,因为

很容易知道我们离交易数据有多远,

因为我只有两个数字,

位置只是

纬度和经度,对吧 街道

地址,所以很容易说,

是的,一个世界贸易中心离

我所有的训练数据所在的地方只有几英里远,

但是当你的输入变得更加复杂时会发生什么

,这将我们带到

我们的第二个例子中,

所以当你查看没有坚果的图像时

通常用于进行

分类你得到的是

你有

一行一行一行的点,

每个点都有不同的

数字

,告诉你这些点的颜色

,所以如果你有一个 高分辨率

图像,您实际上可以

有数百万个输入,这

意味着我们校准

我们提出的公式的模型

的复杂性显着增加,

现在他正在增加复杂性

,我 复杂性的增加

使得这些模型实际上非常

脆弱,

它们对

你可能呈现给它的数据中的微小变化非常敏感,所以

让我举一两个例子

这是我们的第二个例子

,学术界的流行语

文学是

针对这类示例的对抗性攻击,而

对抗性攻击中发生的情况

是,您想象有人建立了一个

神经网络,他们已经

用一堆图像对其进行了训练以进行

分类

,然后他们想看看

您可以做哪些小的更改

为了欺骗神经

网络,假设一些对手试图

攻击这个人工智能系统,

所以第一批

真正做这项工作的第一批人是

日本的一组研究人员,

他们把

一个 有一堆动物图像的模型

,他们建立了一个分类器,然后

他们向它展示了这张马的图像,

所以这张图片不在训练

数据集中

,他们问了类 sifier 告诉

它是什么并说它是一匹马

然后他们去做什么

实际上只是改变了一个

像素他们说一个像素然后他们

把它从白色变成黑色

然后他们把它放在分类器中并

问它是什么 分类器

回来

说它是一只青蛙,所以这

对你来说应该有点麻烦,但它不应该像

下一个例子那样麻烦,因为下一个例子

很好,

如果你曾经的话,下一个例子对你来说应该是非常麻烦的 想

在未来的某个时候,您可能希望

让人工智能系统为您驾驶您的

汽车,所以这个系统专注于

街道标志,

并像

一堆街道标志和街道标志图像一样对

街道标志进行分类构建了一个神经网络

分类器,然后它显示 它

是停车标志的图像,它

能够将其归类为停车标志,

但随后它拍摄了图像并随机

在其上贴了几块胶带,

并要求神经网络

继续对其进行分类

以及何时拍摄 回到

分类它说这不是一个停止

标志它是一个前进

标志说你现在以每小时 45 英里的速度

行驶这些示例中

发生了什么情况是

它们的中间原始图像在

某种意义上

接近

训练的图像有点像

在无人机示例中由布莱恩特公园升起,

而我们进行微小更改的其他图像

突然远离,

并且进行图像

分析的部分问题真的很难知道

两张图像有多远 是因为输入中有

很多不同的

部分可以更改

,所以

本次演讲的第二个要点是

,当

人工智能系统遇到新事物时,

他们不知道自己遇到了新事物,

人工智能系统只是没有好的

找出他们何时丢失的方法

这是一个真正的问题这是一个

我们需要考虑的问题这是一个

人工智能社区

必须找到解决方法的问题所以一个

明显的解决方案是收集更多数据

但它

不仅仅是那么容易,你不能直接

说好,我会收集更多数据,

因为如果你在已有数据的地方收集更多数据,

那么你

在纽约公共图书馆周围的布莱恩特公园周围收集更多数据

当你去世界贸易中心这样的新事物时对你没有多大帮助 当你遇到新事物

时你必须知道

然后收集

数据 通常会发生的事情

通常会发生的事情是你尽力而为

建立一个人工智能系统,你收集

尽可能多的数据,然后你

运行人工智能系统

,你怎么知道你

遇到了一些新的东西?

人工智能系统崩溃

通常很糟糕,有时它是

灾难性的

,我认为我们

可以做得更好 眼睛

系统在那里,

所以有什么替代方法

替代方法是尝试

以某种方式模仿大脑,所以这真的很

有趣,我们的大脑不断地

为世界制作模型,但我

发现令人着迷的是它实际上

在做多个模型,所以只有一个

模型 它是由我们的脑干创建的,

它是我们世界的一种快速而肮脏的

模型

,还有另一个版本是

由我们的皮层创建的,它更加

深思熟虑和抽象

,然后在我们大脑的中心,

大脑的一部分

被称为 海马体及其

工作之一就是不断比较这两个

模型

,看看他们是否排队,如果他们排队,

你就继续,

如果他们不排队,但你

会突然觉得

有什么不对劲,有

异常 在心理学中有一些你需要关注的东西

,我们称之为定向

反射,

所以有一个异常,你的

大脑是一个非常快速的计算

丹 如果危险,你进入

战斗或逃跑模式

肾上腺素启动,

你进入生存模式,但如果它

不危险,有趣的是

你仍然进入不同的模式你

进入探索者模式,

因为你想过去 并找出

为什么你的两个模型不

匹配你去

那里收集更多数据并更新你的

模型

,我认为这可以提供一种非常

强大的方式

来思考如何

解决帮助人工智能系统

理解的问题 当它们丢失

时,您可以将其称为人工智能

定向反射

,如果模型同意,则您可以运行多个模型,

那么您就知道在您已知的

领域内继续进行,

但是如果模型不同意,那么您知道

您遇到了一些新的东西,

您知道 有未知的领域

你需要小心你需要

真正

放慢速度并收集更多数据

所以

当他们遇到未知的时候人工智能系统做得很好

没有系统知道他们什么时候迷路

没有 这是一个我们必须绝对解决的问题,

如果我们不能解决这个问题

,人工智能系统很有可能

会像 30 年前一样消失,

不仅仅是这些人工智能系统

不知道它们何时丢失

你知道这确实严重限制

了我们可以用人工智能系统做什么,但是

人工智能系统可以做什么有一个更基本的限制

,你会注意到,到目前为止,在本次演讲中,

我实际上

并没有给人工智能命名 人 AI

对我来说意味着人工智能 我认为这具有误导性

归根结底,它们确实是

工具,它们是强大的工具,它们是

增强我们

人类利用我们的智慧

解决问题的能力的工具

,如果您考虑一下这种差异

它实际上会帮助你

理解人工智能系统遇到的基本边界

,我想通过

谈论问题

或奇点的概念来说明这一点,所以这个

想法

或者更好的说法

是对奇点的恐惧最终

机器会变得如此聪明

他们将

直接从人工智能的

角度接管

这有点像说

特斯拉的超级计算机负责所有

神经网络并对所有

汽车进行更新突然发展出自己的思维

突然出现 自己的目标 自己的

价值观 独立于任何人

,谁知道它会做什么 也许它

继续告诉特斯拉把他们的

司机赶出去,聚集在北达科他州的

斯特吉斯参加一场大型集会

某种大型电子汽车舞会

从增强

智能的角度来看

,奇点风险是不同的,

因为

当您拥有增强智能

系统时,我 这样做是因为它将

越来越多的控制权集中

在越来越少的人手中,

因此风险类似于

一些青少年在他们的地下室它

可能在任何地方它可能是俄罗斯

中国美国入侵那

超级计算机的特斯拉计算机和 接管并

告诉特斯拉计算机

超级计算机将您发送到哪里知道

踢出让汽车踢出他们的司机

并参加大型舞会

,这就是我认为

就奇点而言我们应该真正

担心

的风险 一个社会,它不是

机器人接管的人工智能风险

现在是人工智能留在这里也许

希望如果它消失我毫不

怀疑它

很可能会回来

,我们将解决的问题将是惊人的,

但你应该 不要被愚弄 人工智能系统

总会有一个基本

限制 总会有一个

边界,这个基本

边界就是人工智能系统是工具

,意味着它们反映了虚拟现实

创造和使用它们的人的线索

他们没有创造自己的

价值

最终确定

什么是有价值的 什么是道德的 什么是

道德的

这是人类的任务 这是

我们永远不能外包的东西

谢谢