Tackling AI Bias is a human problem

[Music]

[Music]

hi

i’m excited to have the opportunity to

speak with you i want to provide my

unbiased perspective on the critical

role of managing bias and ai

does that sound ridiculous well that’s

because it should

since there is no conversation without

bias

as humans we perceive our environments

our experiences and that affects our

perspective this lens

this perspective is bias similarly

an ai perceives its environment its

experiences

and this is in the form of data and this

affects its perspective

this perspective that the ai has learned

from the data influences and creates

bias in the resulting ai

but we want to know we can trust the ai

and be able to understand

how and why it came to a recommendation

but what does ai look like in action a

good example

is an algorithm that predicts credit

risk from your credit history and assets

this is based on how people similar to

you have behaved in the past

in other words the algorithm is scoring

you

based on the behavior of other similar

individuals

and algorithms are really good at this

in 2019

apple and goldman sachs released a

credit card aptly named the apple card

but when janet hill pulled out her

iphone and applied for the apple card

instantaneously an algorithm issued her

a card

but she was given a credit limit that

was 10 times lower than her husband

they share all the same assets and they

share all the same accounts

yet her husband who happened to be steve

wozniak

got 10 times the credit that she did now

the algorithm

doesn’t even use gender as a factor

they didn’t want by this bias to exist

in fact

they made sure gender wasn’t even a

consideration in the algorithm

but before we address how this likely

happened i want to use a story to

explain some of the different types of

biases that exist

in late 2017 i started a team

to help companies climb the ladder to ai

when we set out on this mission we

established a cognitive bias

by having a goal to create a team that

was as diverse as possible

we created a framing bias by centering

the entire hiring process

to support our cognitive bias after that

we did something quite simple

we worked to minimize a selection bias

by using a set of job postings that were

designed to be more inclusive

we took this approach as there’s a

difference in the types of job

descriptions

that different groups will and will not

apply to

because of an implicit bias in addition

to these simple

adjustments to the job postings we set a

requirement

before anyone could start interviewing

we needed to make sure we had a pool of

qualified candidates that’s at least

50 percent diverse this decision

of consciously establishing a selection

bias

proved to be one of the most important

parts of bringing this team together

it was so simple but it was so impactful

and i want to be clear we had no

requirement on the mix of candidates

that we would

or would not offer jobs to in fact we

had a very rigorous process

and only the most qualified candidates

got hired with no exceptions

when we finished the team of a hundred

data scientists was nearly twice as

diverse as the industry average

with nearly an equal number of men and

women who spoke more than 26 different

languages

and were from a wide variety of cultural

religious

and geopolitical backgrounds underneath

all this was a confirmation bias

in the form of a set of academic and

real world research

that was hand selected but why is this

so important when we start to think

about the development of ai

the same types of biases that are

expressed in the story about the data

science team

apply directly to ai that’s because

humans build ai

and we come to the table with a set of

biases and data is used to teach

or train the ai and data has bias that’s

based on the historical decisions

and actions of humans these factors

impact the level of bias an ai will have

so if we go back to the apple card

example

while they explicitly removed gender

there were

other factors associated with gender

that the ai algorithm

identified in order to classify

individuals based on a perceived risk

these features of the data are based on

a historical bias

but we’re not explicitly are you a man

or are you a woman

however they were still associated with

being a man or being a woman

and led to the unintended bias that

occurred

as we look at the pandemic we’re facing

it’s only increasing the rate

and pace of adoption of ai i know

sometimes ai can see him abstract

but you are already impacted by ai every

day

it’s in every industry from media and

entertainment

when you think about netflix

recommending a movie to you

to automotive with the autonomous and

semi-autonomous cars driving around

even telecommunications with the need to

automatically allocate

and prioritize network capacity based on

demand

and finally to the response to the

pandemic

as we are forecasting the impact that

rates of infection have on economies

being able to reopen or not

and bias in any of these interactions

can have a direct impact on our lives

from the simply annoying when netflix

tells you to watch bridget jones and you

really just want to watch fifth element

to predicting what happens to the

availability of critical supplies

if the pandemic shuts down an entire

economy

as ai becomes more pervasive we need to

understand

mitigate and remediate bias and ai

and the development of an ai algorithm

can be thought of as similar to our own

human development

as we are born mostly without biases ai

algorithms are not inherently biased

and environments introduce bias to us

both positive and negative

and data is the algorithmic equivalent

to an environment for ai

external pressure causes us to adjust

our biases over time just as data

changes over time

and as we mature we’re introduced into

new environments

just as ai algorithms learn from new

data over time

ai is very good at picking up on small

details in their environment

and classifying groups or individuals

based on the data which is sometimes

biased

so we think back again to the apple card

gender was

absolutely not a feature of the

algorithm and if ai has no inherent bias

then where did this bias come from

while data sets may seem like the most

obvious source of bias in ai

bias can also be introduced by teams

that don’t have proper training

and if teams are not sufficiently

diverse they’re more

likely to introduce a cognitive bias

when they’re setting up the problem

or framing bias is they lay out the

experimental design

or a selection bias when they start

picking algorithms

and this is in addition to whatever

biases may already exist in the data

from the historical actions of humans

so let’s go through another story

several years ago

a bank started using ai to decide if

mortgages should be given to applicants

or not

since this decision can have serious

long-term impact

on the lives of families the bank was

very careful

to ensure that the algorithms didn’t

have

any gender racial religious or ethnic

biases

however even when everything else was

the same the algorithm started to deny

people of color mortgages at a higher

rate

now remember race gender religion and

ethnicity were specifically excluded

from the algorithm

so what could have happened here it

turns out that address was collected

and in the united states and other parts

of the world

people of similar backgrounds tend to

live together in communities

and it also turns out that people of

color have been has

been historically denied credit at much

higher rates

and have historically lived in certain

zip codes

so if you happen to live in one of these

zip codes

you are more likely to be denied credit

this example demonstrates a few things

first that two or more pieces of data or

features

are likely to be very tightly connected

or correlated

in this case race address and zip code

are tightly correlated

and even though race was excluded the

algorithm found the result of historical

bias

in brace correlated to a separate set of

features

address and zip code second

it highlights the effect of biased data

and how biases introduced into the data

because the model was tainted by a

historical bias in the data

they were making a bunch of bad

decisions really fast

so what does it look like to truly

understand bias

it takes highly talented and diverse

teams

building the ai it is a set of separate

ai algorithms that are used to identify

outliers that represent both known and

unknown biases

it’s making sure we frame projects in a

way that’s as unbiased as possible

and it’s being transparent with the ai

in order to create a trusted foundation

for the ai

now the process of mitigating a

remediating bias

can still be challenging and to

understand why let’s go back to the

example of the bank

they were trying to look they took this

very seriously and their intent was not

to be biased against people of color

or anyone else for that matter in fact

the whole point of implementing the ai

was to remove the potential of implicit

bias

affecting the decisions of their

underwriters and mortgage brokers

now it was not easy for them to

understand what happened

remember they went to great lanes to try

to prevent this from happening in the

first place

it was such a challenge primarily

because their teams had to manually go

back through the algorithms and untangle

what happened

and this isn’t an easy process but we’re

at an amazing point in time

technologically we are at the point

where the bias mitigation process

can be automatic and can be integral to

the entire development life cycle of ai

the right people with the right tools

and the right technology

can start creating a fantastic future of

less biased

and more ethical ai in the form of a

fully automated end-to-end process that

accounts for bias at every step of the

way

and is overseen by talented and diverse

teams

thinking about the new reality we live

in today thanks to the pandemic

entire economies are shutting down and

this is resulting in a world that was

once physically connected via air travel

becoming completely remote where

millions of people are without work

and many others are working remotely

these fundamental changes in the world

are exponentially accelerating the rate

and pace

of the implementation and adoption of ai

whatever trajectory we were on has been

accelerated

because we’re relying on data and ai

more than ever

but as we interact with ai in our daily

lives

we need to be cognizant of the fact that

biases

can be and often are impacting us

directly and indirectly

we need to keep in mind that even

ethical ai is biased

even if it’s consciously biased in a way

that’s aligned to societal norms

now we’ve all heard about the possible

benefits of ai

everything from better customer service

to more efficient and resilient supply

chains

to faster and smarter drug discovery

process and so on

and many of us look forward to the

innovations and impact that ai can offer

organizations and individuals want to

know that they can trust their data

and their ai and explain how it came to

a recommendation

think about how many more organizations

would be ready to use ai

if they could rely on a trusted and

transparent process

now when i think about the task at hand

i know we have a lot of work to do

but i am very hopeful humans created the

bias that’s out there in the world today

and together we have a shared

responsibility

to make sure that ai reflects the best

in human thinking

not the worst done right

done ethically ai will help us emerge

for the pandemic

by us towards a better more equitable

society

[音乐]

[音乐]

嗨,

我很高兴有机会

与您交谈

偏见

作为人类我们感知我们的环境

我们的经验,这会影响我们的

观点这个镜头

这个观点同样是偏见

人工智能感知它的环境它的

经验

,这是以数据的形式,这

会影响它的

观点人工智能

从 数据会影响并

在生成的 AI 中产生偏差,

但我们想知道我们可以信任 AI,

并且能够理解

它是如何以及为什么会产生推荐的,

但是 AI 在实际中是什么样的一个

很好的例子

是预测信用风险的算法

根据您的信用记录和资产,

这是基于与您相似的人

过去的行为方式,

换句话说,算法

根据您对您进行评分 其他类似的

个人

和算法的行为真的很擅长

在 2019 年

苹果和高盛发布了一张

信用卡,恰如其分地命名为苹果卡,

但是当珍妮特希尔拿出她的

iPhone 并立即申请苹果卡时

,算法给她发了

一张卡,

但是 她的信用额度

比她丈夫低 10 倍,

他们共享所有相同的资产,他们

共享所有相同的账户,

但她的丈夫恰好是史蒂夫·

沃兹尼亚克,

获得的信用额度是她现在的 10 倍

算法

没有 甚至使用性别作为

他们不希望这种偏见存在

的因素事实上

他们确保性别甚至不是

算法中的考虑因素,

但在我们解决这可能

发生的原因之前,我想用一个故事来

解释一些 2017 年末存在的不同类型的

偏见

我成立了一个团队

来帮助公司爬上人工智能的阶梯

当我们开始执行这项任务时,我们

通过制定目标来建立认知偏见 一个尽可能多样化的团队

我们通过

集中整个招聘过程

来支持我们的认知偏见,从而产生了框架偏见之后

我们做了一些非常简单的事情,我们

通过使用一组旨在减少选择偏见的工作职位

更具包容性,

我们采用了这种方法

因为除了

职位发布的这些简单调整之外,由于隐含的偏见,不同群体将适用和不会适用的职位描述类型存在差异,我们

在任何人开始面试之前设定了要求

需要确保我们拥有

至少

50% 多样化的合格候选人库 这个

有意识地建立选择

偏见的决定

被证明是

让这个团队团结起来的最重要的部分之一,

它是如此简单,但它是如此有影响力

,我 想明确一点,我们

对我们愿意

或不会提供工作的候选人组合没有要求,事实上我们

有一个非常 过程艰巨

,只有最合格的候选人才能

在我们完成工作时毫无例外地被录用。由 100 名数据科学家组成的团队的

多样性几乎是行业平均水平的两倍

,其中男性和女性的人数几乎相等,

他们会说超过 26 种不同的

语言

,并且 从各种各样的文化

宗教

和地缘政治背景来看,

这一切都是

以一系列学术和现实世界研究的形式出现的确认偏见,这些

研究

是手工挑选的,但是

当我们开始考虑人工智能的发展时,为什么这如此重要?

关于数据科学团队的故事中表达的相同类型的偏见

直接适用于人工智能,这是因为

人类构建了人工智能

,我们带着一组偏见来到谈判桌前

,数据被用来教授

或训练人工智能,而数据有偏见

基于人类的历史决策

和行为,这些因素

会影响人工智能的偏见程度,

所以如果我们回到苹果卡的

例子 如果他们明确删除

了性别

,人工智能算法还

确定了其他与性别相关的因素,以便

根据感知到的风险对

个人进行分类 你是一个女人,

但是他们仍然

与男人或女人联系在一起,

并导致了

当我们看到我们面临的大流行时发生的意外偏见,

它只会增加

人工智能的采用率和速度我知道

有时人工智能可以 看到他很抽象,

但你已经每天都受到人工智能的影响,

从媒体和娱乐到每个行业,

当你想到 netflix

向你推荐电影

时,自动和

半自动汽车在汽车周围行驶,

甚至需要

自动分配的电信

并根据需求优先考虑网络容量

,最后在

我们预测鼠疫的影响时对大流行的反应

感染对

经济能否重新开放

有影响,任何这些互动中的偏见都会对我们的生活产生直接影响,

从 netflix

告诉你看布里奇特琼斯而你

真的只想看第五元素

到预测

如果随着人工智能变得越来越普遍,大流行导致整个经济关闭,关键供应的可用性会发生什么

我们需要

了解

减轻和纠正偏见,人工智能

和人工智能算法的开发

可以被认为与我们自己的

人类发展

类似 我们生来大多没有偏见人工智能

算法并不是天生就有

偏见,环境会给我们

带来积极和消极的偏见

,数据是

人工智能环境的算法等价物,

外部压力导致我们

随着时间的推移调整我们的偏见,就像数据

随着时间的推移而变化一样

随着我们的成熟,我们被引入

新的环境

,就像人工智能算法

随着时间的推移从新数据中学习一样,

人工智能非常有效 OD 善于

在他们的环境中提取小细节

并根据有时有偏见的数据对群体或个人进行分类

所以我们再次回想一下苹果卡

性别

绝对不是

算法的一个特征,如果 ai 没有固有的偏见,

那么在哪里 这种偏见是否来自,

而数据集似乎是

人工智能偏见中最明显的偏见来源,

也可能由

没有经过适当培训的团队引入

,如果团队不够

多样化,他们更有

可能引入认知偏见

当他们设置问题

或框架

偏差时,他们会在开始选择算法时布置实验设计或选择偏差

,这是除了人类历史行为

数据中可能已经存在的任何偏差之外,

让我们开始吧 几年前通过另一个故事,

一家银行开始使用人工智能来决定是否

应该向申请人提供抵押贷款,

因为这个决定可能会产生严重

的长期影响

关于家庭生活的协议银行

非常小心,

以确保算法没有

任何性别种族宗教或种族

偏见

但是即使其他一切

都相同,算法开始

以更高的利率拒绝有色人种抵押贷款

现在记住了 种族、性别、宗教和

种族被明确

排除在算法之外,

所以这里可能发生的事情

事实证明,地址是被收集的

,在美国和

世界其他地区,

具有相似背景的人倾向于

一起生活在社区中

,结果也是 历史上,

有色人种被

拒绝信用的

比率要高得多

,并且历史上一直住在某些

邮政编码中,

所以如果你碰巧住在这些邮政编码中的一个,

你更有可能被拒绝信用

这个例子

首先展示了一些事情 在这种情况下,两条或更多条数据或

特征可能非常紧密地连接

或相关

ess 和

邮政编码密切相关

,即使排除了种族,

算法也发现

括号中的历史偏差结果与一组单独的

特征

地址和邮政编码相关,其次

它突出了有偏差数据的影响以及偏差是

如何引入数据的,

因为 该模型受到

数据中的历史偏见的影响,

他们非常快地做出了一堆错误的

决定,

所以真正理解偏见是什么样子的

呢?它需要非常有才华和多元化的

团队来

构建人工智能它是一组独立的

人工智能算法, 用于识别

代表已知和

未知偏差的异常值

确保我们以尽可能公正的方式构建项目

并且它对人工智能是透明的

,以便为人工智能创建一个值得信赖的基础

现在缓解补救的过程

偏见仍然具有挑战性,要

理解为什么让我们回到他们试图寻找

的银行的例子,

他们采取了这个

非常认真,他们的意图不是

对有色人种

或其他任何人产生偏见,事实上

,实施人工智能的全部目的

是消除

影响其

承销商和抵押贷款经纪人决策的隐性偏见的可能性,

现在不是 他们很容易

理解发生了什么

记得他们一开始就

试图阻止这种情况发生,

这是一个很大的挑战,主要是

因为他们的团队必须手动

返回算法并解开

发生的事情,

而这不是 一个简单的过程,但我们

在技术上处于一个惊人的时间点,

我们正处于

消除偏见过程

可以是自动的,并且可以成为

人工智能整个开发生命周期的组成部分

的地步

可以开始

以完全自动化的端到端流程的形式创造一个更少偏见和更道德的人工智能的美好未来,

该流程

占 b ias 在每一步,

并由才华横溢的多元化团队监督,

思考我们今天生活的新现实

,由于大流行,

整个经济体都在关闭,

这导致一个

曾经通过航空旅行物理连接的世界

变得完全 偏远地区,

数百万人没有工作

,还有许多人在远程工作

,世界上这些根本性的变化

正在成倍地加快

人工智能的实施和采用的速度和步伐,

无论我们走在哪条轨道上,都

因为我们依赖数据和 人工智能

比以往任何时候都多,

但当我们在日常生活中与人工智能互动时,

我们需要认识到

偏见

可能而且经常

直接和间接地影响我们,

我们需要记住,即使

是有意识的道德人工智能也存在偏见

符合社会规范的方式存在偏见

现在我们都听说过

人工智能可能带来的好处,

一切来自于更好的客户 r

为更高效和有弹性的供应链提供服务,

以更快和更智能的药物发现

过程等等

,我们中的许多人期待

人工智能可以提供的创新和影响,

组织和个人想

知道他们可以信任他们的数据

和他们的人工智能和 解释它是如何

提出建议的

想想如果现在有多少组织

可以依靠可信和

透明的流程,他们会准备好使用人工智能

当我想到手头的任务时,

我知道我们还有很多工作要做,

但我 我非常希望人类创造

了当今世界上存在的偏见,

我们有共同的

责任确保人工智能反映

人类思想中最好的一面,

而不是在道德上做的最坏的正确

人工智能将帮助我们

应对大流行病

由我们走向 一个更美好更公平的

社会