How Im fighting bias in algorithms Joy Buolamwini

Hello, I’m Joy, a poet of code,

on a mission to stop
an unseen force that’s rising,

a force that I called “the coded gaze,”

my term for algorithmic bias.

Algorithmic bias, like human bias,
results in unfairness.

However, algorithms, like viruses,
can spread bias on a massive scale

at a rapid pace.

Algorithmic bias can also lead
to exclusionary experiences

and discriminatory practices.

Let me show you what I mean.

(Video) Joy Buolamwini: Hi, camera.
I’ve got a face.

Can you see my face?

No-glasses face?

You can see her face.

What about my face?

I’ve got a mask. Can you see my mask?

Joy Buolamwini: So how did this happen?

Why am I sitting in front of a computer

in a white mask,

trying to be detected by a cheap webcam?

Well, when I’m not fighting the coded gaze

as a poet of code,

I’m a graduate student
at the MIT Media Lab,

and there I have the opportunity to work
on all sorts of whimsical projects,

including the Aspire Mirror,

a project I did so I could project
digital masks onto my reflection.

So in the morning, if I wanted
to feel powerful,

I could put on a lion.

If I wanted to be uplifted,
I might have a quote.

So I used generic
facial recognition software

to build the system,

but found it was really hard to test it
unless I wore a white mask.

Unfortunately, I’ve run
into this issue before.

When I was an undergraduate
at Georgia Tech studying computer science,

I used to work on social robots,

and one of my tasks was to get a robot
to play peek-a-boo,

a simple turn-taking game

where partners cover their face
and then uncover it saying, “Peek-a-boo!”

The problem is, peek-a-boo
doesn’t really work if I can’t see you,

and my robot couldn’t see me.

But I borrowed my roommate’s face
to get the project done,

submitted the assignment,

and figured, you know what,
somebody else will solve this problem.

Not too long after,

I was in Hong Kong
for an entrepreneurship competition.

The organizers decided
to take participants

on a tour of local start-ups.

One of the start-ups had a social robot,

and they decided to do a demo.

The demo worked on everybody
until it got to me,

and you can probably guess it.

It couldn’t detect my face.

I asked the developers what was going on,

and it turned out we had used the same
generic facial recognition software.

Halfway around the world,

I learned that algorithmic bias
can travel as quickly

as it takes to download
some files off of the internet.

So what’s going on?
Why isn’t my face being detected?

Well, we have to look
at how we give machines sight.

Computer vision uses
machine learning techniques

to do facial recognition.

So how this works is, you create
a training set with examples of faces.

This is a face. This is a face.
This is not a face.

And over time, you can teach a computer
how to recognize other faces.

However, if the training sets
aren’t really that diverse,

any face that deviates too much
from the established norm

will be harder to detect,

which is what was happening to me.

But don’t worry – there’s some good news.

Training sets don’t just
materialize out of nowhere.

We actually can create them.

So there’s an opportunity to create
full-spectrum training sets

that reflect a richer
portrait of humanity.

Now you’ve seen in my examples

how social robots

was how I found out about exclusion
with algorithmic bias.

But algorithmic bias can also lead
to discriminatory practices.

Across the US,

police departments are starting to use
facial recognition software

in their crime-fighting arsenal.

Georgetown Law published a report

showing that one in two adults
in the US – that’s 117 million people –

have their faces
in facial recognition networks.

Police departments can currently look
at these networks unregulated,

using algorithms that have not
been audited for accuracy.

Yet we know facial recognition
is not fail proof,

and labeling faces consistently
remains a challenge.

You might have seen this on Facebook.

My friends and I laugh all the time
when we see other people

mislabeled in our photos.

But misidentifying a suspected criminal
is no laughing matter,

nor is breaching civil liberties.

Machine learning is being used
for facial recognition,

but it’s also extending beyond the realm
of computer vision.

In her book, “Weapons
of Math Destruction,”

data scientist Cathy O’Neil
talks about the rising new WMDs –

widespread, mysterious
and destructive algorithms

that are increasingly being used
to make decisions

that impact more aspects of our lives.

So who gets hired or fired?

Do you get that loan?
Do you get insurance?

Are you admitted into the college
you wanted to get into?

Do you and I pay the same price
for the same product

purchased on the same platform?

Law enforcement is also starting
to use machine learning

for predictive policing.

Some judges use machine-generated
risk scores to determine

how long an individual
is going to spend in prison.

So we really have to think
about these decisions.

Are they fair?

And we’ve seen that algorithmic bias

doesn’t necessarily always
lead to fair outcomes.

So what can we do about it?

Well, we can start thinking about
how we create more inclusive code

and employ inclusive coding practices.

It really starts with people.

So who codes matters.

Are we creating full-spectrum teams
with diverse individuals

who can check each other’s blind spots?

On the technical side,
how we code matters.

Are we factoring in fairness
as we’re developing systems?

And finally, why we code matters.

We’ve used tools of computational creation
to unlock immense wealth.

We now have the opportunity
to unlock even greater equality

if we make social change a priority

and not an afterthought.

And so these are the three tenets
that will make up the “incoding” movement.

Who codes matters,

how we code matters

and why we code matters.

So to go towards incoding,
we can start thinking about

building platforms that can identify bias

by collecting people’s experiences
like the ones I shared,

but also auditing existing software.

We can also start to create
more inclusive training sets.

Imagine a “Selfies for Inclusion” campaign

where you and I can help
developers test and create

more inclusive training sets.

And we can also start thinking
more conscientiously

about the social impact
of the technology that we’re developing.

To get the incoding movement started,

I’ve launched the Algorithmic
Justice League,

where anyone who cares about fairness
can help fight the coded gaze.

On codedgaze.com, you can report bias,

request audits, become a tester

and join the ongoing conversation,

#codedgaze.

So I invite you to join me

in creating a world where technology
works for all of us,

not just some of us,

a world where we value inclusion
and center social change.

Thank you.

(Applause)

But I have one question:

Will you join me in the fight?

(Laughter)

(Applause)

你好,我是代码诗人乔伊,我

的任务是阻止
一股看不见的力量在上升

,我称之为“编码凝视”的力量,

我对算法偏见的称呼。

算法偏见,就像人类偏见一样,
会导致不公平。

然而,算法,如病毒,
可以快速大规模地传播偏见

算法偏差也可能
导致排他性经验

和歧视性做法。

让我告诉你我的意思。

(视频)Joy Buolamwini:嗨,相机。
我有一张脸。

你能看到我的脸吗?

不戴眼镜的脸?

你可以看到她的脸。

我的脸呢?

我有一个面具。 你能看到我的面具吗?

Joy Buolamwini:那么这是怎么发生的呢?

为什么我

戴着白色面具坐在电脑前,

试图被廉价的网络摄像头检测到?

好吧,当我不以代码诗人的身份对抗编码凝视时

我是
麻省理工学院媒体实验室的研究生,在

那里我有机会
从事各种异想天开的项目,

包括 Aspire Mirror

、 我做了这个项目,以便我可以将
数字面具投射到我的倒影上。

所以早上,如果我
想感觉强大,

我可以戴上一头狮子。

如果我想振作起来,
我可能会有一个报价。

所以我使用通用的
面部识别软件

来构建系统,

但发现
除非我戴上白色口罩,否则很难对其进行测试。

不幸的是,我以前遇到过
这个问题。

当我
在乔治亚理工大学攻读计算机科学的本科时,

我曾经从事社交机器人的工作

,我的任务之一是让
机器人玩躲猫猫,这

是一个简单的回合制游戏

,合作伙伴遮住脸
和 然后打开它说:“Peek-a-boo!”

问题是,
如果我看不到你,躲猫猫真的不起作用,

而我的机器人也看不到我。

但是我借了我室友的脸
来完成这个项目,

提交了作业,

然后我想,你知道吗,
其他人会解决这个问题。

不久之后,

我在香港
参加了一个创业比赛。

组织者
决定带

参与者参观当地的初创企业。

其中一家初创公司有一个社交机器人

,他们决定做一个演示。

该演示对每个人都有效,
直到我收到为止

,您可能已经猜到了。

它无法检测到我的脸。

我问开发人员发生了什么事

,结果发现我们使用了相同的
通用面部识别软件。

在地球的另一端,

我了解到算法偏差
的传播速度


从互联网上下载一些文件的速度一样快。

发生什么了?
为什么没有检测到我的脸?

好吧,我们必须
看看我们如何让机器看到。

计算机视觉使用
机器学习

技术进行面部识别。

所以这是如何工作的,你创建
一个包含人脸示例的训练集。

这是一张脸。 这是一张脸。
这不是一张脸。

随着时间的推移,你可以教计算机
如何识别其他面孔。

然而,如果训练集
不是真的那么多样化,那么

任何偏离既定规范太多的面孔

都将更难被发现,

这就是发生在我身上的事情。

但别担心——有一些好消息。

训练集
不会凭空出现。

我们实际上可以创建它们。

因此,有机会

创建反映更丰富
的人类画像的全方位训练集。

现在你已经在我的例子中

看到了社交机器人

是如何让我发现
算法偏见排除的。

但算法偏见也可能
导致歧视性做法。

在美国各地,

警察部门开始

在打击犯罪的武器库中使用面部识别软件。

Georgetown Law 发表的一份报告

显示,美国有二分之一的
成年人——即 1.17 亿人——

在面部识别网络中拥有自己的面孔。

警察部门目前可以使用
未经准确性审核的算法来查看这些不受监管的网络

然而,我们知道面部
识别并不是万无一失的,给

面部贴上标签
始终是一个挑战。

您可能已经在 Facebook 上看到了这一点。

当我们看到其他人

在我们的照片中贴错标签时,我和我的朋友们一直在笑。

但是,错误地识别犯罪嫌疑人
并不是笑话,

也不是侵犯公民自由。

机器学习被
用于面部识别,

但它也超出
了计算机视觉领域。 数据科学家 Cathy O’Neil

在她的书《
数学破坏的武器》中

谈到了新兴的大规模杀伤性武器——

广泛的、神秘的
和破坏性的算法

,越来越多地被用于

做出影响我们生活更多方面的决策。

那么谁被雇用或解雇呢?

你得到那笔贷款了吗?
你有保险吗?

你被你想进入的大学录取
了吗?

你和我在同

一个平台上购买的同一个产品支付相同的价格吗?

执法部门也
开始使用机器学习

进行预测性警务。

一些法官使用机器生成的
风险评分来确定

个人将在监狱中度过多长时间。

所以我们真的必须
考虑这些决定。

他们公平吗?

我们已经看到,算法偏差

并不一定总
能带来公平的结果。

那么我们能做些什么呢?

好吧,我们可以开始思考
如何创建更具包容性的代码

并采用包容性编码实践。

它真的从人开始。

所以谁编码很重要。

我们是否正在创建

由可以检查彼此盲点的多元化个人组成的全方位团队?

在技术方面,
我们如何编码很重要。

我们在开发系统时是否考虑了公平性?

最后,为什么我们编码很重要。

我们已经使用计算创造工具
来解锁巨大的财富。

如果我们将社会变革作为优先事项

而不是事后考虑,我们现在就有机会实现更大的平等。

这就是构成“编码”运动的三个原则。

谁编码很重要,

我们如何编码很重要,

以及为什么我们编码很重要。

因此,为了实现编码,
我们可以开始考虑

构建平台,该平台可以

通过收集
我分享的人们的经验来识别偏见,

但也可以审计现有的软件。

我们也可以开始创建
更具包容性的训练集。

想象一下“Selfies for Inclusion”活动

,您和我可以帮助
开发人员测试和创建

更具包容性的训练集。

我们也可以开始
更加认真地

思考
我们正在开发的技术的社会影响。

为了启动编码运动,

我发起了算法
正义联盟

,任何关心公平的人都
可以帮助对抗编码凝视。

在 codedgaze.com 上,您可以报告偏见、

请求审核、成为测试人员

并加入正在进行的对话

#codedgaze。

因此,我邀请您与我

一起创造一个
技术为我们

所有人服务的世界,而不仅仅是我们中的一些人,

一个我们重视包容
并以社会变革为中心的世界。

谢谢你。

(鼓掌)

但是我有一个问题:

你愿意和我一起打吗?

(笑声)

(掌声)