Rana el Kaliouby This app knows how you feel from the look on your face

Our emotions influence
every aspect of our lives,

from our health and how we learn,
to how we do business and make decisions,

big ones and small.

Our emotions also influence
how we connect with one another.

We’ve evolved to live
in a world like this,

but instead, we’re living
more and more of our lives like this –

this is the text message
from my daughter last night –

in a world that’s devoid of emotion.

So I’m on a mission to change that.

I want to bring emotions
back into our digital experiences.

I started on this path 15 years ago.

I was a computer scientist in Egypt,

and I had just gotten accepted to
a Ph.D. program at Cambridge University.

So I did something quite unusual

for a young newlywed Muslim Egyptian wife:

With the support of my husband,
who had to stay in Egypt,

I packed my bags and I moved to England.

At Cambridge, thousands of miles
away from home,

I realized I was spending
more hours with my laptop

than I did with any other human.

Yet despite this intimacy, my laptop
had absolutely no idea how I was feeling.

It had no idea if I was happy,

having a bad day, or stressed, confused,

and so that got frustrating.

Even worse, as I communicated
online with my family back home,

I felt that all my emotions
disappeared in cyberspace.

I was homesick, I was lonely,
and on some days I was actually crying,

but all I had to communicate
these emotions was this.

(Laughter)

Today’s technology
has lots of I.Q., but no E.Q.;

lots of cognitive intelligence,
but no emotional intelligence.

So that got me thinking,

what if our technology
could sense our emotions?

What if our devices could sense
how we felt and reacted accordingly,

just the way an emotionally
intelligent friend would?

Those questions led me and my team

to create technologies that can read
and respond to our emotions,

and our starting point was the human face.

So our human face happens to be
one of the most powerful channels

that we all use to communicate
social and emotional states,

everything from enjoyment, surprise,

empathy and curiosity.

In emotion science, we call each
facial muscle movement an action unit.

So for example, action unit 12,

it’s not a Hollywood blockbuster,

it is actually a lip corner pull,
which is the main component of a smile.

Try it everybody. Let’s get
some smiles going on.

Another example is action unit 4.
It’s the brow furrow.

It’s when you draw your eyebrows together

and you create all
these textures and wrinkles.

We don’t like them, but it’s
a strong indicator of a negative emotion.

So we have about 45 of these action units,

and they combine to express
hundreds of emotions.

Teaching a computer to read
these facial emotions is hard,

because these action units,
they can be fast, they’re subtle,

and they combine in many different ways.

So take, for example,
the smile and the smirk.

They look somewhat similar,
but they mean very different things.

(Laughter)

So the smile is positive,

a smirk is often negative.

Sometimes a smirk
can make you become famous.

But seriously, it’s important
for a computer to be able

to tell the difference
between the two expressions.

So how do we do that?

We give our algorithms

tens of thousands of examples
of people we know to be smiling,

from different ethnicities, ages, genders,

and we do the same for smirks.

And then, using deep learning,

the algorithm looks for all these
textures and wrinkles

and shape changes on our face,

and basically learns that all smiles
have common characteristics,

all smirks have subtly
different characteristics.

And the next time it sees a new face,

it essentially learns that

this face has the same
characteristics of a smile,

and it says, “Aha, I recognize this.
This is a smile expression.”

So the best way to demonstrate
how this technology works

is to try a live demo,

so I need a volunteer,
preferably somebody with a face.

(Laughter)

Cloe’s going to be our volunteer today.

So over the past five years, we’ve moved
from being a research project at MIT

to a company,

where my team has worked really hard
to make this technology work,

as we like to say, in the wild.

And we’ve also shrunk it so that
the core emotion engine

works on any mobile device
with a camera, like this iPad.

So let’s give this a try.

As you can see, the algorithm
has essentially found Cloe’s face,

so it’s this white bounding box,

and it’s tracking the main
feature points on her face,

so her eyebrows, her eyes,
her mouth and her nose.

The question is,
can it recognize her expression?

So we’re going to test the machine.

So first of all, give me your poker face.
Yep, awesome. (Laughter)

And then as she smiles,
this is a genuine smile, it’s great.

So you can see the green bar
go up as she smiles.

Now that was a big smile.

Can you try a subtle smile
to see if the computer can recognize?

It does recognize subtle smiles as well.

We’ve worked really hard
to make that happen.

And then eyebrow raised,
indicator of surprise.

Brow furrow, which is
an indicator of confusion.

Frown. Yes, perfect.

So these are all the different
action units. There’s many more of them.

This is just a slimmed-down demo.

But we call each reading
an emotion data point,

and then they can fire together
to portray different emotions.

So on the right side of the demo –
look like you’re happy.

So that’s joy. Joy fires up.

And then give me a disgust face.

Try to remember what it was like
when Zayn left One Direction.

(Laughter)

Yeah, wrinkle your nose. Awesome.

And the valence is actually quite
negative, so you must have been a big fan.

So valence is how positive
or negative an experience is,

and engagement is how
expressive she is as well.

So imagine if Cloe had access
to this real-time emotion stream,

and she could share it
with anybody she wanted to.

Thank you.

(Applause)

So, so far, we have amassed
12 billion of these emotion data points.

It’s the largest emotion
database in the world.

We’ve collected it
from 2.9 million face videos,

people who have agreed
to share their emotions with us,

and from 75 countries around the world.

It’s growing every day.

It blows my mind away

that we can now quantify something
as personal as our emotions,

and we can do it at this scale.

So what have we learned to date?

Gender.

Our data confirms something
that you might suspect.

Women are more expressive than men.

Not only do they smile more,
their smiles last longer,

and we can now really quantify
what it is that men and women

respond to differently.

Let’s do culture: So in the United States,

women are 40 percent
more expressive than men,

but curiously, we don’t see any difference
in the U.K. between men and women.

(Laughter)

Age: People who are 50 years and older

are 25 percent more emotive
than younger people.

Women in their 20s smile a lot more
than men the same age,

perhaps a necessity for dating.

But perhaps what surprised us
the most about this data

is that we happen
to be expressive all the time,

even when we are sitting
in front of our devices alone,

and it’s not just when we’re watching
cat videos on Facebook.

We are expressive when we’re emailing,
texting, shopping online,

or even doing our taxes.

Where is this data used today?

In understanding how we engage with media,

so understanding virality
and voting behavior;

and also empowering
or emotion-enabling technology,

and I want to share some examples
that are especially close to my heart.

Emotion-enabled wearable glasses
can help individuals

who are visually impaired
read the faces of others,

and it can help individuals
on the autism spectrum interpret emotion,

something that they really struggle with.

In education, imagine
if your learning apps

sense that you’re confused and slow down,

or that you’re bored, so it’s sped up,

just like a great teacher
would in a classroom.

What if your wristwatch tracked your mood,

or your car sensed that you’re tired,

or perhaps your fridge
knows that you’re stressed,

so it auto-locks to prevent you
from binge eating. (Laughter)

I would like that, yeah.

What if, when I was in Cambridge,

I had access to my real-time
emotion stream,

and I could share that with my family
back home in a very natural way,

just like I would’ve if we were all
in the same room together?

I think five years down the line,

all our devices are going
to have an emotion chip,

and we won’t remember what it was like
when we couldn’t just frown at our device

and our device would say, “Hmm,
you didn’t like that, did you?”

Our biggest challenge is that there are
so many applications of this technology,

my team and I realize that we can’t
build them all ourselves,

so we’ve made this technology available
so that other developers

can get building and get creative.

We recognize that
there are potential risks

and potential for abuse,

but personally, having spent
many years doing this,

I believe that the benefits to humanity

from having emotionally
intelligent technology

far outweigh the potential for misuse.

And I invite you all to be
part of the conversation.

The more people who know
about this technology,

the more we can all have a voice
in how it’s being used.

So as more and more
of our lives become digital,

we are fighting a losing battle
trying to curb our usage of devices

in order to reclaim our emotions.

So what I’m trying to do instead
is to bring emotions into our technology

and make our technologies more responsive.

So I want those devices
that have separated us

to bring us back together.

And by humanizing technology,
we have this golden opportunity

to reimagine how we
connect with machines,

and therefore, how we, as human beings,

connect with one another.

Thank you.

(Applause)

我们的情绪影响
着我们生活的方方面面,

从我们的健康和我们的学习方式,
到我们如何做生意和做决定,

无论大小。

我们的情绪也会
影响我们彼此联系的方式。

我们已经进化到生活
在这样的世界中,

但相反,我们
越来越多地生活在这样的生活中——


是我女儿昨晚发来的短信——

在一个没有情感的世界里。

所以我的使命是改变这一点。

我想将情感
带回我们的数字体验中。

15 年前,我开始走这条路。

我是埃及的一名计算机科学家,

刚刚获得
博士学位。 剑桥大学的课程。

因此,我

为一位年轻的新婚穆斯林埃及妻子做了一件非常不寻常的事情:


不得不留在埃及的丈夫的支持下,

我收拾行装,搬到了英国。

在离家数千英里的剑桥,

我意识到我花在
笔记本电脑上的时间

比我花在任何其他人身上的时间都多。

然而,尽管如此亲密,我的笔记本电脑
完全不知道我的感受。

它不知道我是快乐的

,度过了糟糕的一天,还是压力大,困惑

,所以这让我很沮丧。

更糟糕的是,当我
和家里的家人在网上交流时,

我觉得我所有的情绪都
消失在了网络空间中。

我想家,我很孤单
,有时我真的在哭,

但我只需要传达
这些情绪。

(笑声)

今天的技术
有很多智商,但没有情商;

很多认知智能,
但没有情商。

所以这让我想到

,如果我们的技术
能够感知我们的情绪呢?

如果我们的设备能够感知
我们的感受并做出相应的反应,

就像一个情商高的
朋友一样,那会怎样?

这些问题促使我和我的

团队创造了可以读取
和响应我们情绪的技术

,我们的出发点是人脸。

因此,我们的人脸恰好

是我们用来交流
社交和情感状态的最强大的渠道之一,

从享受、惊喜、

同理心和好奇心等一切。

在情感科学中,我们将每个
面部肌肉运动称为一个动作单元。

所以举个例子,动作单元12,

它不是好莱坞大片,

它实际上是一个唇角拉扯
,是微笑的主要组成部分。

大家试试看。 让我们继续
微笑吧。

另一个例子是动作单元 4。
它是眉间沟。

当你把眉毛画在一起时

,你会创造出所有
这些纹理和皱纹。

我们不喜欢它们,但它
是负面情绪的强烈指标。

所以我们有大约 45 个这样的动作单元

,它们结合起来表达
数百种情绪。

教计算机阅读
这些面部情绪是很困难的,

因为这些动作单元,
它们可以很快,它们很微妙,

并且它们以许多不同的方式结合起来。


微笑和假笑为例。

它们看起来有些相似,
但它们的含义却截然不同。

(笑声)

所以微笑是积极的,

假笑往往是消极的。

有时候一个傻笑
可以让你成名。

但说真的
,计算机能够

区分这两种表达方式是很重要的。

那么我们该怎么做呢?

我们为我们的算法提供了

成千上万个
我们认识的微笑的人的例子,他们

来自不同的种族、年龄、性别

,我们对傻笑也做同样的事情。

然后,通过深度学习

,算法寻找我们脸上的所有这些
纹理、皱纹

和形状变化,

基本上知道所有的笑容
都有共同的特征,

所有的笑容都有微妙的
不同特征。

下一次它看到一张新面孔时,

它基本上知道

这张面孔具有与微笑相同的
特征,

并说:“啊哈,我认得这个。
这是一个微笑的表情。”

所以展示
这项技术如何工作的最好方法

是尝试现场演示,

所以我需要一个志愿者,
最好是有脸的人。

(笑声)

Cloe 今天将成为我们的志愿者。

所以在过去的五年里,我们已经
从麻省理工学院的一个研究项目

转变为一家公司

,我的团队非常努力
地让这项技术

在野外工作,就像我们喜欢说的那样。

我们还对其进行了缩小,
以便核心情感引擎

可以在任何
带有摄像头的移动设备上运行,比如这款 iPad。

所以让我们试一试。

正如你所看到的,
算法基本上找到了克洛的脸,

所以它是这个白色的边界框

,它正在跟踪
她脸上的主要特征点,也

就是她的眉毛、眼睛、
嘴巴和鼻子。

问题是
,它能认出她的表情吗?

所以我们要测试机器。

所以首先,给我你的扑克脸。
是的,棒极了。 (笑声

) 然后当她微笑时,
这是一个真诚的微笑,非常棒。

因此
,当她微笑时,您可以看到绿色条上升。

现在那是一个大大的笑容。

能不能试着微微一笑
,看看电脑能不能识别?

它也能识别微妙的微笑。

我们非常
努力地实现这一目标。

然后扬起眉毛,
表示惊讶。

眉头皱起,这
是混乱的标志。

皱眉。 是的,完美。

所以这些都是不同的
行动单位。 还有很多。

这只是一个精简的演示。

但是我们把每次阅读都称为
一个情绪数据点,

然后它们可以一起触发
来描绘不同的情绪。

所以在演示的右侧——
看起来你很开心。

所以这很快乐。 乔伊火了。

然后给我一个厌恶的表情。

试着回忆
一下 Zayn 离开 One Direction 时的情景。

(笑声)

是的,皱起你的鼻子。 惊人的。

而且价实际上是相当
负的,所以你一定是个大粉丝。

因此,效价
是体验的积极或消极程度

,参与度也是
她的表现力。

所以想象一下,如果 Cloe 能够
访问这种实时情感流

,她可以
与任何她想分享的人分享。

谢谢你。

(掌声)

所以,到目前为止,我们已经积累了
120 亿个这样的情绪数据点。

它是世界上最大的情感
数据库。

我们
从 290 万个面部视频、

同意
与我们分享情感的人

以及来自全球 75 个国家/地区的人中收集了这些信息。

它每天都在增长。

让我大吃一惊的

是,我们现在可以量化
像我们的情绪这样个人的东西,

而且我们可以在这个规模上做到这一点。

那么到目前为止,我们学到了什么?

性别。

我们的数据证实
了您可能怀疑的事情。

女性比男性更有表现力。

他们不仅笑得更多,笑得更
久,

而且我们现在可以真正量化
男人和女人

对不同反应的反应。

让我们做文化吧:所以在美国,

女性的
表达能力比男性高 40%,

但奇怪的是,我们没有看到
英国的男性和女性之间有任何差异。

(笑声)

年龄:50 岁及以上

的人比年轻人的情绪化程度高 25%。

20多岁的女性比同龄男性笑得更多

这也许是约会的必需品。

但也许
这些数据最让我们感到惊讶的

是,我们
碰巧一直都在表达,

即使
我们独自坐在设备前,

而不仅仅是当我们
在 Facebook 上观看猫视频时。

当我们发送电子邮件、
发短信、在线购物,

甚至纳税时,我们都会表现得淋漓尽致。

今天在哪里使用这些数据?

了解我们如何与媒体互动,

从而了解病毒式传播
和投票行为;

还有赋能
或增强情感的技术

,我想分享
一些特别贴近我的心的例子。

具有情感功能的可穿戴眼镜
可以帮助

视障人士
阅读他人的面孔

,它可以
帮助自闭症患者解读情绪,

这是他们真正难以解决的问题。

在教育方面,想象一下,
如果您的学习应用程序

感觉到您感到困惑并放慢了速度,

或者您感到无聊,那么它就会加快速度,

就像一位优秀的
老师在课堂上一样。

如果您的手表跟踪您的心情,

或者您的汽车感觉到您累了,

或者您的冰箱
知道您有压力,

那么它会自动锁定以防止
您暴饮暴食。 (笑声)

我愿意,是的。

如果当我在剑桥时,

我可以访问我的实时
情感流,

并且我可以
以非常自然的方式与我的家人分享,

就像我们都
在同一个房间里一样 一起?

我想五年后,

我们所有的设备
都会有一个情感芯片,

当我们不能只是皱眉头看着我们的设备

,我们的设备会说,“嗯,
你没有 不喜欢那样,是吗?”

我们最大的挑战
是这项技术有如此多的应用,

我和我的团队意识到我们
不能自己构建它们,

所以我们提供了这项技术,
以便其他开发人员

可以构建并发挥创意。

我们认识到
存在潜在的风险

和滥用的可能性,

但就我个人而言,
多年来一直这样做,

我相信拥有情商智能技术对人类的好处

远远超过滥用的可能性。

我邀请大家
参与对话。

了解这项技术

的人越多,我们就越能
对它的使用方式发表意见。

因此,
随着我们的生活越来越数字化,

我们正在打一场失败的战斗,
试图遏制我们对设备的使用

以恢复我们的情绪。

因此,我正在尝试做的
是将情感带入我们的技术

,并使我们的技术更具响应性。

所以我希望
那些将我们分开的设备

将我们重新聚在一起。

通过将技术人性化,
我们有这个千载难逢的机会

来重新想象我们如何
与机器连接

,因此,我们作为人类如何

相互连接。

谢谢你。

(掌声)