Machine intelligence makes human morals more important Zeynep Tufekci

So, I started my first job
as a computer programmer

in my very first year of college –

basically, as a teenager.

Soon after I started working,

writing software in a company,

a manager who worked at the company
came down to where I was,

and he whispered to me,

“Can he tell if I’m lying?”

There was nobody else in the room.

“Can who tell if you’re lying?
And why are we whispering?”

The manager pointed
at the computer in the room.

“Can he tell if I’m lying?”

Well, that manager was having
an affair with the receptionist.

(Laughter)

And I was still a teenager.

So I whisper-shouted back to him,

“Yes, the computer can tell
if you’re lying.”

(Laughter)

Well, I laughed, but actually,
the laugh’s on me.

Nowadays, there are computational systems

that can suss out
emotional states and even lying

from processing human faces.

Advertisers and even governments
are very interested.

I had become a computer programmer

because I was one of those kids
crazy about math and science.

But somewhere along the line
I’d learned about nuclear weapons,

and I’d gotten really concerned
with the ethics of science.

I was troubled.

However, because of family circumstances,

I also needed to start working
as soon as possible.

So I thought to myself, hey,
let me pick a technical field

where I can get a job easily

and where I don’t have to deal
with any troublesome questions of ethics.

So I picked computers.

(Laughter)

Well, ha, ha, ha!
All the laughs are on me.

Nowadays, computer scientists
are building platforms

that control what a billion
people see every day.

They’re developing cars
that could decide who to run over.

They’re even building machines, weapons,

that might kill human beings in war.

It’s ethics all the way down.

Machine intelligence is here.

We’re now using computation
to make all sort of decisions,

but also new kinds of decisions.

We’re asking questions to computation
that have no single right answers,

that are subjective

and open-ended and value-laden.

We’re asking questions like,

“Who should the company hire?”

“Which update from which friend
should you be shown?”

“Which convict is more
likely to reoffend?”

“Which news item or movie
should be recommended to people?”

Look, yes, we’ve been using
computers for a while,

but this is different.

This is a historical twist,

because we cannot anchor computation
for such subjective decisions

the way we can anchor computation
for flying airplanes, building bridges,

going to the moon.

Are airplanes safer?
Did the bridge sway and fall?

There, we have agreed-upon,
fairly clear benchmarks,

and we have laws of nature to guide us.

We have no such anchors and benchmarks

for decisions in messy human affairs.

To make things more complicated,
our software is getting more powerful,

but it’s also getting less
transparent and more complex.

Recently, in the past decade,

complex algorithms
have made great strides.

They can recognize human faces.

They can decipher handwriting.

They can detect credit card fraud

and block spam

and they can translate between languages.

They can detect tumors in medical imaging.

They can beat humans in chess and Go.

Much of this progress comes
from a method called “machine learning.”

Machine learning is different
than traditional programming,

where you give the computer
detailed, exact, painstaking instructions.

It’s more like you take the system
and you feed it lots of data,

including unstructured data,

like the kind we generate
in our digital lives.

And the system learns
by churning through this data.

And also, crucially,

these systems don’t operate
under a single-answer logic.

They don’t produce a simple answer;
it’s more probabilistic:

“This one is probably more like
what you’re looking for.”

Now, the upside is:
this method is really powerful.

The head of Google’s AI systems called it,

“the unreasonable effectiveness of data.”

The downside is,

we don’t really understand
what the system learned.

In fact, that’s its power.

This is less like giving
instructions to a computer;

it’s more like training
a puppy-machine-creature

we don’t really understand or control.

So this is our problem.

It’s a problem when this artificial
intelligence system gets things wrong.

It’s also a problem
when it gets things right,

because we don’t even know which is which
when it’s a subjective problem.

We don’t know what this thing is thinking.

So, consider a hiring algorithm –

a system used to hire people,
using machine-learning systems.

Such a system would have been trained
on previous employees' data

and instructed to find and hire

people like the existing
high performers in the company.

Sounds good.

I once attended a conference

that brought together
human resources managers and executives,

high-level people,

using such systems in hiring.

They were super excited.

They thought that this would make hiring
more objective, less biased,

and give women
and minorities a better shot

against biased human managers.

And look – human hiring is biased.

I know.

I mean, in one of my early jobs
as a programmer,

my immediate manager would sometimes
come down to where I was

really early in the morning
or really late in the afternoon,

and she’d say, “Zeynep,
let’s go to lunch!”

I’d be puzzled by the weird timing.

It’s 4pm. Lunch?

I was broke, so free lunch. I always went.

I later realized what was happening.

My immediate managers
had not confessed to their higher-ups

that the programmer they hired
for a serious job was a teen girl

who wore jeans and sneakers to work.

I was doing a good job,
I just looked wrong

and was the wrong age and gender.

So hiring in a gender- and race-blind way

certainly sounds good to me.

But with these systems,
it is more complicated, and here’s why:

Currently, computational systems
can infer all sorts of things about you

from your digital crumbs,

even if you have not
disclosed those things.

They can infer your sexual orientation,

your personality traits,

your political leanings.

They have predictive power
with high levels of accuracy.

Remember – for things
you haven’t even disclosed.

This is inference.

I have a friend who developed
such computational systems

to predict the likelihood
of clinical or postpartum depression

from social media data.

The results are impressive.

Her system can predict
the likelihood of depression

months before the onset of any symptoms –

months before.

No symptoms, there’s prediction.

She hopes it will be used
for early intervention. Great!

But now put this in the context of hiring.

So at this human resources
managers conference,

I approached a high-level manager
in a very large company,

and I said to her, “Look,
what if, unbeknownst to you,

your system is weeding out people
with high future likelihood of depression?

They’re not depressed now,
just maybe in the future, more likely.

What if it’s weeding out women
more likely to be pregnant

in the next year or two
but aren’t pregnant now?

What if it’s hiring aggressive people
because that’s your workplace culture?”

You can’t tell this by looking
at gender breakdowns.

Those may be balanced.

And since this is machine learning,
not traditional coding,

there is no variable there
labeled “higher risk of depression,”

“higher risk of pregnancy,”

“aggressive guy scale.”

Not only do you not know
what your system is selecting on,

you don’t even know
where to begin to look.

It’s a black box.

It has predictive power,
but you don’t understand it.

“What safeguards,” I asked, “do you have

to make sure that your black box
isn’t doing something shady?”

She looked at me as if I had
just stepped on 10 puppy tails.

(Laughter)

She stared at me and she said,

“I don’t want to hear
another word about this.”

And she turned around and walked away.

Mind you – she wasn’t rude.

It was clearly: what I don’t know
isn’t my problem, go away, death stare.

(Laughter)

Look, such a system
may even be less biased

than human managers in some ways.

And it could make monetary sense.

But it could also lead

to a steady but stealthy
shutting out of the job market

of people with higher risk of depression.

Is this the kind of society
we want to build,

without even knowing we’ve done this,

because we turned decision-making
to machines we don’t totally understand?

Another problem is this:

these systems are often trained
on data generated by our actions,

human imprints.

Well, they could just be
reflecting our biases,

and these systems
could be picking up on our biases

and amplifying them

and showing them back to us,

while we’re telling ourselves,

“We’re just doing objective,
neutral computation.”

Researchers found that on Google,

women are less likely than men
to be shown job ads for high-paying jobs.

And searching for African-American names

is more likely to bring up ads
suggesting criminal history,

even when there is none.

Such hidden biases
and black-box algorithms

that researchers uncover sometimes
but sometimes we don’t know,

can have life-altering consequences.

In Wisconsin, a defendant
was sentenced to six years in prison

for evading the police.

You may not know this,

but algorithms are increasingly used
in parole and sentencing decisions.

He wanted to know:
How is this score calculated?

It’s a commercial black box.

The company refused to have its algorithm
be challenged in open court.

But ProPublica, an investigative
nonprofit, audited that very algorithm

with what public data they could find,

and found that its outcomes were biased

and its predictive power
was dismal, barely better than chance,

and it was wrongly labeling
black defendants as future criminals

at twice the rate of white defendants.

So, consider this case:

This woman was late
picking up her godsister

from a school in Broward County, Florida,

running down the street
with a friend of hers.

They spotted an unlocked kid’s bike
and a scooter on a porch

and foolishly jumped on it.

As they were speeding off,
a woman came out and said,

“Hey! That’s my kid’s bike!”

They dropped it, they walked away,
but they were arrested.

She was wrong, she was foolish,
but she was also just 18.

She had a couple of juvenile misdemeanors.

Meanwhile, that man had been arrested
for shoplifting in Home Depot –

85 dollars' worth of stuff,
a similar petty crime.

But he had two prior
armed robbery convictions.

But the algorithm scored her
as high risk, and not him.

Two years later, ProPublica found
that she had not reoffended.

It was just hard to get a job
for her with her record.

He, on the other hand, did reoffend

and is now serving an eight-year
prison term for a later crime.

Clearly, we need to audit our black boxes

and not have them have
this kind of unchecked power.

(Applause)

Audits are great and important,
but they don’t solve all our problems.

Take Facebook’s powerful
news feed algorithm –

you know, the one that ranks everything
and decides what to show you

from all the friends and pages you follow.

Should you be shown another baby picture?

(Laughter)

A sullen note from an acquaintance?

An important but difficult news item?

There’s no right answer.

Facebook optimizes
for engagement on the site:

likes, shares, comments.

In August of 2014,

protests broke out in Ferguson, Missouri,

after the killing of an African-American
teenager by a white police officer,

under murky circumstances.

The news of the protests was all over

my algorithmically
unfiltered Twitter feed,

but nowhere on my Facebook.

Was it my Facebook friends?

I disabled Facebook’s algorithm,

which is hard because Facebook
keeps wanting to make you

come under the algorithm’s control,

and saw that my friends
were talking about it.

It’s just that the algorithm
wasn’t showing it to me.

I researched this and found
this was a widespread problem.

The story of Ferguson
wasn’t algorithm-friendly.

It’s not “likable.”

Who’s going to click on “like?”

It’s not even easy to comment on.

Without likes and comments,

the algorithm was likely showing it
to even fewer people,

so we didn’t get to see this.

Instead, that week,

Facebook’s algorithm highlighted this,

which is the ALS Ice Bucket Challenge.

Worthy cause; dump ice water,
donate to charity, fine.

But it was super algorithm-friendly.

The machine made this decision for us.

A very important
but difficult conversation

might have been smothered,

had Facebook been the only channel.

Now, finally, these systems
can also be wrong

in ways that don’t resemble human systems.

Do you guys remember Watson,
IBM’s machine-intelligence system

that wiped the floor
with human contestants on Jeopardy?

It was a great player.

But then, for Final Jeopardy,
Watson was asked this question:

“Its largest airport is named
for a World War II hero,

its second-largest
for a World War II battle.”

(Hums Final Jeopardy music)

Chicago.

The two humans got it right.

Watson, on the other hand,
answered “Toronto” –

for a US city category!

The impressive system also made an error

that a human would never make,
a second-grader wouldn’t make.

Our machine intelligence can fail

in ways that don’t fit
error patterns of humans,

in ways we won’t expect
and be prepared for.

It’d be lousy not to get a job
one is qualified for,

but it would triple suck
if it was because of stack overflow

in some subroutine.

(Laughter)

In May of 2010,

a flash crash on Wall Street
fueled by a feedback loop

in Wall Street’s “sell” algorithm

wiped a trillion dollars
of value in 36 minutes.

I don’t even want to think
what “error” means

in the context of lethal
autonomous weapons.

So yes, humans have always made biases.

Decision makers and gatekeepers,

in courts, in news, in war …

they make mistakes;
but that’s exactly my point.

We cannot escape
these difficult questions.

We cannot outsource
our responsibilities to machines.

(Applause)

Artificial intelligence does not give us
a “Get out of ethics free” card.

Data scientist Fred Benenson
calls this math-washing.

We need the opposite.

We need to cultivate algorithm suspicion,
scrutiny and investigation.

We need to make sure we have
algorithmic accountability,

auditing and meaningful transparency.

We need to accept
that bringing math and computation

to messy, value-laden human affairs

does not bring objectivity;

rather, the complexity of human affairs
invades the algorithms.

Yes, we can and we should use computation

to help us make better decisions.

But we have to own up
to our moral responsibility to judgment,

and use algorithms within that framework,

not as a means to abdicate
and outsource our responsibilities

to one another as human to human.

Machine intelligence is here.

That means we must hold on ever tighter

to human values and human ethics.

Thank you.

(Applause)