- 
Not Synced
So I started my first job
as a computer programmer
 
- 
Not Synced
in my very first year of college,
 
- 
Not Synced
basically as a teenager.
 
- 
Not Synced
Soon after I started working,
 
- 
Not Synced
writing software in a company,
 
- 
Not Synced
a manager who worked at the company
came down to where I was,
 
- 
Not Synced
and he whispered to me,
 
- 
Not Synced
"Can he tell if I'm lying?"
 
- 
Not Synced
There was nobody else in the room.
 
- 
Not Synced
"Can who tell if you're lying,
and why are we whispering?"
 
- 
Not Synced
The manager pointed
at the computer in the room.
 
- 
Not Synced
"Can he tell if I'm lying?"
 
- 
Not Synced
Well, that manager was having
an affair with the receptionist,
 
- 
Not Synced
and I was still a teenager,
 
- 
Not Synced
so I whisper-shouted back to him,
 
- 
Not Synced
"Yes, the computer can tell
if you're lying."
 
- 
Not Synced
(Laughter)
 
- 
Not Synced
Well, I laughed, but actually
the laugh's on me.
 
- 
Not Synced
Nowadays, there are computational systems
that can suss out emotional states,
 
- 
Not Synced
and even lying,
 
- 
Not Synced
from processing human faces.
 
- 
Not Synced
Advertisers and even governments
are very interested.
 
- 
Not Synced
I had become a computer programmer
because I was one of those kids
 
- 
Not Synced
crazed about math and science,
 
- 
Not Synced
but somewhere along the line
I'd learned about nuclear weapons,
 
- 
Not Synced
and I'd gotten really concerned
with the ethics of science.
 
- 
Not Synced
I was troubled.
 
- 
Not Synced
However, because of family circumstances,
 
- 
Not Synced
I also needed to start working
as soon as possible.
 
- 
Not Synced
So I thought to myself, hey,
let me pick a technical field
 
- 
Not Synced
where I can get a job easily
 
- 
Not Synced
and where I don't have to deal
with any troublesome questions of ethics.
 
- 
Not Synced
So I picked computer.
 
- 
Not Synced
Well ha ha ha, all the laughs are on me.
 
- 
Not Synced
Nowadays, computer scientists
are building platforms that control
 
- 
Not Synced
what a billion people see every day.
 
- 
Not Synced
They're developing cars that
could decide who to run over.
 
- 
Not Synced
They're even building machines, weapons,
that might kill human beings in war.
 
- 
Not Synced
It's ethics all the way down.
 
- 
Not Synced
Machine intelligence is here.
 
- 
Not Synced
We're now using computation
to make all sort of decisions,
 
- 
Not Synced
but also new kinds of decisions.
 
- 
Not Synced
We're asking questions to computation
that have no single right answers,
 
- 
Not Synced
that are subjective and open-ended
and value-laden.
 
- 
Not Synced
We're asking questions like,
 
- 
Not Synced
"Who should the company hire?"
 
- 
Not Synced
"Which update from which friend
should you be shown?"
 
- 
Not Synced
"Which convict is more
likely to re-offend?"
 
- 
Not Synced
"Which news item or movie
should be recommended to people?"
 
- 
Not Synced
Look, yes, we've been using
computers for a while,
 
- 
Not Synced
but this is different.
 
- 
Not Synced
This is a historical twist,
 
- 
Not Synced
because we cannot anchor
computation for such subjective decisions
 
- 
Not Synced
the way we can anchor computation
 
- 
Not Synced
for flying airplanes, building bridges,
going to the moon.
 
- 
Not Synced
Are airplanes safer?
Did the bridge sway and fall?
 
- 
Not Synced
There we have agreed-upon,
fairly clear benchmarks,
 
- 
Not Synced
and we have laws of nature to guide us.
 
- 
Not Synced
We have no such anchors and benchmarks
 
- 
Not Synced
for decisions in messy human affairs.
 
- 
Not Synced
To make things more complicated,
our software is getting more powerful,
 
- 
Not Synced
but it's also getting less transparent
and more complex.
 
- 
Not Synced
Recently, in the past decade,
 
- 
Not Synced
complex algorithms have made
great strides.
 
- 
Not Synced
They can recognize human faces.
 
- 
Not Synced
They can decipher handwriting.
 
- 
Not Synced
They can detect credit card fraud
and block spam
 
- 
Not Synced
and they can translate between languages.
 
- 
Not Synced
They can detect tumors in medical imaging.
 
- 
Not Synced
They can beat humans in chess and go.
 
- 
Not Synced
Much of this progress comes from
a method called machine learning.
 
- 
Not Synced
Machine learning is different
than traditional programming,
 
- 
Not Synced
where you give the computer
detailed, exact, painstaking instructions.
 
- 
Not Synced
It's more like you take the system
and you feed it lot of data,
 
- 
Not Synced
including unstructured data,
 
- 
Not Synced
like the kind we generate
in our digital lives.
 
- 
Not Synced
And the system learns
by churning through this data.
 
- 
Not Synced
And also crucially, these systems
don't operate under a single answer logic.
 
- 
Not Synced
They don't produce a simple answer.
 
- 
Not Synced
It's more probabilistic.
 
- 
Not Synced
This one is probably more
like what you're looking for.
 
- 
Not Synced
Now the upside is, this method
is really powerful.
 
- 
Not Synced
The head of Google's AI systems called it
the unreasonable effectiveness of data.
 
- 
Not Synced
The downside is we don't really
understand what the system learned.
 
- 
Not Synced
In fact, that's its power.
 
- 
Not Synced
This is less like giving instructions
to a computer.
 
- 
Not Synced
It's more like training a puppy
machine creature
 
- 
Not Synced
we don't really understand or control.
 
- 
Not Synced
So this is our problem.
 
- 
Not Synced
It's a problem when this artificial
intelligence system gets things wrong.
 
- 
Not Synced
It's also a problem
 
- 
Not Synced
when it gets things right,
 
- 
Not Synced
because we don't even know which is which
when it's a subjective problem.
 
- 
Not Synced
We don't know what this thing is thinking.
 
- 
Not Synced
So consider a hiring algorithm,
 
- 
Not Synced
a system used to hire people
using machine learning systems.
 
- 
Not Synced
Such a system would have been trained
on previous employees' data
 
- 
Not Synced
an instructed to find and hire
people like the existing
 
- 
Not Synced
high performers in the company.
 
- 
Not Synced
Sounds good.
 
- 
Not Synced
I once attended a conference
that brought together
 
- 
Not Synced
human resources, managers,
and executives, high-level people,
 
- 
Not Synced
using such systems in hiring.
 
- 
Not Synced
They were super-excited.
 
- 
Not Synced
They thought that this would make hiring
more objective, less biased,
 
- 
Not Synced
and give women
and minorities a better shot
 
- 
Not Synced
against biased human managers.
 
- 
Not Synced
And look, human hiring is biased.
 
- 
Not Synced
I know.
 
- 
Not Synced
In one of my early jobs as a programmer,
 
- 
Not Synced
my immediate manager would sometimes
come down to where I was
 
- 
Not Synced
really early in the morning
 
- 
Not Synced
or really late in the afternoon,
 
- 
Not Synced
and she'd say, "Zeynep,
let's go to lunch!"
 
- 
Not Synced
I'd be puzzled by the weird timing.
 
- 
Not Synced
It's 4 pm? Lunch?
 
- 
Not Synced
I was broke, so free lunch. I always went.
 
- 
Not Synced
I later realized what was happening.
 
- 
Not Synced
My immediate managers
had not confessed to their higher-ups
 
- 
Not Synced
that the programmer they hired
for a serious job was a teen girl
 
- 
Not Synced
who wore jeans and sneakers to work.
 
- 
Not Synced
I was doing a good job.
 
- 
Not Synced
I just looked wrong and was
the wrong age and gender.
 
- 
Not Synced
So hiring in a gender- and race-blind way
 
- 
Not Synced
certainly sounds good to me.
 
- 
Not Synced
But with these systems,
it is more complicated, and here's why.
 
- 
Not Synced
Currently, computational systems
can infer all sorts of things about you
 
- 
Not Synced
from your digital crumbs,
 
- 
Not Synced
even if you have not
disclosed those things.
 
- 
Not Synced
They can infer your sexual orientation,
 
- 
Not Synced
your personality traits,
 
- 
Not Synced
your political leanings.
 
- 
Not Synced
They have predictive power
with high levels of accuracy,
 
- 
Not Synced
remember for things
you haven't even disclosed.
 
- 
Not Synced
This is inference.
 
- 
Not Synced
I have a friend who developed
such computational systems
 
- 
Not Synced
to predict the likelihood of clinical
or post-partum depression
 
- 
Not Synced
from social media data.
 
- 
Not Synced
The results were impressive.
 
- 
Not Synced
Her system can predict
the likelihood of depression
 
- 
Not Synced
months before the onset of any symptoms,
 
- 
Not Synced
months before.
 
- 
Not Synced
No symptoms, there's prediction.
 
- 
Not Synced
She hopes it will be used
for early intervention.
 
- 
Not Synced
Great.
 
- 
Not Synced
But now put this in the context of hiring.
 
- 
Not Synced
So at this human resources
manager's conference,
 
- 
Not Synced
I approached a high-level manager
 
- 
Not Synced
in a very large company,
 
- 
Not Synced
and I said to her, "Look,
what if, unbeknownst to you,
 
- 
Not Synced
your system is weeding out people with
high future likelihood of depression?
 
- 
Not Synced
They're not depressed now,
just maybe in the future, more likely.
 
- 
Not Synced
What if it's weeding out women
more likely to be pregnant
 
- 
Not Synced
in the next year or two
but aren't pregnant now?
 
- 
Not Synced
What if it's hiring aggressive people
because that's your workplace culture?"
 
- 
Not Synced
You can't tell this by looking
at gender breakdowns.
 
- 
Not Synced
Those may be balanced.
 
- 
Not Synced
And since this is machine learning,
not traditional coding,
 
- 
Not Synced
there is no variable there
 
- 
Not Synced
labeled "higher risk of depression,"
 
- 
Not Synced
"higher risk of pregnancy,"
 
- 
Not Synced
"aggressive guy scale."
 
- 
Not Synced
Not only do you not know
what your system is selecting on,
 
- 
Not Synced
you don't even know
where to begin to look.
 
- 
Not Synced
It's a black box.
 
- 
Not Synced
It has predictive power,
but you don't understand it.
 
- 
Not Synced
"What safeguards," I asked, "do you have
to make sure that your black box
 
- 
Not Synced
isn't doing something shady?"
 
- 
Not Synced
So she looked at me
 
- 
Not Synced
as if I had just stepped
on 10 puppy tails.
 
- 
Not Synced
She stared at me and she said,
 
- 
Not Synced
"I don't want to hear
another word about this."
 
- 
Not Synced
And she turned around and walked away.
 
- 
Not Synced
Mind you, she wasn't rude.
 
- 
Not Synced
It was clearly what I don't know
isn't my problem, go away, death stare.
 
- 
Not Synced
Look, such a system may even be less
biased than human managers in some ways,
 
- 
Not Synced
and it could make monetary sense,
 
- 
Not Synced
but it could also lead to a steady but
stealthy shutting out of the job market
 
- 
Not Synced
of people with higher risk of depression.
 
- 
Not Synced
Is this the kind of society
we want to build
 
- 
Not Synced
without even knowing we've done this
 
- 
Not Synced
because we turned decision-making
to machines we don't totally understand?
 
- 
Not Synced
Another problem is this:
 
- 
Not Synced
these systems are often trained
on data generated by our actions,
 
- 
Not Synced
human imprints.
 
- 
Not Synced
Well, they could just be
reflecting our biases,
 
- 
Not Synced
and these systems could be
picking up on our biases
 
- 
Not Synced
and amplifying them
 
- 
Not Synced
and showing them back to us,
 
- 
Not Synced
while we're telling ourselves, "We're just
doing objective neutral computation."
 
- 
Not Synced
Researchers found that on Google,
 
- 
Not Synced
women are less likely than men
 
- 
Not Synced
to be shown job ads for high-paying jobs,
 
- 
Not Synced
and searching for African-American names
 
- 
Not Synced
is more likely to bring up ads
suggesting criminal history,
 
- 
Not Synced
even when there is none.
 
- 
Not Synced
Such hidden biases
and black box algorithms
 
- 
Not Synced
that researchers uncover sometimes
 
- 
Not Synced
but sometimes we don't know
 
- 
Not Synced
can have life-altering consequences.