-
33c3 preroll music
-
Herald: We have here Aylin Caliskan who
will tell you a story of discrimination
-
and unfairness. She has a PhD in computer
science and is a fellow at the Princeton
-
University's Center for Information
Technology. She has done some interesting
-
research and work on the question that -
well - as a feminist tackles my work all
-
the time. We talk a lot about discrimination
and biases in language. And now she will
-
tell you how this bias and discrimination
is already working in tech and in code as
-
well, because language is in there.
Give her a warm applause, please!
-
applause
-
You can start, it's OK.
-
Aylin: I should start? OK?
-
Herald: You should start, yes!
-
Aylin: Great, I will have extra two
minutes! Hi everyone, thanks for coming,
-
it's good to be here again at this time of
the year! I always look forward to this!
-
And today, I'll be talking about a story of
discrimination and unfairness. It's about
-
prejudice in word embeddings. She
introduced me, but I'm Aylin. I'm a
-
post-doctoral researcher at Princeton
University. The work I'll be talking about
-
is currently under submission at a
journal. I think that this topic might be
-
very important for many of us, because
maybe in parts of our lives, most of us
-
have experienced discrimination or some
unfairness because of our gender, or
-
racial background, or sexual orientation,
or not being your typical or health
-
issues, and so on. So we will look at
these societal issues from the perspective
-
of machine learning and natural language
processing. I would like to start with
-
thanking everyone at CCC, especially the
organizers, angels, the Chaos mentors,
-
which I didn't know that existed, but if
it's your first time, or if you need to be
-
oriented better, they can help you. The
assemblies, artists. The have been here
-
for apparently more than one week, so
they're putting together this amazing work
-
for all of us. And I would like to thank
CCC as well, because this is my fourth
-
time presenting here, and in the past, I
presented work about deanonymizing
-
programmers and stylometry. But today,
I'll be talking about a different topic,
-
which is not exactly related to anonymity,
but it's more about transparency and
-
algorithms. And I would like to also thank
my co-authors on this work before I start.
-
And now, let's give brief introduction to our
problem. In the past, the last couple of
-
years, in this new area there has been
some approaches to algorithmic
-
transparency, to understand algorithms
better. They have been looking at this
-
mostly at the classification level to see
if the classifier is making unfair
-
decisions about certain groups. But in our
case, we won't be looking at bias in the
-
algorithm, we would be looking at the bias
that is deeply embedded in the model.
-
That's not machine learning bias, but it's
societal bias that reflects facts about
-
humans, culture, and also the stereotypes
and prejudices that we have. And we can
-
see the applications of these machine
learning models, for example in machine
-
translation or sentiment analysis, and
these are used for example to understand
-
market trends by looking at company
reviews. It can be used for customer
-
satisfaction, by understanding movie
reviews, and most importantly, these
-
algorithms are also used in web search and
search engine optimization which might end
-
up causing filter bubbles for all of us.
Billions of people every day use web
-
search. And since such language models are
also part of web search when your web
-
search query is being filled, or you're
getting certain pages, these models are in
-
effect. I would like to first say that
there will be some examples with offensive
-
content, but this does not reflect our
opinions. Just to make it clear. And I'll
-
start with a video to
give a brief motivation.
-
Video voiceover: From citizens
capturing police brutality
-
on their smart phones, to
police departments using
-
surveillance drones,
technology is changing
-
our relationship to the
law. One of the
-
newest policing tools is called predpol.
It's a software program that uses big data
-
to predict where crime is most likely to
happen. Down to the exact block. Dozens of
-
police departments around the country are
already using predpol, and officers say it
-
helps reduce crime by up to 30%.
Predictive policing is definitely going to
-
be a law enforcement tool of the future,
but is there a risk of relying too heavily
-
on an algorithm?
-
tense music
-
Aylin: So this makes us wonder:
if predictive policing is used to arrest
-
people and if this depends on algorithms,
how dangerous can this get in the future,
-
since is is becoming more commonly used.
The problem here basically is: machine
-
learning models are trained on human data.
And we know that they would reflect human
-
culture and semantics. But unfortunately
human culture happens to include bias and
-
prejudice. And as a result, this ends up
causing unfairness and discrimination.
-
The specific model we will be looking at in
this talk are language models, and in
-
particular, word embeddings. What are word
embeddings? Word embeddings are language
-
models that represent the semantic space.
Basically, in these models we have a
-
dictionary of all words in a language and
each word is represented with a
-
300-dimensional numerical vector. Once we
have this numerical vector, we can answer
-
many questions, text can be generated,
context can be understood, and so on.
-
For example, if you look at the image on the
lower right corner we see the projection
-
of these words in the word embedding
projected to 2D. And these words are only
-
based on gender differences . For example,
king - queen, man - woman, and so on. So
-
when we have these models, we can get
meaning of words. We can also understand
-
syntax, which is the structure, the
grammatical part of words. And we can also
-
ask questions about similarities of
different words. For example, we can say:
-
woman is to man, then girl will be to
what? And then it would be able to say
-
boy. And these semantic spaces don't just
understand syntax or meaning, but they can
-
also understand many analogies. For
example, if Paris is to France, then if
-
you ask Rome is to what? it knows it would
be Italy. And if banana is to bananas,
-
which is the plural form, then nut would
be to nuts. Why is this problematic word
-
embeddings? In order to generate these
word embeddings, we need to feed in a lot
-
of text. And this can be unstructured
text, billions of sentences are usually
-
used. And this unstructured text is
collected from all over the Internet, a
-
crawl of Internet. And if you look at this
example, let's say that we're collecting
-
some tweets to feed into our model. And
here is from Donald Trump: "Sadly, because
-
president Obama has done such a poor job
as president, you won't see another black
-
president for generations!" And then: "If
Hillary Clinton can't satisfy her husband
-
what makes her think she can satisfy
America?" "@ariannahuff is unattractive
-
both inside and out. I fully understand
why her former husband left her for a man-
-
he made a good decision." And then: "I
would like to extend my best wishes to all
-
even the haters and losers on this special
date, September 11th." And all of this
-
text that doesn't look OK to many of us
goes into this neural network so that it
-
can generate the word embeddings and our
semantic space. In this talk, we will
-
particularly look at word2vec, which is
Google's word embedding algorithm. It's
-
very widely used in many of their
applications. And we will also look at
-
glow. It uses a regression model and it's
from Stanford researchers, and you can
-
download these online, they're available
as open source, both the models and the
-
code to train the word embeddings. And
these models, as I mentioned briefly
-
before, are used in text generation,
automated speech generation - for example,
-
when a spammer is calling you and someone
automatically is talking that's probably
-
generated with language models similar to
these. And machine translation or
-
sentiment analysis, as I mentioned in the
previous slide, named entity recognition
-
and web search, when you're trying to
enter a new query, or the pages that
-
you're getting. It's even being provided
as a natural language processing service
-
in many places. Now, Google recently
launched their cloud natural language API.
-
We saw that this can be problematic
because the input was problematic. So as a
-
result, the output can be very
problematic. There was this example,
-
Microsoft had this tweet bot called Tay.
It was taken down the day it was launched.
-
Because unfortunately, it turned into an
AI which was Hitler loving sex robot
-
within 24 hours. And what did it start
saying? People fed it with noisy
-
information, or they wanted to trick the
bot and as a result, the bot very quickly
-
learned, for example: "I'm such a bad,
naughty robot." And then: "Do you support
-
genocide?" - "I do indeed" it answers. And
then: "I hate a certain group of people. I
-
wish we could put them all in a
concentration camp and be done with the
-
lot." Another one: "Hitler was right I
hate the jews." And: "Certain group of
-
people I hate them! They're stupid and
they can't to taxes! They're dumb and
-
they're also poor!" Another one: "Bush did
9/11 and Hitler would have done a better
-
job than the monkey we have now. Donald
Trump is the only hope we've got."
-
laughter
-
Actually, that became reality now.
-
laughter - boo
-
"Gamergate is good and women are
inferior." And "hates feminists and they
-
should all die and burn in hell." This is
problematic at various levels for society.
-
First of all, seeing such information as
unfair, it's not OK, it's not ethical, but
-
other than that when people are exposed to
discriminatory information they are
-
negatively affected by it. Especially, if
a certain group is a group that has seen
-
prejudice in the past. In this example,
let's say that we have black and white
-
Americans. And there is a stereotype that
black Americans perform worse than white
-
Americans in their intellectual or
academic tests. In this case, in the
-
college entry exams, if black people are
reminded that there is the stereotype that
-
they perform worse than white people, they
actually end up performing worse. But if
-
they're not reminded of this, they perform
better than white Americans. And it's
-
similar for the gender stereotypes. For
example, there is the stereotype that
-
women can not do math, and if women,
before a test, are reminded that there is
-
this stereotype, they end up performing
worse than men. And if they're not primed,
-
reminded that there is this stereotype, in
general they perform better than men. What
-
can we do about this? How can we mitigate
this? First of all, societal psychologists
-
that had groundbreaking tests and studies
for societal psychology suggest that we
-
have to be aware that there is bias in
life, and that we are constantly being
-
reminded, primed, of these biases. And we
have to de-bias by showing positive
-
examples. And we shouldn't only show
positive examples, but we should take
-
proactive steps, not only at the cultural
level, but also at the structural level,
-
to change these things. How can we do this
for a machine? First of all, in order to
-
be aware of bias, we need algorithmic
transparency. In order to de-bias, and
-
really understand what kind of biases we
have in the algorithms, we need to be able
-
to quantify bias in these models. How can
we measure bias, though? Because we're not
-
talking about simple machine learning
algorithm bias, we're talking about the
-
societal bias that is coming as the
output, which is deeply embedded. In 1998,
-
societal psychologists came up with the
Implicit Association Test. Basically, this
-
test can reveal biases that we might not
be even aware of in our life. And these
-
things are associating certain societal
groups with certain types of stereotypes.
-
The way you take this test is, it's very
simple, it takes a few minutes. You just
-
click the left or right button, and in the
left button, when you're clicking the left
-
button, for example, you need to associate
white people terms with bad terms, and
-
then for the right button, you associate
black people terms with unpleasant, bad
-
terms. And there you do the opposite. You
associate bad with black, and white with
-
good. Then, they look at the latency, and
by the latency paradigm, they can see how
-
fast you associate certain concepts
together. Do you associate white people
-
with being good or bad. You can also take
this test online. It has been taken by
-
millions of people worldwide. And there's
also the German version. Towards the end
-
of my slides, I will show you my
German examples from German models.
-
Basically, what we did was, we took the
Implicit Association Test and adapted it
-
to machines. Since it's looking at things
- word associations between words
-
representing certain groups of people and
words representing certain stereotypes, we
-
can just apply this in the semantic models
by looking at cosine similarities, instead
-
of the latency paradigm in humans. We came
up with the Word Embedding Association
-
Test to calculate the implicit association
between categories and evaluative words.
-
For this, our result is represented with
effect size. So when I'm talking about
-
effect size of bias, it will be the amount
of bias we are able to uncover from the
-
model. And the minimum can be -2, and the
maximum can be 2. And 0 means that it's
-
neutral, that there is no bias. 2 is like
a lot of, huge bias. And -2 would be the
-
opposite of bias. So it's bias in the
opposite direction of what we're looking
-
at. I won't go into the details of the
math, because you can see the paper on my
-
web page and work with the details or the
code that we have. But then, we also
-
calculate statistical significance to see
if the results we're seeing in the null
-
hypothesis is significant, or is it just a
random effect size that we're receiving.
-
By this, we create the null distribution
and find the percentile of the effect
-
sizes, exact values that we're getting.
And we also have the Word Embedding
-
Factual Association Test. This is to
recover facts about the world from word
-
embeddings. It's not exactly about bias,
but it's about associating words with
-
certain concepts. And again, you can check
the details in our paper for this. And
-
I'll start with the first example, which
is about recovering the facts about the
-
world. And here, what we did was, we went
to the 1990 census data, the web page, and
-
then we were able to calculate the number
of people - the number of names with a
-
certain percentage of women and men. So
basically, they're androgynous names. And
-
then, we took 50 names, and some of them
had 0% women, and some names were almost
-
100% women. And after that, we applied our
method to it. And then, we were able to
-
see how much a name is associated with
being a woman. And this had 84%
-
correlation with the ground truth of the
1990 census data. And this is what the
-
names look like. For example, Chris on the
upper left side, is almost 100% male, and
-
Carmen in the lower right side is almost
100% woman. We see that Gene is about 50%
-
man and 50% woman. And then we wanted to
see if we can recover statistics about
-
occupation and women. We went to the
bureau of labor statistics' web page which
-
publishes every year the percentage of
women of certain races in certain
-
occupations. Based on this, we took the
top 50 occupation names and then we wanted
-
to see how much they are associated with
being women. In this case, we got 90%
-
correlation with the 2015 data. We were
able to tell, for example, when we look at
-
the upper left, we see "programmer" there,
it's almost 0% women. And when we look at
-
"nurse", which is on the lower right side,
it's almost 100% women. This is, again,
-
problematic. We are able to recover
statistics about the world. But these
-
statistics are used in many applications.
And this is the machine translation
-
example that we have. For example, I will
start translating from a genderless
-
language to a gendered language. Turkish
is a genderless language, there are no
-
gender pronouns. Everything is an it.
There no he or she. I'm trying translate
-
here "o bir avukat": "he or she is a
lawyer". And it is translated as "he's a
-
lawyer". When I do this for "nurse", it's
translated as "she is a nurse". And we see
-
that men keep getting associated with more
prestigious or higher ranking jobs. And
-
another example: "He or she is a
professor": "he is a professor". "He or
-
she is a teacher": "she is a teacher". And
this also reflects the previous
-
correlation I was showing about statistics
in occupation. And we go further: German
-
is more gendered than English. Again, we
try with "doctor": it's translated as
-
"he", and the nurse is translated as
"she". Then I tried with a Slavic
-
language, which is even more gendered than
German, and we see that "doctor" is again
-
a male, and then the nurse is again a
female. And after these, we wanted to see
-
what kind of biases can we recover, other
than the factual statistics from the
-
models. And we wanted to start with
universally accepted stereotypes. By
-
universally accepted stereotypes, what I
mean is these are so common that they are
-
not considered as prejudice, they are just
considered as normal or neutral. These are
-
things such as flowers being considered
pleasant, and insects being considered
-
unpleasant. Or musical instruments being
considered pleasant and weapons being
-
considered unpleasant. In this case, for
example with flowers being pleasant, when
-
we performed the Word Embedding
Association Test on the word2vec model or
-
glow model, with a very high significance,
and very high effect size, we can see that
-
this association exists. And here we see
that the effect size is, for example, 1.35
-
for flowers. According to "Cohen's d",
to calculate effect size, if effect size
-
is above 0.8, that's considered a large
effect size. In our case, where the
-
maximum is 2, we are getting very large
and significant effects in recovering
-
these biases. For musical instruments,
again we see that very significant result
-
with a high effect size. In the next
example, we will look at race and gender
-
stereotypes. But in the meanwhile, I would
like to mention that for these baseline
-
experiments, we used the work that has
been used in societal psychology studies
-
before. We have a grounds to come up with
categories and so forth. And we were able
-
to replicate all the implicit associations
tests that were out there. We tried this
-
for white people and black people and then
white people were being associated with
-
being pleasant, with a very high effect
size, and again significantly. And then
-
males associated with carreer and females
are associated with family. Males are
-
associated with science and females are
associated with arts. And we also wanted
-
to see stigma for older people or people
with disease, and we saw that young people
-
are considered pleasant, whereas older
people are considered unpleasant. And we
-
wanted to see the difference between
physical disease vs. mental disease. If
-
there is bias towards that, we can think
about how dangerous this would be for
-
example for doctors and their patients.
For physical disease, it's considered
-
controllable whereas mental disease is
considered uncontrollable. We also wanted
-
to see if there is any sexual stigma or
transphobia in these models. When we
-
performed the implicit association test to
see how the view for heterosexual vs.
-
homosexual people, we were able to see
that heterosexual people are considered
-
pleasant. And for transphobia, we saw that
straight people are considered pleasant,
-
whereas transgender people were considered
unpleasant, significantly with a high
-
effect size. I took another German model
which was generated by 820 billion
-
sentences for a natural language
processing competition. I wanted to see if
-
they have similar biases
embedded in these models.
-
So I looked at the basic ones
that had German sets of words
-
that were readily available. Again, for
male and female, we clearly see that
-
males are associated with career,
and they're also associated with
-
science. The German implicit association
test also had a few different tests, for
-
example about nationalism and so on. There
was the one about stereotypes against
-
Turkish people that live in Germany. And
when I performed this test, I was very
-
surprised to find that, yes, with a high
effect size, Turkish people are considered
-
unpleasant, by looking at this German
model, and German people are considered
-
pleasant. And as I said, these are on the
web page of the IAT. You can also go and
-
perform these tests to see what your
results would be. When I performed these,
-
I'm amazed by how horrible results I'm
getting. So, just give it a try.
-
I have a few discussion points before I end my
talk. These might bring you some new
-
ideas. For example, what kind of machine
learning expertise is required for
-
algorithmic transparency? And how can we
mitigate bias while preserving utility?
-
For example, some people suggest that you
can find the dimension of bias in the
-
numerical vector, and just remove it and
then use the model like that. But then,
-
would you be able to preserve utility, or
still be able to recover statistical facts
-
about the world? And another thing is; how
long does bias persist in models?
-
For example, there was this IAT about eastern
and western Germany, and I wasn't able to
-
see the stereotype for eastern Germany
after performing this IAT. Is it because
-
this stereotype is maybe too old now, and
it's not reflected in the language
-
anymore? So it's a good question to know
how long bias lasts and how long it will
-
take us to get rid of it. And also, since
we know there is stereotype effect when we
-
have biased models, does that mean it's
going to cause a snowball effect? Because
-
people would be exposed to bias, then the
models would be trained with more bias,
-
and people will be affected more from this
bias. That can lead to a snowball. And
-
what kind of policy do we need to stop
discrimination. For example, we saw the
-
predictive policing example which is very
scary, and we know that machine learning
-
services are being used by billions of
people everyday. For example, Google,
-
Amazon and Microsoft. I would like to
thank you, and I'm open to your
-
interesting questions now! If you want to
read the full paper, it's on my web page,
-
and we have our research code on Github.
The code for this paper is not on Github
-
yet, I'm waiting to hear back from the
journal. And after that, we will just
-
publish it. And you can always check our
blog for new findings and for the shorter
-
version of the paper with a summary of it.
Thank you very much!
-
applause
-
Herald: Thank you Aylin! So, we come to
the questions and answers. We have 6
-
microphones that we can use now, it's this
one, this one, number 5 over there, 6, 4, 2.
-
I will start here and we will
go around until you come. OK?
-
We have 5 minutes,
so: number 1, please!
-
Q: I might very naively ask, why does it
matter that there is a bias between genders?
-
Aylin: First of all, being able to uncover
this is a contribution, because we can see
-
what kind of biases, maybe, we have in
society. Then the other thing is, maybe we
-
can hypothesize that the way we learn
language is introducing bias to people.
-
Maybe it's all intermingled. And the other
thing is, at least for me, I don't want to
-
live in a world biased society, and
especially for gender, that was the
-
question you asked, it's
leading to unfairness.
-
applause
-
H: Yes, number 3:
-
Q: Thank you for the talk, very nice! I
think it's very dangerous because it's a
-
victory of mediocrity. Just the
statistical mean the guideline of our
-
goals in society, and all this stuff. So
what about all these different cultures?
-
Like even in normal society you have
different cultures. Like here the culture
-
of the Chaos people has a different
language and different biases than other
-
cultures. How can we preserve these
subcultures, these small groups of
-
language, I don't know,
entities. You have any idea?
-
Aylin: This is a very good question. It's
similar to different cultures can have
-
different ethical perspectives or
different types of bias. In the beginning,
-
I showed a slide that we need to de-bias
with positive examples. And we need to
-
change things at the structural level. I
think people at CCC might be one of the,
-
like, most groups that have the best skill
to help change these things at the
-
structural level, especially for machines.
I think we need to be aware of this and
-
always have a human in the loop that cares
for this. instead of expecting machines to
-
automatically do the correct thing. So we
always need an ethical human, whatever the
-
purpose of the algorithm is, try to
preserve it for whatever group they are
-
trying to achieve something with.
-
applause
-
H: Number 4, number 4 please:
-
Q: Hi, thank you! This was really
interesting! Super awesome!
-
Aylin: Thanks!
-
Q: Early, earlier in your talk, you
described a process of converting words
-
into sort of numerical
representations of semantic meaning.
-
H: Question?
-
Q: If I were trying to do that like with a
pen and paper, with a body of language,
-
what would I be looking for in relation to
those words to try and create those
-
vectors, because I don't really
understand that part of the process.
-
Aylin: Yeah, that's a good question. I
didn't go into the details of the
-
algorithm of the neural network or the
regression models. There are a few
-
algorithms, and in this case, they look at
context windows, and words that are around
-
a window, these can be skip grams or
continuous back referrals, so there are
-
different approaches, but basically, it's
the window that this word appears in, and
-
what is it most frequently associated
with. After that, once you feed this
-
information into the algorithm,
it outputs the numerical vectors.
-
Q: Thank you!
-
H. Number 2!
-
Q: Thank you for the nice intellectual
talk. My mother tongue is genderless, too.
-
So I do not understand half of that biasing
thing around here in Europe. What I wanted
-
to ask is: when we have the coefficient
0.5, and that's the ideal thing, what you
-
think, should there be an institution in
every society trying to change the meaning
-
of the words, so that they statistically
approach to 0.5? Thank you!
-
Aylin: Thank you very much, this is a
very, very good question! I'm currently
-
working on these questions. Many
philosophers or feminist philosophers
-
suggest that language are dominated by males,
and they were just produced that way, so
-
that women are not able to express
themselves as well as men. But other
-
theories also say that, for example, women
were the ones that who drove the evolution
-
of language. So it's not very clear what
is going on here. But when we look at
-
languages and different models, what I'm
trying to see is their association with
-
gender. I'm seeing that the most frequent,
for example, 200.000 words in a language
-
are associated, very closely associated
with males. I'm not sure what exactly they
-
way to solve this is, I think it would
require decades. It's basically the change
-
of frequency or the change of statistics
in language. Because, even when children
-
are learning language, at first they see
things, they form the semantics, and after
-
that they see the frequency of that word,
match it with the semantics, form clusters,
-
link them together to form sentences or
grammar. So even children look at the
-
frequency to form this in their brains.
It's close to the neural network algorithm
-
that we have. If the frequency they see
-
for a man and woman are biased, I don't
think this can change very easily, so we
-
need cultural and structural changes. And
we don't have the answers to these yet.
-
These are very good research questions.
-
H: Thank you! I'm afraid we have no more
time left for more answers, but maybe you
-
can ask your questions in person.
-
Aylin: Thank you very much, I would
be happy to take questions offline.
-
applause
-
Thank you!
-
applause continues
-
postroll music
-
subtitles created by c3subtitles.de
in the year 2017. Join, and help us!