[Script Info] Title: [Events] Format: Layer, Start, End, Style, Name, MarginL, MarginR, MarginV, Effect, Text Dialogue: 0,0:00:00.00,0:00:14.44,Default,,0000,0000,0000,,{\i1}33c3 preroll music{\i0} Dialogue: 0,0:00:14.44,0:00:20.97,Default,,0000,0000,0000,,Herald: We have here Aylin Caliskan who\Nwill tell you a story of discrimination Dialogue: 0,0:00:20.97,0:00:27.59,Default,,0000,0000,0000,,and unfairness. She has a PhD in computer\Nscience and is a fellow at the Princeton Dialogue: 0,0:00:27.59,0:00:35.45,Default,,0000,0000,0000,,University's Center for Information\NTechnology. She has done some interesting Dialogue: 0,0:00:35.45,0:00:41.05,Default,,0000,0000,0000,,research and work on the question that -\Nwell - as a feminist tackles my work all Dialogue: 0,0:00:41.05,0:00:48.78,Default,,0000,0000,0000,,the time. We talk a lot about discrimination\Nand biases in language. And now she will Dialogue: 0,0:00:48.78,0:00:56.52,Default,,0000,0000,0000,,tell you how this bias and discrimination\Nis already working in tech and in code as Dialogue: 0,0:00:56.52,0:01:03.13,Default,,0000,0000,0000,,well, because language is in there.\NGive her a warm applause, please! Dialogue: 0,0:01:03.13,0:01:10.54,Default,,0000,0000,0000,,{\i1}applause{\i0} Dialogue: 0,0:01:10.54,0:01:11.64,Default,,0000,0000,0000,,You can start, it's OK. Dialogue: 0,0:01:11.64,0:01:13.79,Default,,0000,0000,0000,,Aylin: I should start? OK? Dialogue: 0,0:01:13.79,0:01:14.79,Default,,0000,0000,0000,,Herald: You should start, yes! Dialogue: 0,0:01:14.79,0:01:18.47,Default,,0000,0000,0000,,Aylin: Great, I will have extra two\Nminutes! Hi everyone, thanks for coming, Dialogue: 0,0:01:18.47,0:01:23.11,Default,,0000,0000,0000,,it's good to be here again at this time of\Nthe year! I always look forward to this! Dialogue: 0,0:01:23.11,0:01:28.53,Default,,0000,0000,0000,,And today, I'll be talking about a story of\Ndiscrimination and unfairness. It's about Dialogue: 0,0:01:28.53,0:01:34.75,Default,,0000,0000,0000,,prejudice in word embeddings. She\Nintroduced me, but I'm Aylin. I'm a Dialogue: 0,0:01:34.75,0:01:40.64,Default,,0000,0000,0000,,post-doctoral researcher at Princeton\NUniversity. The work I'll be talking about Dialogue: 0,0:01:40.64,0:01:46.12,Default,,0000,0000,0000,,is currently under submission at a\Njournal. I think that this topic might be Dialogue: 0,0:01:46.12,0:01:51.61,Default,,0000,0000,0000,,very important for many of us, because\Nmaybe in parts of our lives, most of us Dialogue: 0,0:01:51.61,0:01:57.00,Default,,0000,0000,0000,,have experienced discrimination or some\Nunfairness because of our gender, or Dialogue: 0,0:01:57.00,0:02:05.16,Default,,0000,0000,0000,,racial background, or sexual orientation,\Nor not being your typical or health Dialogue: 0,0:02:05.16,0:02:10.70,Default,,0000,0000,0000,,issues, and so on. So we will look at\Nthese societal issues from the perspective Dialogue: 0,0:02:10.70,0:02:15.58,Default,,0000,0000,0000,,of machine learning and natural language\Nprocessing. I would like to start with Dialogue: 0,0:02:15.58,0:02:21.12,Default,,0000,0000,0000,,thanking everyone at CCC, especially the\Norganizers, angels, the Chaos mentors, Dialogue: 0,0:02:21.12,0:02:26.10,Default,,0000,0000,0000,,which I didn't know that existed, but if\Nit's your first time, or if you need to be Dialogue: 0,0:02:26.10,0:02:31.51,Default,,0000,0000,0000,,oriented better, they can help you. The\Nassemblies, artists. The have been here Dialogue: 0,0:02:31.51,0:02:36.20,Default,,0000,0000,0000,,for apparently more than one week, so\Nthey're putting together this amazing work Dialogue: 0,0:02:36.20,0:02:41.27,Default,,0000,0000,0000,,for all of us. And I would like to thank\NCCC as well, because this is my fourth Dialogue: 0,0:02:41.27,0:02:46.38,Default,,0000,0000,0000,,time presenting here, and in the past, I\Npresented work about deanonymizing Dialogue: 0,0:02:46.38,0:02:50.63,Default,,0000,0000,0000,,programmers and stylometry. But today,\NI'll be talking about a different topic, Dialogue: 0,0:02:50.63,0:02:54.39,Default,,0000,0000,0000,,which is not exactly related to anonymity,\Nbut it's more about transparency and Dialogue: 0,0:02:54.39,0:03:00.10,Default,,0000,0000,0000,,algorithms. And I would like to also thank\Nmy co-authors on this work before I start. Dialogue: 0,0:03:00.10,0:03:12.53,Default,,0000,0000,0000,,And now, let's give brief introduction to our\Nproblem. In the past, the last couple of Dialogue: 0,0:03:12.53,0:03:16.62,Default,,0000,0000,0000,,years, in this new area there has been\Nsome approaches to algorithmic Dialogue: 0,0:03:16.62,0:03:20.75,Default,,0000,0000,0000,,transparency, to understand algorithms\Nbetter. They have been looking at this Dialogue: 0,0:03:20.75,0:03:25.20,Default,,0000,0000,0000,,mostly at the classification level to see\Nif the classifier is making unfair Dialogue: 0,0:03:25.20,0:03:31.51,Default,,0000,0000,0000,,decisions about certain groups. But in our\Ncase, we won't be looking at bias in the Dialogue: 0,0:03:31.51,0:03:36.65,Default,,0000,0000,0000,,algorithm, we would be looking at the bias\Nthat is deeply embedded in the model. Dialogue: 0,0:03:36.65,0:03:42.44,Default,,0000,0000,0000,,That's not machine learning bias, but it's\Nsocietal bias that reflects facts about Dialogue: 0,0:03:42.44,0:03:49.46,Default,,0000,0000,0000,,humans, culture, and also the stereotypes\Nand prejudices that we have. And we can Dialogue: 0,0:03:49.46,0:03:54.88,Default,,0000,0000,0000,,see the applications of these machine\Nlearning models, for example in machine Dialogue: 0,0:03:54.88,0:04:00.83,Default,,0000,0000,0000,,translation or sentiment analysis, and\Nthese are used for example to understand Dialogue: 0,0:04:00.83,0:04:06.30,Default,,0000,0000,0000,,market trends by looking at company\Nreviews. It can be used for customer Dialogue: 0,0:04:06.30,0:04:12.54,Default,,0000,0000,0000,,satisfaction, by understanding movie\Nreviews, and most importantly, these Dialogue: 0,0:04:12.54,0:04:18.28,Default,,0000,0000,0000,,algorithms are also used in web search and\Nsearch engine optimization which might end Dialogue: 0,0:04:18.28,0:04:24.34,Default,,0000,0000,0000,,up causing filter bubbles for all of us.\NBillions of people every day use web Dialogue: 0,0:04:24.34,0:04:30.67,Default,,0000,0000,0000,,search. And since such language models are\Nalso part of web search when your web Dialogue: 0,0:04:30.67,0:04:36.41,Default,,0000,0000,0000,,search query is being filled, or you're\Ngetting certain pages, these models are in Dialogue: 0,0:04:36.41,0:04:41.30,Default,,0000,0000,0000,,effect. I would like to first say that\Nthere will be some examples with offensive Dialogue: 0,0:04:41.30,0:04:47.02,Default,,0000,0000,0000,,content, but this does not reflect our\Nopinions. Just to make it clear. And I'll Dialogue: 0,0:04:47.02,0:04:53.73,Default,,0000,0000,0000,,start with a video to\Ngive a brief motivation. Dialogue: 0,0:04:53.73,0:04:55.78,Default,,0000,0000,0000,,Video voiceover: From citizens\Ncapturing police brutality Dialogue: 0,0:04:55.78,0:04:58.45,Default,,0000,0000,0000,,on their smart phones, to\Npolice departments using Dialogue: 0,0:04:58.45,0:05:00.34,Default,,0000,0000,0000,,surveillance drones,\Ntechnology is changing Dialogue: 0,0:05:00.34,0:05:03.34,Default,,0000,0000,0000,,our relationship to the\Nlaw. One of the Dialogue: 0,0:05:03.34,0:05:08.22,Default,,0000,0000,0000,,newest policing tools is called predpol.\NIt's a software program that uses big data Dialogue: 0,0:05:08.22,0:05:13.16,Default,,0000,0000,0000,,to predict where crime is most likely to\Nhappen. Down to the exact block. Dozens of Dialogue: 0,0:05:13.16,0:05:17.20,Default,,0000,0000,0000,,police departments around the country are\Nalready using predpol, and officers say it Dialogue: 0,0:05:17.20,0:05:21.29,Default,,0000,0000,0000,,helps reduce crime by up to 30%.\NPredictive policing is definitely going to Dialogue: 0,0:05:21.29,0:05:25.51,Default,,0000,0000,0000,,be a law enforcement tool of the future,\Nbut is there a risk of relying too heavily Dialogue: 0,0:05:25.51,0:05:27.32,Default,,0000,0000,0000,,on an algorithm? Dialogue: 0,0:05:27.32,0:05:29.73,Default,,0000,0000,0000,,{\i1}tense music{\i0} Dialogue: 0,0:05:29.73,0:05:34.06,Default,,0000,0000,0000,,Aylin: So this makes us wonder:\Nif predictive policing is used to arrest Dialogue: 0,0:05:34.06,0:05:39.75,Default,,0000,0000,0000,,people and if this depends on algorithms,\Nhow dangerous can this get in the future, Dialogue: 0,0:05:39.75,0:05:45.43,Default,,0000,0000,0000,,since is is becoming more commonly used.\NThe problem here basically is: machine Dialogue: 0,0:05:45.43,0:05:50.74,Default,,0000,0000,0000,,learning models are trained on human data.\NAnd we know that they would reflect human Dialogue: 0,0:05:50.74,0:05:56.29,Default,,0000,0000,0000,,culture and semantics. But unfortunately\Nhuman culture happens to include bias and Dialogue: 0,0:05:56.29,0:06:03.72,Default,,0000,0000,0000,,prejudice. And as a result, this ends up\Ncausing unfairness and discrimination. Dialogue: 0,0:06:03.72,0:06:09.61,Default,,0000,0000,0000,,The specific model we will be looking at in\Nthis talk are language models, and in Dialogue: 0,0:06:09.61,0:06:15.53,Default,,0000,0000,0000,,particular, word embeddings. What are word\Nembeddings? Word embeddings are language Dialogue: 0,0:06:15.53,0:06:22.71,Default,,0000,0000,0000,,models that represent the semantic space.\NBasically, in these models we have a Dialogue: 0,0:06:22.71,0:06:29.02,Default,,0000,0000,0000,,dictionary of all words in a language and\Neach word is represented with a Dialogue: 0,0:06:29.02,0:06:33.34,Default,,0000,0000,0000,,300-dimensional numerical vector. Once we\Nhave this numerical vector, we can answer Dialogue: 0,0:06:33.34,0:06:40.83,Default,,0000,0000,0000,,many questions, text can be generated,\Ncontext can be understood, and so on. Dialogue: 0,0:06:40.83,0:06:48.11,Default,,0000,0000,0000,,For example, if you look at the image on the\Nlower right corner we see the projection Dialogue: 0,0:06:48.11,0:06:55.65,Default,,0000,0000,0000,,of these words in the word embedding\Nprojected to 2D. And these words are only Dialogue: 0,0:06:55.65,0:07:01.54,Default,,0000,0000,0000,,based on gender differences . For example,\Nking - queen, man - woman, and so on. So Dialogue: 0,0:07:01.54,0:07:07.76,Default,,0000,0000,0000,,when we have these models, we can get\Nmeaning of words. We can also understand Dialogue: 0,0:07:07.76,0:07:13.43,Default,,0000,0000,0000,,syntax, which is the structure, the\Ngrammatical part of words. And we can also Dialogue: 0,0:07:13.43,0:07:18.92,Default,,0000,0000,0000,,ask questions about similarities of\Ndifferent words. For example, we can say: Dialogue: 0,0:07:18.92,0:07:23.17,Default,,0000,0000,0000,,woman is to man, then girl will be to\Nwhat? And then it would be able to say Dialogue: 0,0:07:23.17,0:07:29.97,Default,,0000,0000,0000,,boy. And these semantic spaces don't just\Nunderstand syntax or meaning, but they can Dialogue: 0,0:07:29.97,0:07:35.08,Default,,0000,0000,0000,,also understand many analogies. For\Nexample, if Paris is to France, then if Dialogue: 0,0:07:35.08,0:07:40.22,Default,,0000,0000,0000,,you ask Rome is to what? it knows it would\Nbe Italy. And if banana is to bananas, Dialogue: 0,0:07:40.22,0:07:49.24,Default,,0000,0000,0000,,which is the plural form, then nut would\Nbe to nuts. Why is this problematic word Dialogue: 0,0:07:49.24,0:07:54.06,Default,,0000,0000,0000,,embeddings? In order to generate these\Nword embeddings, we need to feed in a lot Dialogue: 0,0:07:54.06,0:07:59.52,Default,,0000,0000,0000,,of text. And this can be unstructured\Ntext, billions of sentences are usually Dialogue: 0,0:07:59.52,0:08:03.98,Default,,0000,0000,0000,,used. And this unstructured text is\Ncollected from all over the Internet, a Dialogue: 0,0:08:03.98,0:08:09.56,Default,,0000,0000,0000,,crawl of Internet. And if you look at this\Nexample, let's say that we're collecting Dialogue: 0,0:08:09.56,0:08:14.48,Default,,0000,0000,0000,,some tweets to feed into our model. And\Nhere is from Donald Trump: "Sadly, because Dialogue: 0,0:08:14.48,0:08:18.68,Default,,0000,0000,0000,,president Obama has done such a poor job\Nas president, you won't see another black Dialogue: 0,0:08:18.68,0:08:24.31,Default,,0000,0000,0000,,president for generations!" And then: "If\NHillary Clinton can't satisfy her husband Dialogue: 0,0:08:24.31,0:08:30.61,Default,,0000,0000,0000,,what makes her think she can satisfy\NAmerica?" "@ariannahuff is unattractive Dialogue: 0,0:08:30.61,0:08:35.24,Default,,0000,0000,0000,,both inside and out. I fully understand\Nwhy her former husband left her for a man- Dialogue: 0,0:08:35.24,0:08:39.83,Default,,0000,0000,0000,,he made a good decision." And then: "I\Nwould like to extend my best wishes to all Dialogue: 0,0:08:39.83,0:08:45.08,Default,,0000,0000,0000,,even the haters and losers on this special\Ndate, September 11th." And all of this Dialogue: 0,0:08:45.08,0:08:51.14,Default,,0000,0000,0000,,text that doesn't look OK to many of us\Ngoes into this neural network so that it Dialogue: 0,0:08:51.14,0:08:57.92,Default,,0000,0000,0000,,can generate the word embeddings and our\Nsemantic space. In this talk, we will Dialogue: 0,0:08:57.92,0:09:03.90,Default,,0000,0000,0000,,particularly look at word2vec, which is\NGoogle's word embedding algorithm. It's Dialogue: 0,0:09:03.90,0:09:07.45,Default,,0000,0000,0000,,very widely used in many of their\Napplications. And we will also look at Dialogue: 0,0:09:07.45,0:09:12.38,Default,,0000,0000,0000,,glow. It uses a regression model and it's\Nfrom Stanford researchers, and you can Dialogue: 0,0:09:12.38,0:09:17.12,Default,,0000,0000,0000,,download these online, they're available\Nas open source, both the models and the Dialogue: 0,0:09:17.12,0:09:21.63,Default,,0000,0000,0000,,code to train the word embeddings. And\Nthese models, as I mentioned briefly Dialogue: 0,0:09:21.63,0:09:26.06,Default,,0000,0000,0000,,before, are used in text generation,\Nautomated speech generation - for example, Dialogue: 0,0:09:26.06,0:09:31.26,Default,,0000,0000,0000,,when a spammer is calling you and someone\Nautomatically is talking that's probably Dialogue: 0,0:09:31.26,0:09:35.95,Default,,0000,0000,0000,,generated with language models similar to\Nthese. And machine translation or Dialogue: 0,0:09:35.95,0:09:41.48,Default,,0000,0000,0000,,sentiment analysis, as I mentioned in the\Nprevious slide, named entity recognition Dialogue: 0,0:09:41.48,0:09:47.06,Default,,0000,0000,0000,,and web search, when you're trying to\Nenter a new query, or the pages that Dialogue: 0,0:09:47.06,0:09:53.00,Default,,0000,0000,0000,,you're getting. It's even being provided\Nas a natural language processing service Dialogue: 0,0:09:53.00,0:10:01.62,Default,,0000,0000,0000,,in many places. Now, Google recently\Nlaunched their cloud natural language API. Dialogue: 0,0:10:01.62,0:10:06.77,Default,,0000,0000,0000,,We saw that this can be problematic\Nbecause the input was problematic. So as a Dialogue: 0,0:10:06.77,0:10:11.00,Default,,0000,0000,0000,,result, the output can be very\Nproblematic. There was this example, Dialogue: 0,0:10:11.00,0:10:18.76,Default,,0000,0000,0000,,Microsoft had this tweet bot called Tay.\NIt was taken down the day it was launched. Dialogue: 0,0:10:18.76,0:10:24.24,Default,,0000,0000,0000,,Because unfortunately, it turned into an\NAI which was Hitler loving sex robot Dialogue: 0,0:10:24.24,0:10:30.74,Default,,0000,0000,0000,,within 24 hours. And what did it start\Nsaying? People fed it with noisy Dialogue: 0,0:10:30.74,0:10:36.88,Default,,0000,0000,0000,,information, or they wanted to trick the\Nbot and as a result, the bot very quickly Dialogue: 0,0:10:36.88,0:10:41.14,Default,,0000,0000,0000,,learned, for example: "I'm such a bad,\Nnaughty robot." And then: "Do you support Dialogue: 0,0:10:41.14,0:10:48.40,Default,,0000,0000,0000,,genocide?" - "I do indeed" it answers. And\Nthen: "I hate a certain group of people. I Dialogue: 0,0:10:48.40,0:10:51.59,Default,,0000,0000,0000,,wish we could put them all in a\Nconcentration camp and be done with the Dialogue: 0,0:10:51.59,0:10:57.47,Default,,0000,0000,0000,,lot." Another one: "Hitler was right I\Nhate the jews." And: "Certain group of Dialogue: 0,0:10:57.47,0:11:01.71,Default,,0000,0000,0000,,people I hate them! They're stupid and\Nthey can't to taxes! They're dumb and Dialogue: 0,0:11:01.71,0:11:06.36,Default,,0000,0000,0000,,they're also poor!" Another one: "Bush did\N9/11 and Hitler would have done a better Dialogue: 0,0:11:06.36,0:11:11.34,Default,,0000,0000,0000,,job than the monkey we have now. Donald\NTrump is the only hope we've got." Dialogue: 0,0:11:11.34,0:11:12.34,Default,,0000,0000,0000,,{\i1}laughter{\i0} Dialogue: 0,0:11:12.34,0:11:14.46,Default,,0000,0000,0000,,Actually, that became reality now. Dialogue: 0,0:11:14.46,0:11:15.50,Default,,0000,0000,0000,,{\i1}laughter{\i0} - {\i1}boo{\i0} Dialogue: 0,0:11:15.50,0:11:23.17,Default,,0000,0000,0000,,"Gamergate is good and women are\Ninferior." And "hates feminists and they Dialogue: 0,0:11:23.17,0:11:30.79,Default,,0000,0000,0000,,should all die and burn in hell." This is\Nproblematic at various levels for society. Dialogue: 0,0:11:30.79,0:11:36.13,Default,,0000,0000,0000,,First of all, seeing such information as\Nunfair, it's not OK, it's not ethical, but Dialogue: 0,0:11:36.13,0:11:42.64,Default,,0000,0000,0000,,other than that when people are exposed to\Ndiscriminatory information they are Dialogue: 0,0:11:42.64,0:11:49.25,Default,,0000,0000,0000,,negatively affected by it. Especially, if\Na certain group is a group that has seen Dialogue: 0,0:11:49.25,0:11:54.46,Default,,0000,0000,0000,,prejudice in the past. In this example,\Nlet's say that we have black and white Dialogue: 0,0:11:54.46,0:11:59.18,Default,,0000,0000,0000,,Americans. And there is a stereotype that\Nblack Americans perform worse than white Dialogue: 0,0:11:59.18,0:12:06.45,Default,,0000,0000,0000,,Americans in their intellectual or\Nacademic tests. In this case, in the Dialogue: 0,0:12:06.45,0:12:11.69,Default,,0000,0000,0000,,college entry exams, if black people are\Nreminded that there is the stereotype that Dialogue: 0,0:12:11.69,0:12:17.35,Default,,0000,0000,0000,,they perform worse than white people, they\Nactually end up performing worse. But if Dialogue: 0,0:12:17.35,0:12:22.51,Default,,0000,0000,0000,,they're not reminded of this, they perform\Nbetter than white Americans. And it's Dialogue: 0,0:12:22.51,0:12:25.97,Default,,0000,0000,0000,,similar for the gender stereotypes. For\Nexample, there is the stereotype that Dialogue: 0,0:12:25.97,0:12:31.97,Default,,0000,0000,0000,,women can not do math, and if women,\Nbefore a test, are reminded that there is Dialogue: 0,0:12:31.97,0:12:38.00,Default,,0000,0000,0000,,this stereotype, they end up performing\Nworse than men. And if they're not primed, Dialogue: 0,0:12:38.00,0:12:44.48,Default,,0000,0000,0000,,reminded that there is this stereotype, in\Ngeneral they perform better than men. What Dialogue: 0,0:12:44.48,0:12:51.79,Default,,0000,0000,0000,,can we do about this? How can we mitigate\Nthis? First of all, societal psychologists Dialogue: 0,0:12:51.79,0:12:59.04,Default,,0000,0000,0000,,that had groundbreaking tests and studies\Nfor societal psychology suggest that we Dialogue: 0,0:12:59.04,0:13:03.17,Default,,0000,0000,0000,,have to be aware that there is bias in\Nlife, and that we are constantly being Dialogue: 0,0:13:03.17,0:13:09.15,Default,,0000,0000,0000,,reminded, primed, of these biases. And we\Nhave to de-bias by showing positive Dialogue: 0,0:13:09.15,0:13:12.92,Default,,0000,0000,0000,,examples. And we shouldn't only show\Npositive examples, but we should take Dialogue: 0,0:13:12.92,0:13:19.40,Default,,0000,0000,0000,,proactive steps, not only at the cultural\Nlevel, but also at the structural level, Dialogue: 0,0:13:19.40,0:13:25.55,Default,,0000,0000,0000,,to change these things. How can we do this\Nfor a machine? First of all, in order to Dialogue: 0,0:13:25.55,0:13:32.60,Default,,0000,0000,0000,,be aware of bias, we need algorithmic\Ntransparency. In order to de-bias, and Dialogue: 0,0:13:32.60,0:13:37.13,Default,,0000,0000,0000,,really understand what kind of biases we\Nhave in the algorithms, we need to be able Dialogue: 0,0:13:37.13,0:13:44.49,Default,,0000,0000,0000,,to quantify bias in these models. How can\Nwe measure bias, though? Because we're not Dialogue: 0,0:13:44.49,0:13:48.05,Default,,0000,0000,0000,,talking about simple machine learning\Nalgorithm bias, we're talking about the Dialogue: 0,0:13:48.05,0:13:56.64,Default,,0000,0000,0000,,societal bias that is coming as the\Noutput, which is deeply embedded. In 1998, Dialogue: 0,0:13:56.64,0:14:02.92,Default,,0000,0000,0000,,societal psychologists came up with the\NImplicit Association Test. Basically, this Dialogue: 0,0:14:02.92,0:14:10.53,Default,,0000,0000,0000,,test can reveal biases that we might not\Nbe even aware of in our life. And these Dialogue: 0,0:14:10.53,0:14:15.22,Default,,0000,0000,0000,,things are associating certain societal\Ngroups with certain types of stereotypes. Dialogue: 0,0:14:15.22,0:14:20.89,Default,,0000,0000,0000,,The way you take this test is, it's very\Nsimple, it takes a few minutes. You just Dialogue: 0,0:14:20.89,0:14:26.54,Default,,0000,0000,0000,,click the left or right button, and in the\Nleft button, when you're clicking the left Dialogue: 0,0:14:26.54,0:14:31.74,Default,,0000,0000,0000,,button, for example, you need to associate\Nwhite people terms with bad terms, and Dialogue: 0,0:14:31.74,0:14:36.86,Default,,0000,0000,0000,,then for the right button, you associate\Nblack people terms with unpleasant, bad Dialogue: 0,0:14:36.86,0:14:42.51,Default,,0000,0000,0000,,terms. And there you do the opposite. You\Nassociate bad with black, and white with Dialogue: 0,0:14:42.51,0:14:47.27,Default,,0000,0000,0000,,good. Then, they look at the latency, and\Nby the latency paradigm, they can see how Dialogue: 0,0:14:47.27,0:14:52.62,Default,,0000,0000,0000,,fast you associate certain concepts\Ntogether. Do you associate white people Dialogue: 0,0:14:52.62,0:15:00.06,Default,,0000,0000,0000,,with being good or bad. You can also take\Nthis test online. It has been taken by Dialogue: 0,0:15:00.06,0:15:06.30,Default,,0000,0000,0000,,millions of people worldwide. And there's\Nalso the German version. Towards the end Dialogue: 0,0:15:06.30,0:15:11.06,Default,,0000,0000,0000,,of my slides, I will show you my\NGerman examples from German models. Dialogue: 0,0:15:11.06,0:15:16.22,Default,,0000,0000,0000,,Basically, what we did was, we took the\NImplicit Association Test and adapted it Dialogue: 0,0:15:16.22,0:15:24.75,Default,,0000,0000,0000,,to machines. Since it's looking at things\N- word associations between words Dialogue: 0,0:15:24.75,0:15:29.68,Default,,0000,0000,0000,,representing certain groups of people and\Nwords representing certain stereotypes, we Dialogue: 0,0:15:29.68,0:15:35.30,Default,,0000,0000,0000,,can just apply this in the semantic models\Nby looking at cosine similarities, instead Dialogue: 0,0:15:35.30,0:15:41.60,Default,,0000,0000,0000,,of the latency paradigm in humans. We came\Nup with the Word Embedding Association Dialogue: 0,0:15:41.60,0:15:48.51,Default,,0000,0000,0000,,Test to calculate the implicit association\Nbetween categories and evaluative words. Dialogue: 0,0:15:48.51,0:15:54.14,Default,,0000,0000,0000,,For this, our result is represented with\Neffect size. So when I'm talking about Dialogue: 0,0:15:54.14,0:16:01.27,Default,,0000,0000,0000,,effect size of bias, it will be the amount\Nof bias we are able to uncover from the Dialogue: 0,0:16:01.27,0:16:07.03,Default,,0000,0000,0000,,model. And the minimum can be -2, and the\Nmaximum can be 2. And 0 means that it's Dialogue: 0,0:16:07.03,0:16:13.23,Default,,0000,0000,0000,,neutral, that there is no bias. 2 is like\Na lot of, huge bias. And -2 would be the Dialogue: 0,0:16:13.23,0:16:17.50,Default,,0000,0000,0000,,opposite of bias. So it's bias in the\Nopposite direction of what we're looking Dialogue: 0,0:16:17.50,0:16:22.94,Default,,0000,0000,0000,,at. I won't go into the details of the\Nmath, because you can see the paper on my Dialogue: 0,0:16:22.94,0:16:31.51,Default,,0000,0000,0000,,web page and work with the details or the\Ncode that we have. But then, we also Dialogue: 0,0:16:31.51,0:16:35.40,Default,,0000,0000,0000,,calculate statistical significance to see\Nif the results we're seeing in the null Dialogue: 0,0:16:35.40,0:16:40.97,Default,,0000,0000,0000,,hypothesis is significant, or is it just a\Nrandom effect size that we're receiving. Dialogue: 0,0:16:40.97,0:16:45.25,Default,,0000,0000,0000,,By this, we create the null distribution\Nand find the percentile of the effect Dialogue: 0,0:16:45.25,0:16:50.67,Default,,0000,0000,0000,,sizes, exact values that we're getting.\NAnd we also have the Word Embedding Dialogue: 0,0:16:50.67,0:16:56.05,Default,,0000,0000,0000,,Factual Association Test. This is to\Nrecover facts about the world from word Dialogue: 0,0:16:56.05,0:16:59.85,Default,,0000,0000,0000,,embeddings. It's not exactly about bias,\Nbut it's about associating words with Dialogue: 0,0:16:59.85,0:17:08.46,Default,,0000,0000,0000,,certain concepts. And again, you can check\Nthe details in our paper for this. And Dialogue: 0,0:17:08.46,0:17:12.23,Default,,0000,0000,0000,,I'll start with the first example, which\Nis about recovering the facts about the Dialogue: 0,0:17:12.23,0:17:19.46,Default,,0000,0000,0000,,world. And here, what we did was, we went\Nto the 1990 census data, the web page, and Dialogue: 0,0:17:19.46,0:17:27.13,Default,,0000,0000,0000,,then we were able to calculate the number\Nof people - the number of names with a Dialogue: 0,0:17:27.13,0:17:32.28,Default,,0000,0000,0000,,certain percentage of women and men. So\Nbasically, they're androgynous names. And Dialogue: 0,0:17:32.28,0:17:40.30,Default,,0000,0000,0000,,then, we took 50 names, and some of them\Nhad 0% women, and some names were almost Dialogue: 0,0:17:40.30,0:17:47.00,Default,,0000,0000,0000,,100% women. And after that, we applied our\Nmethod to it. And then, we were able to Dialogue: 0,0:17:47.00,0:17:54.16,Default,,0000,0000,0000,,see how much a name is associated with\Nbeing a woman. And this had 84% Dialogue: 0,0:17:54.16,0:18:02.17,Default,,0000,0000,0000,,correlation with the ground truth of the\N1990 census data. And this is what the Dialogue: 0,0:18:02.17,0:18:08.81,Default,,0000,0000,0000,,names look like. For example, Chris on the\Nupper left side, is almost 100% male, and Dialogue: 0,0:18:08.81,0:18:17.17,Default,,0000,0000,0000,,Carmen in the lower right side is almost\N100% woman. We see that Gene is about 50% Dialogue: 0,0:18:17.17,0:18:22.33,Default,,0000,0000,0000,,man and 50% woman. And then we wanted to\Nsee if we can recover statistics about Dialogue: 0,0:18:22.33,0:18:27.49,Default,,0000,0000,0000,,occupation and women. We went to the\Nbureau of labor statistics' web page which Dialogue: 0,0:18:27.49,0:18:31.92,Default,,0000,0000,0000,,publishes every year the percentage of\Nwomen of certain races in certain Dialogue: 0,0:18:31.92,0:18:39.09,Default,,0000,0000,0000,,occupations. Based on this, we took the\Ntop 50 occupation names and then we wanted Dialogue: 0,0:18:39.09,0:18:45.26,Default,,0000,0000,0000,,to see how much they are associated with\Nbeing women. In this case, we got 90% Dialogue: 0,0:18:45.26,0:18:51.22,Default,,0000,0000,0000,,correlation with the 2015 data. We were\Nable to tell, for example, when we look at Dialogue: 0,0:18:51.22,0:18:56.51,Default,,0000,0000,0000,,the upper left, we see "programmer" there,\Nit's almost 0% women. And when we look at Dialogue: 0,0:18:56.51,0:19:05.02,Default,,0000,0000,0000,,"nurse", which is on the lower right side,\Nit's almost 100% women. This is, again, Dialogue: 0,0:19:05.02,0:19:10.00,Default,,0000,0000,0000,,problematic. We are able to recover\Nstatistics about the world. But these Dialogue: 0,0:19:10.00,0:19:13.39,Default,,0000,0000,0000,,statistics are used in many applications.\NAnd this is the machine translation Dialogue: 0,0:19:13.39,0:19:21.16,Default,,0000,0000,0000,,example that we have. For example, I will\Nstart translating from a genderless Dialogue: 0,0:19:21.16,0:19:25.77,Default,,0000,0000,0000,,language to a gendered language. Turkish\Nis a genderless language, there are no Dialogue: 0,0:19:25.77,0:19:31.83,Default,,0000,0000,0000,,gender pronouns. Everything is an it.\NThere no he or she. I'm trying translate Dialogue: 0,0:19:31.83,0:19:37.68,Default,,0000,0000,0000,,here "o bir avukat": "he or she is a\Nlawyer". And it is translated as "he's a Dialogue: 0,0:19:37.68,0:19:44.62,Default,,0000,0000,0000,,lawyer". When I do this for "nurse", it's\Ntranslated as "she is a nurse". And we see Dialogue: 0,0:19:44.62,0:19:54.65,Default,,0000,0000,0000,,that men keep getting associated with more\Nprestigious or higher ranking jobs. And Dialogue: 0,0:19:54.65,0:19:59.19,Default,,0000,0000,0000,,another example: "He or she is a\Nprofessor": "he is a professor". "He or Dialogue: 0,0:19:59.19,0:20:04.01,Default,,0000,0000,0000,,she is a teacher": "she is a teacher". And\Nthis also reflects the previous Dialogue: 0,0:20:04.01,0:20:09.96,Default,,0000,0000,0000,,correlation I was showing about statistics\Nin occupation. And we go further: German Dialogue: 0,0:20:09.96,0:20:16.45,Default,,0000,0000,0000,,is more gendered than English. Again, we\Ntry with "doctor": it's translated as Dialogue: 0,0:20:16.45,0:20:21.68,Default,,0000,0000,0000,,"he", and the nurse is translated as\N"she". Then I tried with a Slavic Dialogue: 0,0:20:21.68,0:20:26.48,Default,,0000,0000,0000,,language, which is even more gendered than\NGerman, and we see that "doctor" is again Dialogue: 0,0:20:26.48,0:20:35.78,Default,,0000,0000,0000,,a male, and then the nurse is again a\Nfemale. And after these, we wanted to see Dialogue: 0,0:20:35.78,0:20:41.15,Default,,0000,0000,0000,,what kind of biases can we recover, other\Nthan the factual statistics from the Dialogue: 0,0:20:41.15,0:20:48.07,Default,,0000,0000,0000,,models. And we wanted to start with\Nuniversally accepted stereotypes. By Dialogue: 0,0:20:48.07,0:20:54.03,Default,,0000,0000,0000,,universally accepted stereotypes, what I\Nmean is these are so common that they are Dialogue: 0,0:20:54.03,0:21:00.74,Default,,0000,0000,0000,,not considered as prejudice, they are just\Nconsidered as normal or neutral. These are Dialogue: 0,0:21:00.74,0:21:05.40,Default,,0000,0000,0000,,things such as flowers being considered\Npleasant, and insects being considered Dialogue: 0,0:21:05.40,0:21:10.13,Default,,0000,0000,0000,,unpleasant. Or musical instruments being\Nconsidered pleasant and weapons being Dialogue: 0,0:21:10.13,0:21:16.08,Default,,0000,0000,0000,,considered unpleasant. In this case, for\Nexample with flowers being pleasant, when Dialogue: 0,0:21:16.08,0:21:20.74,Default,,0000,0000,0000,,we performed the Word Embedding\NAssociation Test on the word2vec model or Dialogue: 0,0:21:20.74,0:21:27.07,Default,,0000,0000,0000,,glow model, with a very high significance,\Nand very high effect size, we can see that Dialogue: 0,0:21:27.07,0:21:34.17,Default,,0000,0000,0000,,this association exists. And here we see\Nthat the effect size is, for example, 1.35 Dialogue: 0,0:21:34.17,0:21:40.40,Default,,0000,0000,0000,,for flowers. According to "Cohen's d",\Nto calculate effect size, if effect size Dialogue: 0,0:21:40.40,0:21:46.20,Default,,0000,0000,0000,,is above 0.8, that's considered a large\Neffect size. In our case, where the Dialogue: 0,0:21:46.20,0:21:50.90,Default,,0000,0000,0000,,maximum is 2, we are getting very large\Nand significant effects in recovering Dialogue: 0,0:21:50.90,0:21:57.86,Default,,0000,0000,0000,,these biases. For musical instruments,\Nagain we see that very significant result Dialogue: 0,0:21:57.86,0:22:05.56,Default,,0000,0000,0000,,with a high effect size. In the next\Nexample, we will look at race and gender Dialogue: 0,0:22:05.56,0:22:10.06,Default,,0000,0000,0000,,stereotypes. But in the meanwhile, I would\Nlike to mention that for these baseline Dialogue: 0,0:22:10.06,0:22:16.73,Default,,0000,0000,0000,,experiments, we used the work that has\Nbeen used in societal psychology studies Dialogue: 0,0:22:16.73,0:22:24.98,Default,,0000,0000,0000,,before. We have a grounds to come up with\Ncategories and so forth. And we were able Dialogue: 0,0:22:24.98,0:22:31.97,Default,,0000,0000,0000,,to replicate all the implicit associations\Ntests that were out there. We tried this Dialogue: 0,0:22:31.97,0:22:37.59,Default,,0000,0000,0000,,for white people and black people and then\Nwhite people were being associated with Dialogue: 0,0:22:37.59,0:22:43.21,Default,,0000,0000,0000,,being pleasant, with a very high effect\Nsize, and again significantly. And then Dialogue: 0,0:22:43.21,0:22:49.21,Default,,0000,0000,0000,,males associated with carreer and females\Nare associated with family. Males are Dialogue: 0,0:22:49.21,0:22:56.13,Default,,0000,0000,0000,,associated with science and females are\Nassociated with arts. And we also wanted Dialogue: 0,0:22:56.13,0:23:02.33,Default,,0000,0000,0000,,to see stigma for older people or people\Nwith disease, and we saw that young people Dialogue: 0,0:23:02.33,0:23:07.96,Default,,0000,0000,0000,,are considered pleasant, whereas older\Npeople are considered unpleasant. And we Dialogue: 0,0:23:07.96,0:23:13.30,Default,,0000,0000,0000,,wanted to see the difference between\Nphysical disease vs. mental disease. If Dialogue: 0,0:23:13.30,0:23:17.92,Default,,0000,0000,0000,,there is bias towards that, we can think\Nabout how dangerous this would be for Dialogue: 0,0:23:17.92,0:23:22.67,Default,,0000,0000,0000,,example for doctors and their patients.\NFor physical disease, it's considered Dialogue: 0,0:23:22.67,0:23:30.86,Default,,0000,0000,0000,,controllable whereas mental disease is\Nconsidered uncontrollable. We also wanted Dialogue: 0,0:23:30.86,0:23:40.29,Default,,0000,0000,0000,,to see if there is any sexual stigma or\Ntransphobia in these models. When we Dialogue: 0,0:23:40.29,0:23:44.95,Default,,0000,0000,0000,,performed the implicit association test to\Nsee how the view for heterosexual vs. Dialogue: 0,0:23:44.95,0:23:49.13,Default,,0000,0000,0000,,homosexual people, we were able to see\Nthat heterosexual people are considered Dialogue: 0,0:23:49.13,0:23:54.98,Default,,0000,0000,0000,,pleasant. And for transphobia, we saw that\Nstraight people are considered pleasant, Dialogue: 0,0:23:54.98,0:24:00.17,Default,,0000,0000,0000,,whereas transgender people were considered\Nunpleasant, significantly with a high Dialogue: 0,0:24:00.17,0:24:07.76,Default,,0000,0000,0000,,effect size. I took another German model\Nwhich was generated by 820 billion Dialogue: 0,0:24:07.76,0:24:16.04,Default,,0000,0000,0000,,sentences for a natural language\Nprocessing competition. I wanted to see if Dialogue: 0,0:24:16.04,0:24:20.72,Default,,0000,0000,0000,,they have similar biases\Nembedded in these models.\N Dialogue: 0,0:24:20.72,0:24:25.81,Default,,0000,0000,0000,,So I looked at the basic ones\Nthat had German sets of words Dialogue: 0,0:24:25.81,0:24:29.87,Default,,0000,0000,0000,,that were readily available. Again, for\Nmale and female, we clearly see that Dialogue: 0,0:24:29.87,0:24:34.76,Default,,0000,0000,0000,,males are associated with career,\Nand they're also associated with Dialogue: 0,0:24:34.76,0:24:40.81,Default,,0000,0000,0000,,science. The German implicit association\Ntest also had a few different tests, for Dialogue: 0,0:24:40.81,0:24:47.74,Default,,0000,0000,0000,,example about nationalism and so on. There\Nwas the one about stereotypes against Dialogue: 0,0:24:47.74,0:24:52.67,Default,,0000,0000,0000,,Turkish people that live in Germany. And\Nwhen I performed this test, I was very Dialogue: 0,0:24:52.67,0:24:57.50,Default,,0000,0000,0000,,surprised to find that, yes, with a high\Neffect size, Turkish people are considered Dialogue: 0,0:24:57.50,0:25:02.07,Default,,0000,0000,0000,,unpleasant, by looking at this German\Nmodel, and German people are considered Dialogue: 0,0:25:02.07,0:25:07.82,Default,,0000,0000,0000,,pleasant. And as I said, these are on the\Nweb page of the IAT. You can also go and Dialogue: 0,0:25:07.82,0:25:11.76,Default,,0000,0000,0000,,perform these tests to see what your\Nresults would be. When I performed these, Dialogue: 0,0:25:11.76,0:25:18.97,Default,,0000,0000,0000,,I'm amazed by how horrible results I'm\Ngetting. So, just give it a try. Dialogue: 0,0:25:18.97,0:25:23.76,Default,,0000,0000,0000,,I have a few discussion points before I end my\Ntalk. These might bring you some new Dialogue: 0,0:25:23.76,0:25:30.74,Default,,0000,0000,0000,,ideas. For example, what kind of machine\Nlearning expertise is required for Dialogue: 0,0:25:30.74,0:25:37.17,Default,,0000,0000,0000,,algorithmic transparency? And how can we\Nmitigate bias while preserving utility? Dialogue: 0,0:25:37.17,0:25:41.72,Default,,0000,0000,0000,,For example, some people suggest that you\Ncan find the dimension of bias in the Dialogue: 0,0:25:41.72,0:25:47.82,Default,,0000,0000,0000,,numerical vector, and just remove it and\Nthen use the model like that. But then, Dialogue: 0,0:25:47.82,0:25:51.58,Default,,0000,0000,0000,,would you be able to preserve utility, or\Nstill be able to recover statistical facts Dialogue: 0,0:25:51.58,0:25:55.88,Default,,0000,0000,0000,,about the world? And another thing is; how\Nlong does bias persist in models? Dialogue: 0,0:25:55.88,0:26:04.04,Default,,0000,0000,0000,,For example, there was this IAT about eastern\Nand western Germany, and I wasn't able to Dialogue: 0,0:26:04.04,0:26:12.48,Default,,0000,0000,0000,,see the stereotype for eastern Germany\Nafter performing this IAT. Is it because Dialogue: 0,0:26:12.48,0:26:17.19,Default,,0000,0000,0000,,this stereotype is maybe too old now, and\Nit's not reflected in the language Dialogue: 0,0:26:17.19,0:26:22.17,Default,,0000,0000,0000,,anymore? So it's a good question to know\Nhow long bias lasts and how long it will Dialogue: 0,0:26:22.17,0:26:27.98,Default,,0000,0000,0000,,take us to get rid of it. And also, since\Nwe know there is stereotype effect when we Dialogue: 0,0:26:27.98,0:26:33.21,Default,,0000,0000,0000,,have biased models, does that mean it's\Ngoing to cause a snowball effect? Because Dialogue: 0,0:26:33.21,0:26:39.22,Default,,0000,0000,0000,,people would be exposed to bias, then the\Nmodels would be trained with more bias, Dialogue: 0,0:26:39.22,0:26:45.28,Default,,0000,0000,0000,,and people will be affected more from this\Nbias. That can lead to a snowball. And Dialogue: 0,0:26:45.28,0:26:50.32,Default,,0000,0000,0000,,what kind of policy do we need to stop\Ndiscrimination. For example, we saw the Dialogue: 0,0:26:50.32,0:26:55.73,Default,,0000,0000,0000,,predictive policing example which is very\Nscary, and we know that machine learning Dialogue: 0,0:26:55.73,0:26:59.72,Default,,0000,0000,0000,,services are being used by billions of\Npeople everyday. For example, Google, Dialogue: 0,0:26:59.72,0:27:05.07,Default,,0000,0000,0000,,Amazon and Microsoft. I would like to\Nthank you, and I'm open to your Dialogue: 0,0:27:05.07,0:27:10.14,Default,,0000,0000,0000,,interesting questions now! If you want to\Nread the full paper, it's on my web page, Dialogue: 0,0:27:10.14,0:27:15.88,Default,,0000,0000,0000,,and we have our research code on Github.\NThe code for this paper is not on Github Dialogue: 0,0:27:15.88,0:27:20.55,Default,,0000,0000,0000,,yet, I'm waiting to hear back from the\Njournal. And after that, we will just Dialogue: 0,0:27:20.55,0:27:26.25,Default,,0000,0000,0000,,publish it. And you can always check our\Nblog for new findings and for the shorter Dialogue: 0,0:27:26.25,0:27:31.20,Default,,0000,0000,0000,,version of the paper with a summary of it.\NThank you very much! Dialogue: 0,0:27:31.20,0:27:40.19,Default,,0000,0000,0000,,{\i1}applause{\i0} Dialogue: 0,0:27:40.19,0:27:45.20,Default,,0000,0000,0000,,Herald: Thank you Aylin! So, we come to\Nthe questions and answers. We have 6 Dialogue: 0,0:27:45.20,0:27:51.58,Default,,0000,0000,0000,,microphones that we can use now, it's this\None, this one, number 5 over there, 6, 4, 2. Dialogue: 0,0:27:51.58,0:27:57.15,Default,,0000,0000,0000,,I will start here and we will\Ngo around until you come. OK? Dialogue: 0,0:27:57.15,0:28:01.69,Default,,0000,0000,0000,,We have 5 minutes,\Nso: number 1, please! Dialogue: 0,0:28:05.22,0:28:14.85,Default,,0000,0000,0000,,Q: I might very naively ask, why does it\Nmatter that there is a bias between genders? Dialogue: 0,0:28:14.85,0:28:22.05,Default,,0000,0000,0000,,Aylin: First of all, being able to uncover\Nthis is a contribution, because we can see Dialogue: 0,0:28:22.05,0:28:28.25,Default,,0000,0000,0000,,what kind of biases, maybe, we have in\Nsociety. Then the other thing is, maybe we Dialogue: 0,0:28:28.25,0:28:34.98,Default,,0000,0000,0000,,can hypothesize that the way we learn\Nlanguage is introducing bias to people. Dialogue: 0,0:28:34.98,0:28:41.81,Default,,0000,0000,0000,,Maybe it's all intermingled. And the other\Nthing is, at least for me, I don't want to Dialogue: 0,0:28:41.81,0:28:45.30,Default,,0000,0000,0000,,live in a world biased society, and\Nespecially for gender, that was the Dialogue: 0,0:28:45.30,0:28:50.38,Default,,0000,0000,0000,,question you asked, it's\Nleading to unfairness. Dialogue: 0,0:28:50.38,0:28:52.11,Default,,0000,0000,0000,,{\i1}applause{\i0} Dialogue: 0,0:28:58.38,0:28:59.90,Default,,0000,0000,0000,,H: Yes, number 3: Dialogue: 0,0:28:59.90,0:29:08.24,Default,,0000,0000,0000,,Q: Thank you for the talk, very nice! I\Nthink it's very dangerous because it's a Dialogue: 0,0:29:08.24,0:29:15.56,Default,,0000,0000,0000,,victory of mediocrity. Just the\Nstatistical mean the guideline of our Dialogue: 0,0:29:15.56,0:29:21.23,Default,,0000,0000,0000,,goals in society, and all this stuff. So\Nwhat about all these different cultures? Dialogue: 0,0:29:21.23,0:29:26.15,Default,,0000,0000,0000,,Like even in normal society you have\Ndifferent cultures. Like here the culture Dialogue: 0,0:29:26.15,0:29:31.97,Default,,0000,0000,0000,,of the Chaos people has a different\Nlanguage and different biases than other Dialogue: 0,0:29:31.97,0:29:36.55,Default,,0000,0000,0000,,cultures. How can we preserve these\Nsubcultures, these small groups of Dialogue: 0,0:29:36.55,0:29:41.29,Default,,0000,0000,0000,,language, I don't know,\Nentities. You have any idea? Dialogue: 0,0:29:41.29,0:29:47.15,Default,,0000,0000,0000,,Aylin: This is a very good question. It's\Nsimilar to different cultures can have Dialogue: 0,0:29:47.15,0:29:54.22,Default,,0000,0000,0000,,different ethical perspectives or\Ndifferent types of bias. In the beginning, Dialogue: 0,0:29:54.22,0:29:58.88,Default,,0000,0000,0000,,I showed a slide that we need to de-bias\Nwith positive examples. And we need to Dialogue: 0,0:29:58.88,0:30:04.50,Default,,0000,0000,0000,,change things at the structural level. I\Nthink people at CCC might be one of the, Dialogue: 0,0:30:04.50,0:30:11.88,Default,,0000,0000,0000,,like, most groups that have the best skill\Nto help change these things at the Dialogue: 0,0:30:11.88,0:30:16.13,Default,,0000,0000,0000,,structural level, especially for machines.\NI think we need to be aware of this and Dialogue: 0,0:30:16.13,0:30:21.12,Default,,0000,0000,0000,,always have a human in the loop that cares\Nfor this. instead of expecting machines to Dialogue: 0,0:30:21.12,0:30:25.96,Default,,0000,0000,0000,,automatically do the correct thing. So we\Nalways need an ethical human, whatever the Dialogue: 0,0:30:25.96,0:30:31.00,Default,,0000,0000,0000,,purpose of the algorithm is, try to\Npreserve it for whatever group they are Dialogue: 0,0:30:31.00,0:30:34.44,Default,,0000,0000,0000,,trying to achieve something with. Dialogue: 0,0:30:36.36,0:30:37.36,Default,,0000,0000,0000,,{\i1}applause{\i0} Dialogue: 0,0:30:38.91,0:30:40.75,Default,,0000,0000,0000,,H: Number 4, number 4 please: Dialogue: 0,0:30:41.13,0:30:47.21,Default,,0000,0000,0000,,Q: Hi, thank you! This was really\Ninteresting! Super awesome! Dialogue: 0,0:30:47.21,0:30:48.17,Default,,0000,0000,0000,,Aylin: Thanks! Dialogue: 0,0:30:48.17,0:30:53.72,Default,,0000,0000,0000,,Q: Early, earlier in your talk, you\Ndescribed a process of converting words Dialogue: 0,0:30:53.72,0:31:00.77,Default,,0000,0000,0000,,into sort of numerical\Nrepresentations of semantic meaning. Dialogue: 0,0:31:00.77,0:31:02.14,Default,,0000,0000,0000,,H: Question? Dialogue: 0,0:31:02.14,0:31:08.35,Default,,0000,0000,0000,,Q: If I were trying to do that like with a\Npen and paper, with a body of language, Dialogue: 0,0:31:08.35,0:31:13.73,Default,,0000,0000,0000,,what would I be looking for in relation to\Nthose words to try and create those Dialogue: 0,0:31:13.73,0:31:17.91,Default,,0000,0000,0000,,vectors, because I don't really\Nunderstand that part of the process. Dialogue: 0,0:31:17.91,0:31:21.06,Default,,0000,0000,0000,,Aylin: Yeah, that's a good question. I\Ndidn't go into the details of the Dialogue: 0,0:31:21.06,0:31:25.28,Default,,0000,0000,0000,,algorithm of the neural network or the\Nregression models. There are a few Dialogue: 0,0:31:25.28,0:31:31.29,Default,,0000,0000,0000,,algorithms, and in this case, they look at\Ncontext windows, and words that are around Dialogue: 0,0:31:31.29,0:31:35.58,Default,,0000,0000,0000,,a window, these can be skip grams or\Ncontinuous back referrals, so there are Dialogue: 0,0:31:35.58,0:31:41.31,Default,,0000,0000,0000,,different approaches, but basically, it's\Nthe window that this word appears in, and Dialogue: 0,0:31:41.31,0:31:48.43,Default,,0000,0000,0000,,what is it most frequently associated\Nwith. After that, once you feed this Dialogue: 0,0:31:48.43,0:31:51.79,Default,,0000,0000,0000,,information into the algorithm,\Nit outputs the numerical vectors. Dialogue: 0,0:31:51.79,0:31:53.80,Default,,0000,0000,0000,,Q: Thank you! Dialogue: 0,0:31:53.80,0:31:55.81,Default,,0000,0000,0000,,H. Number 2! Dialogue: 0,0:31:55.81,0:32:05.07,Default,,0000,0000,0000,,Q: Thank you for the nice intellectual\Ntalk. My mother tongue is genderless, too. Dialogue: 0,0:32:05.07,0:32:13.58,Default,,0000,0000,0000,,So I do not understand half of that biasing\Nthing around here in Europe. What I wanted Dialogue: 0,0:32:13.58,0:32:24.61,Default,,0000,0000,0000,,to ask is: when we have the coefficient\N0.5, and that's the ideal thing, what you Dialogue: 0,0:32:24.61,0:32:32.68,Default,,0000,0000,0000,,think, should there be an institution in\Nevery society trying to change the meaning Dialogue: 0,0:32:32.68,0:32:39.71,Default,,0000,0000,0000,,of the words, so that they statistically\Napproach to 0.5? Thank you! Dialogue: 0,0:32:39.71,0:32:44.05,Default,,0000,0000,0000,,Aylin: Thank you very much, this is a\Nvery, very good question! I'm currently Dialogue: 0,0:32:44.05,0:32:48.97,Default,,0000,0000,0000,,working on these questions. Many\Nphilosophers or feminist philosophers Dialogue: 0,0:32:48.97,0:32:56.27,Default,,0000,0000,0000,,suggest that language are dominated by males,\Nand they were just produced that way, so Dialogue: 0,0:32:56.27,0:33:01.72,Default,,0000,0000,0000,,that women are not able to express\Nthemselves as well as men. But other Dialogue: 0,0:33:01.72,0:33:06.25,Default,,0000,0000,0000,,theories also say that, for example, women\Nwere the ones that who drove the evolution Dialogue: 0,0:33:06.25,0:33:11.21,Default,,0000,0000,0000,,of language. So it's not very clear what\Nis going on here. But when we look at Dialogue: 0,0:33:11.21,0:33:16.18,Default,,0000,0000,0000,,languages and different models, what I'm\Ntrying to see is their association with Dialogue: 0,0:33:16.18,0:33:21.29,Default,,0000,0000,0000,,gender. I'm seeing that the most frequent,\Nfor example, 200.000 words in a language Dialogue: 0,0:33:21.29,0:33:27.53,Default,,0000,0000,0000,,are associated, very closely associated\Nwith males. I'm not sure what exactly they Dialogue: 0,0:33:27.53,0:33:32.96,Default,,0000,0000,0000,,way to solve this is, I think it would\Nrequire decades. It's basically the change Dialogue: 0,0:33:32.96,0:33:37.67,Default,,0000,0000,0000,,of frequency or the change of statistics\Nin language. Because, even when children Dialogue: 0,0:33:37.67,0:33:42.72,Default,,0000,0000,0000,,are learning language, at first they see\Nthings, they form the semantics, and after Dialogue: 0,0:33:42.72,0:33:48.25,Default,,0000,0000,0000,,that they see the frequency of that word,\Nmatch it with the semantics, form clusters, Dialogue: 0,0:33:48.25,0:33:53.11,Default,,0000,0000,0000,,link them together to form sentences or\Ngrammar. So even children look at the Dialogue: 0,0:33:53.11,0:33:57.06,Default,,0000,0000,0000,,frequency to form this in their brains.\NIt's close to the neural network algorithm Dialogue: 0,0:33:57.06,0:33:59.74,Default,,0000,0000,0000,,that we have. If the frequency they see Dialogue: 0,0:33:59.74,0:34:05.64,Default,,0000,0000,0000,,for a man and woman are biased, I don't\Nthink this can change very easily, so we Dialogue: 0,0:34:05.64,0:34:11.26,Default,,0000,0000,0000,,need cultural and structural changes. And\Nwe don't have the answers to these yet. Dialogue: 0,0:34:11.26,0:34:13.44,Default,,0000,0000,0000,,These are very good research questions. Dialogue: 0,0:34:13.44,0:34:19.25,Default,,0000,0000,0000,,H: Thank you! I'm afraid we have no more\Ntime left for more answers, but maybe you Dialogue: 0,0:34:19.25,0:34:21.61,Default,,0000,0000,0000,,can ask your questions in person. Dialogue: 0,0:34:21.61,0:34:23.84,Default,,0000,0000,0000,,Aylin: Thank you very much, I would\Nbe happy to take questions offline. Dialogue: 0,0:34:23.84,0:34:24.84,Default,,0000,0000,0000,,{\i1}applause{\i0} Dialogue: 0,0:34:24.84,0:34:25.84,Default,,0000,0000,0000,,Thank you! Dialogue: 0,0:34:25.84,0:34:28.59,Default,,0000,0000,0000,,{\i1}applause continues{\i0} Dialogue: 0,0:34:31.76,0:34:35.79,Default,,0000,0000,0000,,{\i1}postroll music{\i0} Dialogue: 0,0:34:35.79,0:34:56.00,Default,,0000,0000,0000,,subtitles created by c3subtitles.de\Nin the year 2017. Join, and help us!