36c3 preroll music Herald: ...but now we start what we're here for and I'm really happy to be allowed to introduce Anna Mazgal. She will talk about something which a great title. I love it. "Confessions of a Future Terrorist". Terror, terrorism is the one thing you can always shout out and you get everything through. And she will give us a rough guide to overregulate free speech with anti-terrorist measures. Anna works for Wikimedia where she's a lobbyist of human rights into the digital environment and works in Brussels. And she gives a lot of talk. And I think it's the first time at a congress for her, is that right? Anna: Second time. Herald: Second time? I haven't really researched right, because I searched for it. So I have to do this again. It's her 2nd time for congress and I'm really happy to have her here. Please welcome her with a big round of applause. Anna, the stage is yours. Anna: Thank you. Yes. So as you have already heard, I don't do any of the cool things that we comedians and Wikipedians do. I am based in Brussels and the L-word, I do the lobbying on behalf of our community. And today I am here because I wanted to talk to you about one of the proposals of 4 laws that we are now observing the development of. And I wanted to share my concerns also like active, as an activist, because I'm really worried how, if that law passes in its worst possible vision or one of the of the bad versions, how it will affect my work. I'm also concerned how it would it will affect your work. And basically all of ours expression online. And I also want to share with you that this law makes me really angry. So, so I think these are a few good reasons to be here and to talk to you. And I hope after this presentation we can have a conversation about this. And I'm looking forward also to your perspective on it and, and also the things you may not agree with maybe. So, so what is this law? So, in September 2018, the European Commission came out with a proposal of regulation on preventing the dissemination of terrorist content online. So there are a few things to unpack here of for what it is about. First of all, when we see a law that is about Internet and that is about the content and what is about the online environment and it says it will prevent something, this always brings a very difficult and complicated perspective for us, the digital rights activists in Brussels. Because prevention online never means anything good. So, so this is one thing. The other thing is that this very troubled concept of terrorist content, I will be talking about this more, we will talk, I will show you how the European Commission understands it and what are the problems with that understanding and whether this is something that can actually be really defined in the law. So, so these are the, the already the red flags that I have seen and we have seen when we were, when we first got them, the proposal, into our hands. I would like to tell you a little bit about the framework of it. This is probably the, the most dry part of that, but I think it's important to correctly place it. First of all, this is the European Union legislation. So we're talking about the legislation that will influence 27 member states. Maybe 28, but we know about Brexit, so, so that is a debatable what's going to happen there. And it's important to note that whenever we have European legislation in the EU, this is the, these are the laws that actually are shaping the laws of all those countries. And they come before the national laws. So, so should this, should this be implemented in any of the form, when it's implemented in any of the form, this is what is going to happen. The next important part of information that I want to give you is that this particular regulation is a part of the framework that is called Digital Single Market. So the European Union, one of, one of the objectives when a European Commission created the law and when other bodies of the European Union work on it, is that there is, that, that the laws in the member states of the European Union are actually similar. And the Digital Single Market means that what we want, we want to achieve something on the Internet that in a way is already achieved within the European Union geographically, meaning that we don't want the borders on the Internet between people communicating and also delivering goods and services in the European Union. And you may ask how that connects with the terrorist content and how that connects with what today's topics. To be honest, I am also puzzled because I think, that legislation that talks about how people communicate online and what is considered the speech that we wanted there and we don't want this should not be a part of a framework that is about market. So this is also something that brings a concern. Also, as you've seen at the first slide, this piece of legislation, this proposal is called the regulation. And not to go too much into details about what are the forms of legislation in the EU, the important thing to know here is that the regulation is a law that once it is adopted by the EU, once the parliament votes on it, it starts, it is binding directly in all the member states of the European Union, which means that there is no further discussion on how this should be actually used. Of course, in each country there are different decisions being made by different bodies. But it means for us, the people that work on this and that want to influence the legislative process, that once this law is out of Brussels, there is nothing much to be done about how it's going to be implemented. And this is important because for now, the discussion about this, because for us, the discussion about this, is the one that happens in Brussels. There are a few versions of the law. And very quickly, European Commission proposes the law. European Parliament looks at it, debates it, and then produces its own version of it. So amends it or makes it worse. And then the Council of the EU, which is the gathering of all the member states and the representatives of the government of the member states, also creates their own version. And then, of course, when you have 3 versions, you also need to have a lot of conversations and a lot of negotiation, how to put this together into one. And all of those bodies have their own ideas. Every one of those bodies have their own ideas on how any law should look like. So this process is not only complicated, but also this negotiation that is called the trilogs. It's actually very non-transparent. And there is no or almost none, no official information about how those negotiations go, what are the versions of the document and so on. This is the part that we are now in. And I will talk more about this later on. Today I want to talk to you about the potential consequences of diversion. That is the original one, which is the European Commission's version. And it's because it will be very complicated and confusing I guess, if we look at all of the proposals that are on the table. But also, it's important because European Commission has a lot of influence also informally, both on member states and also on - to an extent - on the whole trial process. So whatever gains we have in other versions or whatever better solutions we have there, they are not secure yet. And I promise I'm almost done with this part. There is other relevant legislation that we'll consider. One is the E-Commerce Directive. And in this, the part that is very relevant is for this particular conversation, is that the platforms, according to this law or Internet services or hosting providers are not by default responsible for the content that users play, place online. So it's a very important premise that also protects us, protects our rights, protects our privacy, that the they are not, they cannot go after us or they cannot look for, for the content that could be potentially illegal, which would mean that they would have to look into everything. But of course, they have to react when somebody notifies them and they have to see whether the information that is placed by the users should stay up or not. There is also a directive on combating terrorism. And this is the piece of legislation that is quite recent to my best knowledge. Not all countries in the European Union, not all member states have implemented it yet. So for us, it was also very puzzling that we actually have a new law, a new proposal that is talking about the communication part of, of what already has been mentioned in this directive. When we still don't know how it works, we still don't know because this law is physically not being used at all. So this was for us also difficult to understand why the commission does not want to wait and see how like what comes out from the directive on combating terrorism. So why would the European Commission and why the European legislators would actually want such a law that, again, is about the content that people post through different services and why this is an important issue. If this is, why this issue is actually conflated with the market questions and the harmonization in the digital market. So there are some serious numbers here. 94 % and 99 %. And I'm wondering if you have any idea what those numbers are about. Audience: Persons. Anna: I'm sorry? Audience: Persons. Anna: Yes. It's about people. But the numbers are actually presenting, so there was a survey done by Eurostat and those numbers present the percentage of people first number 94 % presents the percentage of people that say that they have not come across terrorist content online. Right. So, inversely, only 6 % of people actually say that they had access to terrorist content, it's important to underline that they say it because there's no way to check what that content actually was and of course we can, you know, here is the analogy of what a certain American judge said about pornography. I know it when I see it. It's not a very good definition of anything, really. So I would argue that actually 6 % of people being affected by something is not really a big percentage and that the European Union actually has bigger problems to deal with and where they can spend money and energy on. E.g., we are all affected by, I don't know, air pollution and that's much more people. 89% are the people in the age range between 15 and 24. But again, were not affected by something what they would consider terrorist content. Of course, would somebody think of the children? There you go. The children and young people do not also experience it in an overwhelming, overwhelmingly. So, but this rationale is being used, 6 % and 11 % as one of the reasons why this regulation is important, why this law is important. The other is the exposure to, the other reason is the exposure to imagery of violent crimes via social media. So, of course, we know that, that platforms such as Facebook and YouTube contain all sorts of things that people look. We also know that because of their business models, they sometimes push controversial content or violent content into, into peoples proposal, that the proposals that they give to people to, to watch or to read. So this is, actually the 2nd part is not addressed by this, by this proposal at all. But nevertheless, whenever we talk to the representatives of the commission why this law is there, they start waving, that was my experience that the one of the meetings, the person started waving his phone at me and saying, "Well, you know, there are beheading videos online and I can show you how horrible it is", which I consider to be an emotional blackmail at best, but not really a good regulatory impulse. So I guess maybe the commission people are somehow mysteriously affected by that content more than anything else. I don't mean to joke about those, those videos because of course, it is not something that I would want to watch and, and it is very violent. But I would also argue that the problem is not that the video is there, but that somebody has been beheaded. And this is where we should actually direct our attention and look for the sources of that sort of behavior and not only to try and clean the Internet. The other reason why, why this law should be enacted is a radicalisation. Of course, this is a, this is a problem for certain vulnerable populations and people. And we can read about it a lot. And there are organizations that are dealing with strategies to counteract radicalisation. Again, when we look at the evidence, what is the, what is the relationship between content that is available online and the fact that people get radicalized in different level, in different ways, that we didn't see any research and the commission also did not present any research that would actually point to at least the correlation between the two. So again, asked about, so "How did you come up with this idea since without really actually showing the, the support for your claim that radicalisation is connected to that?" This is a quote from a meeting that happened public and journalists were there. Again, the person from the commission said, "We had to make a guess, so we made the guess that way." There is the guess being, yes, there is some sort of connection between the content and the radicalisation. And then finally, when we read the impact assessment and when we look at the different articles, that or different explanations that the European Commission posts about the rationale for this law, of course, they bring the terrorists attack that have been happening and they make, they swiftly go from naming the different violent events that have happened in Europe very recently or quite recently. And they swiftly make a connection between the fact that somebody took a truck and and run into a group of people. Or that somebody was participating in the shooting or organizing the shooting of people enjoying themselves.They swiftly go from this to the fact that regulation of the content is needed. Which also, the fact that you put something in one sentence does not mean it makes sense. Right? So this is also not very well documented. Again, pressed about this, the representative of the European Commission said that, well, "We know that and it has been proven in the investigation, that one of the people that were responsible for the Bataclan attack actually used the Internet before that happened. laughter Yes. No more comment needed on that one. So, well, clearly, there are "very good reasons", quote unquote, to spend time and citizens money on working on the new law. And I always say that basically these laws are created because, not because there is a reason, but because there is a do-something-doctrine. Right. We have a problem, we have to do something. And this is how this law, I think, came to be. And the do-something-doctrine in this particular case, or also, of course, it encompasses a very broad and blurry definition of that law. I will talk about this more in a moment. It also encompasses measures, we, if we define something that we want to counteract to, we have to basically say what should happen. Right. So that the problem is being solved. And there are 3 measures that they will also explain. One is the removal orders. The other is referrals. And the third are so-called proactive measures. This is, I guess, the part where we touch the prevention most. And then the third issue is that, the one of the things I also want to talk about is the links between the content that is being removed and the actual investigations or prosecutions that may occur, because of course it's possible that there will be some content found that actually does document the crime. And then what do we do about that? So, going forward, I do think that the definition and this law is basically, its main principle is to normalize the state control over how people communicate and what they wanted to say. As it was said before, under the premise of terrorism, we can actually pack a lot of different things because people are afraid of this. And we have also examples from other topics, other laws that have been debated in Brussels. One was public sector information directive, where everybody was very happy discussing how much public information should be released and where it should come from and how people should have access to it. And a part of public information is the information that is produced by companies that perform public services, but they may also be private, for example, sometimes transport, public transport is provided that way. And actually public transport providers were the ones that were saying that they cannot release the information that they have, namely timetables and other information about how the system works that could be useful for citizens because then it may be used by terrorists. I guess that maybe prevents the potential terrorists from going from bus stop to bus stop and figuring out how the buses go. But we already know that this does not work that way. So this is something that actually normalizes this approach. And let's first look at the definition of the proposal as presented by the European Commission. So they say basically, let me read: "Terrorist content means one or more of the following information. So a) inciting or advocating, including by glorifying, the commission of terrorist offences". I do apologise for the horrible level of English that they use, I don't know why and that I don't apologise for them, but for the fact that they expose you to it. "The commission of terrorist offences, Clapping thereby causing a danger that such acts be committed". You won't believe how many times I had to read all this to actually understand what all those things mean. "Encouraging the contribution to terrorist offences". So contribution could be money, could be some, I guess material resources. "Promoting the activities of a terrorist group, in particular by encouraging the participation in or support to a terrorist group. Instructing on methods or techniques for the purpose of committing terrorist offenses". And then there is also the definition of "dissemination of terrorist content". That basically means "making terrorist content available to third parties on the hosting service providers services". As you can probably see, the dissemination and the fact that third parties are evoked mean that this law is super broad. So it's not only about social media because making content available through third parties may mean that I am sharing something over some sort of service with my mom and she is a third party in the understanding of this law. So we were actually super troubled to see that not only does it encompass services that make information available to the public. So the one that we all can see like social media, but also that potentially it could be used against services that make, let people communicate privately. So that is a that is a big issue. The second thing I wanted to direct your attention to is the parts that they put in italics. It's how soft those those concepts are, inciting, advocating, glorifying, encouraging, promoting. This is a law that actually potentially can really influence how we talk and how to communicate what we wanted to talk about, whether we agree or disagree with certain policies or certain political decisions. And all those things are super soft. And it's very, very hard to say what does it really mean. And I want to give you an example of the same content used in 3 different cases to illustrate this. So let's imagine we have a group of people that recorded a video and on those videos, they say that, well, basically they call themselves terrorists, to make it easier, and they say that they wanted to commit all sorts of horrible things in specific places, so that constitutes like some sort of a credible threat. And they also bragged that they killed someone. And they also say that they're super happy about this and so on. And they also, of course, encourage others to join them and so on and so on. And the 3 cases would be: 1 would be that this particular group posted videos on, I don't know, their YouTube channel. The other case would be that there is a media outlet that reports on it and either links to this video or maybe present snippets of it. And the third case would be, for example, that there is some sort of group that is actually following what's happening in that region and collects evidence that can then help identify the people and prosecute them for the crimes they commit. Like the crime that's our exemplary terrorists admitted to committing. And do you think that according to this definition, in your opinion, do you think that there is a difference between those 3 types of presenting that content between the terrorist group that is presenting it on their channel, between the media outlet and between the activists? There is none. Because this law has nothing, does not define in any way that the so-called terrorist content is something that is published with an intention of actually advocating and glorifying. So the problem is that not only does the content that, let's say, is as weak as we may call it manifestly illegal. So somebody kills someone and is being recorded and we know it's a crime and perhaps we don't want to watch it, although I do think that we should also have a discussion in our society, what we wanted to see and what we don't want to see. From the fact, from the perspective that the world is complicated and we may have the right to access all sorts of information, even that is not so pleasant and not so easy to digest. So this law does not make this differentiation. There is no mention of how this should be intentional to qualify to be considered so-called terrorist content. And that's a big problem. So as you can see, there is a fallacy in this narrative because these will be the member states and their so-called competent authorities that will be deciding what the terrorist content is. And, of course, Europeans have a tendency, a tendency to think we have the tendency to think of ourselves as the the societies and the nations and the countries that champion the rule of law and that and that actually respect fundamental rights and expect, respect freedom of speech. But we also know that this is changing rapidly. And I also will show you examples of how that changes in this area that we're talking about right now. So. So I do not have great trust in in European governments into making the correct judgment about that. So. So we have this category of very dubious and and very broad terrorist content. And then, so how it's how it's being done. The, the basically all that power to decide what the content, like how to deal with that content, is actually outsourced to private actors. So the platforms that we are talking about becomes kind of mercenaries because both the commission and I guess many member states say, well, it's not possible that the judge will actually look through content that is placed online and give, you know, proper judiciary decisions about what should, what constitute freedom of expression and what goes beyond it. Because it hurts other people or is basically a proof of something illegal. So the platforms will take those decisions. This will be the hosting service providers, as I mentioned. And then also a lot of the reliance that they will do it right is put into the wishful thinking in this proposal that says, well, basically, you have to put in terms of service that you will not host terrorist content. So then that again, there is a layer in there where the platform, let's say Facebook or Twitter or any anyone else actually decides what and how they wanted to deal with that in detail. Also, one thing I didn't mention is that looking for this regulation and looking at who is the platform that should basically have those terms of service we realize that Wikimedia that actually our platforms will actually be in the scope of that. So not only that may affect the way we can document and reference the articles that are appearing on Wikipedia, on all those all those, on the events that are described or the groups or the political situation and what not. But also that, you know, our community of editors will have less and less to say if we have to put a lot of emphasis on terms of service. I just think that we are somehow a collateral damage of this. But also this doesn't console me much because, of course, Internet is bigger than than our projects. And also, we want to make sure that, that content is not being removed elsewhere. So basically the 3 measures are the removal orders, as I mentioned. And this is something that is fairly, fairly straightforward. And actually, I'm wondering why there has to be a special law to bring it, because, to being because the removal order is basically a decision that the competent authority in the member state releases and sends it to the platform. The problem is that according to the commission, the platform should actually act on it in 1 hour. And then again, we ask them why 1 hour and not 74 minutes? And they say, "Well, because we actually know", I don't know how, but they say they do. Let's take it at face value. "We actually know that the content is the most, you know, viral and spreads the fastest within, has the biggest range within the 1 hour from appearance". And then we ask them "Well, but how can you know that? Actually, the people that find the content find it exactly on the moment when it comes up. Maybe it has been around for 2 weeks and this 1 hour window when it went really viral is like long gone". And here they don't really answer, obviously. So this is one of the measures that I guess makes the most sense out of all of that. Then we have the referrals that we call lazy removal orders. And this is this is really something that is very puzzling for me because the referral is a situation in which this competent authority and the person working there goes through the content, through the videos or postings and looks at it and says, "Well, I think, I think it's against the terms of service of this platform, but does not actually release this removal order, but writes to the platform, let's them know and say, "Hey, can you please check this out?" I'm sorry, I'm confused, is this the time that I have left or the time? OK, good, time is important here. So so basically, you know, they are basically, won't spend the time to prepare this removal order and right and take let the platform to, tell the platform actually to remove it. But they will just ask them to please verify whether this content should be there or not. And first of all, this is the real outsourcing of power over the speech and expression. But also we know how platforms take those decisions. They have a very short time. The people that do it are sitting somewhere, most probably where the content is not originating from. So they don't understand the context. Sometimes they don't understand the language. And also, you know, it's better to get rid of it just in case it really is problematic, right? So this is something that is completely increased this great gray area of information that is controversial enough to be flagged, but it's not illegal enough to be removed by the order. By the way, the European Parliament actually kicked this out from their version. So now the fight is in this negotiation between the 3 institutions to actually follow this recommendation and just remove it, because it really does not make sense. And it really makes the people that release those referrals not really accountable for their decisions because they don't take the decision. They just make a suggestion. And then we have the proactive measures, which most definitely will lead to over-policing of content. There is a whole, a very clever description in the law that basically boils down to the point that if you are going to use content filtering and if you're going to prevent content from disappearing, then basically you are doing a good job as a platform. And this is the way to actually deal with terrorist content. Since, however we define it, again, this is very context-oriented, very context-dependent. It's really very difficult to say based on what sort of criteria and based, based on what sort of databases those automated processes will be, will be happening. So, of course, as it happens in today's world, Somebody privatizes the profits, but the losses are always socialized. And this is now no exception from that rule. So, again, when we were talking to the European Commission and asking them, "Why is this not a piece of legislation that belongs to the enforcement of the law?" And that is then not controlled by a heavily by the judiciary system and by any other sort of oversight, that enforcements usually had. "They have, well, because, you know, when we have those videos of beheadings, they usually don't happen in Europe and they are really beyond our jurisdiction". So, of course, nobody will act on it. On the very meaningful level of actually finding the people that that are killing, that are in the business of killing others and making sure they cannot continue with this activity. So it's very clear that this whole law is about cleaning the Internet and not really about a meaningfully tackling societal problems that lead to that sort of violence. Also the redress, which is the mechanism in which the user can say, hey, this is not the right decision. I actually believe this content is not illegal at all. And it's important for me to say this and this is my right and I want it to be up. Those, those provisions are very weak. You cannot actually protest meaningfully against a removal order of your content. Of course, you can always take states, the states to court. But we know how amazingly interesting that is and how fast it happens. So that, so we can I think we can agree that there is no meaningful way to actually protest. Also, the state may ask, well, actually, this this removal order should... the user should not be informed that the content has been has been taken down because of terrorism. So. Or depicting terrorism or glorifying or whatever. So even you may not even know why the content is taken down. It will be a secret. For referrals and for proactive measures, well, you know what? Go talk to the platform and protest with them. And then, of course, the other question is, so who is the terrorist? Right. Because this is a very important question that that we should have answered if we wanted to... if we wanted to have a law that actually is meaningfully engaging with those issues. And of course, well, the as you know already from what I said, the European Commission in that particular case does not provide a very good answer. But we have some other responses to that. For example, Europol has created a report and then there was a blog post based on that. On the title "On the importance of taking down non-violent terrorist content". So we have the European Commission that says, "Yes, it's about the beheadings and about the mutilations". And we have Europol that says, "You know, actually this non-violent terrorist content is super important". So basically what they say and I quote, "Poetry is a literary medium that is widely appreciated across the Arab world and is an important part of their region's identity. Mastering it provides the poet with singular authority in Arabic culture. The most prominent jihadi leaders - including Osama bin Laden and former Islamic State spokesman Abu Muhammad al-Adnadi - frequently included poetry in their speeches or wrote poems of their own. Their charisma was closely intertwined with their mastery of poetry." So we can see the art that is being made by Europol between a very important aspect of a culture that is beautiful and enriching and about the fact that this, that, that Europol wants it to see it weaponized. The other part of the blogpost was about how ISIS presents interesting activities that their members... their, their... their fighters have. And one of them is that they are enjoying themselves and smiling and spending time together and swimming. So what? How do we, what do we make out of that? So the video of some brown people swimming are now terrorist content? This is... the blatant racism of, of this, of this communication really enrages me. And I think it's really a shame that that nobody called Europol out on this, when the blogpost came up. We also have laws in Europe that are different. I mean, this is not the same legislation, but that actually give the... give the the taste of what may happen. One is the the Spanish law against hate speech. And, and this is an important part. It didn't happen online, but it shows the approach, that basically first you have legislators that say, oh, don't worry about this, we really want to go after bad guys. And then what happens is that there was a puppeteer performance done by 2 people, "The Witch and Don Christóbal" and the puppets were actually... this is the kind of Punch and Judy performance in which, this is a genre of, of theater, theatric performances, I'm sorry. That is kind of full of silly jokes and and sometimes excessive and and unjustified violence and, and the, and the full of bad taste. And this is quite serious. And the, the 2 characters in the, the 2 puppets held the banner that featured a made-up terrorist organization. And after that performance, actually, they were charged with, first of all, promoting terrorism, even though there is no terrorist organization like that. And then also with inciting, inciting hatred. And this is what's one of the puppeteers said after describing this whole horrible experience. Finally, the charges were dropped. So this is good. But I think this really sums up who is the terrorists and how those laws are being used against people who actually have nothing to do with, with violence. We were charged with inciting hatred, which is a felony created in theory to protect vulnerable minorities, the minorities in this case where the church, the police and the legal system. laughter Then, again in Spain, I don't want to single out this beautiful country, but actually, unfortunately, they have good examples. This is a very recent one. So Tsunami Democràtic in Catalonia created an app to actually help people organize small action in a decentralized manner. And they placed the documentations on GitHub. And it was taken down by the order of, of the Spanish court. And also the - and this is the practical application of such laws online - also, the website of Tsunami Democràtic was taken down by the court. Of course, both of that on charges of of facilitating terrorist activities and inciting to terrorism. So why is it important? Because of what comes next. So there will be the Digital Services Act, which will be an overhaul of this idea that I mentioned at the beginning, which is that basically platform are not responsible by default, by what we put online. And European Commission and other, the European Commission and other actors in the EU are toying with the idea that maybe platforms should be somehow responsible. So, of course. And it's not only about social media, but basically anybody that any sort of, of a service that helps people place content online. And then, the, one of the ideas, we don't know what it's going to be, it's not there yet. It's going to happen at the beginning of the next year, so quite soon. But we can actually expect that the so-called "Good Samaritan" rule will be 1 of the solutions proposed. What is this rule? This rule basically means if a platform is really going the extra mile and doing a good job in removing the content, that is what... that is either illegal or again or again, a very difficult category, harmful. I also don't know what that exactly means. Then if they behave well, then they will not be held responsible. So this is basically a proposal that you cannot really turn down, because if you run the business, you want to manage the risk of that and you don't want to be fined. And you and you don't want to pay, pay money. So, of course, you try and overpolice and of course you try and you filter the content and of course you take content when it only raises a question what sort of... what sort of content that is. Is it neutral or is it maybe, you know, making somebody offended or... or stirred? And, of course, other attempts, we heard it from Germany. Which is basically that there wasn't a proposal to actually make... oblige.. like make platforms obliged to give passwords of users of social media. The people that are under investigation or prosecution. And also, of course, we see that one of the ideas that supposedly is going to fix everything is that, well, if terrorists communicate through encrypted services, then maybe we should do something about encryption. And there was a petition already on (?) to actually go in to actually forbid encryption for those services after one of the one of the terrorist attacks. So, of course, it sounds, it sounds very extreme. But this is, in my opinion, the next the next frontier here. So what can we do? Because this is all quite difficult. So as I mentioned, the negotiations are still on. So there is still time to talk to your government. And this is very import because, of course, the governments, when they have this idea... they have this proposal on the table, that they will be able to decide finally who is the terrorist and what is the terrorist content. And also, that's on one hand. On the other hand, they know that people don't really care all that much about what happens in the E.U., which is unfortunately true. They are actually supporting very much the commission's proposals. The only thing that they don't like is the fact that somebody from the police, from other country can maybe interfere with content in their language, because that's one of the provisions that that also is there. So, so this is what they don't like. They want to keep their there the territoriality of their enforcement laws intact. But there is still time and we can still do this. And if you want to talk to me about what are the good ways to do it, I'm available here. And I would love to take that conversation up with you. The other is a very simple measure that I believe is is always working. Is one that basically is about telling just 1 friend, even 1 friend, and asking them to do the same to talk to other people about this. And there are 2 reasons to do it. One is because, of course, then we make people aware of what it happens. And the other in this particular case that is very important is that basically people are scared of terrorism and, and they support a lot of measures just because they hear this word. And when we explain, that, what that really means and when we unpack this a little bit, we build the resilience to those arguments. And I think it's important. The other people who should know about this are activists working with vulnerable groups because of the stigmatization that I already mentioned and because of the fact that we need to document horrible things that are happening to people in other places in the world and also here in Europe. And journalists and media organizations, because they will be affected by this law. And by the way, how they can report and where they can they can get the sources for their information. So I think I went massively over time from what it was planned. I hope we can still have some questions. Thank you. So, yeah. Talk to me more about this now and then after the talk. Thank you. applause Herald: Thanks for your talk. We still have time for questions, so please, if you have a question, line up at the mics. We have 1, 2, 3 evenly distributed through the room. I want to remind you really quickly that a question normally is one sentence and ends with a question mark. laughter Not everybody seems to know that. So we start with mic number 2. Mic2: Hello. I... so I run Tor Relays in the United States. It seems like a lot of these laws are focused on the notion of centralized platforms. Do they define what a platform is and are they going to extradite me because I am facilitating Tor Onion service? A: Should we answer, no? H: Yeah. A: Okay, yes, so they do and they don't in a way that the definition it's based on basically what the the hosting provider is in in the European law is actually very broad. So it doesn't take into account the fact how big you are or how you run your services. The bottom line is that if you allow people to put content up and share it with, again, 3rd party, which may be the whole room here, it may be the whole world but it may be just the people I want to share things to with. Then then you're obliged to to use the measures that are... or, or to comply with the measures that are envisioned in this regulation. And there is a there's a debate also in the parliament. It was taken up and narrowed down actually to the communication to the public. So I guess then as you correctly observe, it is more about about the big platforms or about the centralized services. But actually the, in the commission version, nothing makes me believe that, that only then will be affected. On the contrary, also the, the messaging services may be. H: Okay, um, next question, mic number 3. Mic3: Is it, uh, a photo with the upload filters, the copyright directive, it was really similar debate, especially on small companies, because, um, uh, at that time, the question was they tried to push upload filters for copyright content. And the question was, uh, how does that fit with small companies? And they still haven't provided an answer to that. Uh, the problem is they took the copyright directive and basically inspired themselves from the upload filters and applied it to terrorist content. And it's again, the question, how does that work with small Internet companies that have to have someone on call during the nights and things like that. So even big providers, I heard they don't have the means to, to properly enforce that something like this, this is a killer for the European Internet industry. A: Yes. laughter applause H: I want to give a short reminder on the 1 sentence rule. We have a question from the Internet. Signal angel, please. Signal Angel: Yes, what, the question is, wouldn't decentralized social networks bypass these regulations? A: I'm not a lawyer, but I will give a question, I give an answer to this question that the lawyer would give, I maybe spent too much time with lawyers. That depends. laughter A: Because it really does, because this definition of who is obliged is so broad that a lot depends on the context, a lot depends on what is happening, what is being shared and how. So it's, it's very difficult to say. I just want to say that we also had this conversation about copyright and many people came to me last year at Congress. I wasn't giving a talk about it, but I was at the talk about the copyright directive and the filtering. And many people said, well, actually, you know, if you're not using those services, you will not be affected. And actually, when we share peer to peer, then this is not an issue. But actually, this is changing. And there is actually a decision of the European Court of Justice. And the decisions are not like basically the law, but basically they are very often then followed and incorporated. And this is the... and this is the decision on the Pirate Bay and in... on Pirate Bay. And in this decision, the court says that, well, the argument that Pirate Bay made was basically we're not hosting any content. We're just connecting people with it. And in short, and the court said, well, actually, we don't care. Because you organize it, you optimize it, you like the info, you optimize the information, you bring it to people. And the fact that you don't share it does not really mean anything. And you are liable for the, for the copyright infringements. So, again, this is about a different issue, but this is a very relevant way of thinking that we may expect that it will be translated into other types of content. So, again, the fact that that you don't host anything but you just connect people to one another will not be... may not be something that, that will take you off the hook. H: Microphone number 3. Mic3: Do these proposals contain or... what sort of repercussions do these proposals contained for filing a request, removal requests that are later determined to be illegitimate? Is this just a free pass to censor things? Or can.... are there repercussions? A: You... just want to make sure I understand, you mean the removal orders, the ones that say remove content, and that's it? Mic3: Yeah. If somebody files a removal order that is determined later to be completely illegitimate. Are there repercussions? A: Well, the problem starts even before that because again, the removal orders are being issued by competent authorities. So there's like a designated authority that can do it. Not everybody can. And basically, the order says this is the content. This is the URL. This is the legal basis. Take it down. So there is no way to protest it. And the platform can only not follow this order within 1 hour in 2 situations. One is that the force majeure, that is usually the issue. Basically, there's some sort of external circumstance that prevents them from doing it. I don't know. Complete power outage or problem with their servers that basically they cannot access and remove or block access to this content. The other is if the request... the removal order, I'm sorry, contains errors that actually make it impossible to do. So, for example, there is no URL or it's broken and it doesn't lead anywhere. And these are the only 2 situations. In the rest, the content has to be removed. And there is no way for the user and no way for the platform to actually say, well, hold on, this is not the way to do it. And therefore, after it's being implemented to say, well, that was a bad decision. As I said, you can always go to court with the, with your state, but not many people will do it. And this is not really a meaningful way to address this. H: Next question. Mic number 3. Mic3: How many... how much time do we have to contact the parliamentarians to inform them maybe that there is some big issue with this? What's the worst case timetable at the moment? A: That's a very good question. And thank you for asking because... this ... because I forgot to mention this. That actually is quite urgent. So the commission wanted to, like usually, in those situations, the commission wanted to close the thing until the end of the year and they didn't manage it because there is no, no agreement on those most pressing issues. But we expect that the, the best case scenario is that until March, maybe until June, it will probably happen earlier. It may be the next couple of months. And there will be lots of meetings about about that. So this is more or less the timeline. It's, there's no sort of external deadline for this, right, so this is just an estimation and of course, it may change, but, but this is what we expect. H: We have another question from the Internet. S: Does this law considers that such content is used for psychological warfare by big nations? A: I'm sorry. I... Again, please. S: This, this content is, pictures or video of whatsever, does this law consider that such content is used for psychological warfare? A: Well, I'm trying to see how that relates. I think the law is... does not go into details like that in a way. Which means that I can go back to the definition that basically it's just about the fact that if the content appears to be positive about terrorist activities, then that's the basis of taking it down. But there's nothing else that is being actually said about that. It's not more nuanced than that. So I guess the answer is no. H: One last question from mic number 2. Mic2: Are there, in... any case studies published on successful application of alike laws in other countries? I asked because we have alike laws in Russia for 12 years and it's not that useful as far as I see. A: Not that I know of. So I think it's also a very difficult thing to research because we can only research what, what we know that happened. Right? In a way that you have to have people that actually are vocal about this and that complain about these laws not being enforced in the proper way. So, for example, content that taken down is completely about something else, which also sometimes happens. And, and that's very difficult. I think the biggest question here is whether there is an amount of studies documenting that something does not work that would prevent the European Union from actually having this legislative fever. And I would argue that not, because, as they said, they don't have really good arguments or they don't really have good numbers to justify bringing this law at all. Not to mention bringing the ridiculous measures that they propose. So what we say sometimes in Brussels when we're very frustrated that we we were hoping, you know, being there and advocating for for human rights, is that we... we hoped for... that we can contribute to evidence based policy. But actually, what's happening, it's a policy based evidence. And, and this is the difficult part. So I am all for studies and I am all for presenting information that, you know, may possibly help legislators. There are definitely some MEP or some people there, even probably in the commission. Maybe they just are not allowed to voice their opinion on this because it's a highly political issue that would wish to have those studies or would wish to be able to use them. And that believe in that. But it's just, it doesn't translate into the political process. H: Okay. Time's up. If you have any more questions, you can come up and approach Anna later. A: Yes. H: There is please. Thanks. So first for me. Thanks for the talk. Thanks for patiently answer... 36c3 postroll music Subtitles created by c3subtitles.de in the year 2021. Join, and help us!