rC3 preroll music
Herald: And the the talk that is about to
begin now is by Christoph Schmon from EFF
and Eliska Pirkova from Access Now and
they will be talking about this, probably
maybe the biggest legislative initiative
since the GDPR in Brussels. It's called
the Digital Services Act. And onto you
two. You're muted.
Eliska Perkova: I realized. Hello!
First of all, thank you very much for
having us today. It's a great pleasure to
be the part of this event. I think for
both of us, it's the first time we are
actually joining this year. So it's
fantastic to be the part of this great
community. And today we are going to talk
about the legislative proposal, the EU
proposal which causes a lot of noise all
around Europe, but not only in Europe, but
also beyond. And that's the Digital
Services Act legislative package that
today we already know that this
legislative package actually consists of
two acts: Digital Services Act and a
Digital Market Act. And both of them will
significantly change the regulation of
online platforms with a specific focus on
very large online platforms, also often
referred to as gatekeepers. So those who
actually hold a lot of economic dominance,
but also a lot of influence and control
over users rights and the public
discourse. So I'm going to start with
giving you a quick introduction into
what's the fuss about, what is actually
the DSA about, why we are also interested
in it and why we keep talking about it,
and why this legislation will keep us
preoccupied for the years to come. DSA was
already announced two years ago as a part
of European Union digital strategy, and it
was appointed as one of those actions that
the digital strategy will be actually
consisting of. And it was the promise that
the European Commission gave us already at
that time to create the systemic
regulation of online platforms that
actually places hopefully the users and
their rights into the center of this
upcoming legislation. So the promise
behind DSA is that these ad-based Internet
bill, and I'm speaking now about Ad-Tech
and Online-Targeting. The
Internet, as we actually knew, will be
actually replaced by something that puts
user and users controls, and users right
as a priority. So both of these
legislations implemented and drafted
right, should be actually achieving that
goal in the future. Now, previously,
before DSA actually was drafted, there was
some so-called e-Commerce Directive in
place that actually established the basic
principles, especially in the field of
content governance. I won't go into
details on that because I don't want to
make this too legalistic. But ultimately,
DSA legislation is supposed to not
completely replace but build up on the top
of this legislation that actually created
the ground and the main legal regime for
almost 20 years in Europe to regulate the
user generated content. So, DSA and DMA,
as the legislation will seek to harmonize
the rules addressing the problem, such as
online hate speech, disinformation. But it
also puts emphasis finally on increased
meaningful transparency in online
advertising, the way how the content is
actually being distributed across
platforms and also will develop a specific
enforcement mechanism that will be
actually looking into it. Now, before I
will actually go into the details on DSA
and why DSA matters, and do we actually
need such a big new legislative reform
that is coming from the European
Commission? I want to just unpack it for
you a little bit what this legislative
package actually consists of. So, as I
already mentioned, two regulations... the
regulation, the strongest legal instrument
European Commission actually has in its
hands, which is supposed to achieve the
highest level of harmonization across the
member states. And we all can imagine how
difficult that will be to achieve,
especially in the realm of freedom of
expression and particular categories of
user generated content, which is so deeply
complex dependance. All of those
related to content moderation and content
curation will be mainly in the realm of
Digital Services Act. And then the second
regulation, the Digital Market Act, will
be specifically looking at the dominance
of online platforms, economic dominance,
competitive environment for smaller
players, the fairness in the competition.
And it will also establish the list of
do's and don'ts for gatekeeper's
platforms. So something that... is so the
platforms that actually hold a relevant
dominance and now based on these new
proposals, we know that these platforms
are mainly called as very large online
platforms. So this is exactly how the
legislation refers to gatekeepers now. And
now I think one more point that I want to
make is that the DSA and the DMA were
launched on the 15th of December 2020. So it
was literally a Christmas present given to
the digital rights community by the
European Commission, a long time
anticipated one. The work on DSA
started, however much earlier. Electronic
Frontiers Foundations, as much as Access
Now, together with EDRi, were working very
hard to come up with the priorities and
recommendations what we would like to see
within these legislations to be
enshrined, because from the beginning we
understood the importance and the far
reaching consequences this legislation
will have not only inside of the European
Union, but also beyond. And that brings me
to the final introductory point that I
want to make before I will hand over to
Chris, which is why and do we actually
need DSA? We strongly believe that there
is a big justification and good reason to
actually establish a systemic regulation
of online platforms in order to secure
users fundamental rights, empower them,
and also to protect our democratic
discourse. And this is due to the fact
that for many years we're witnessing this
phase: quite bad regulatory practices in
platform governance that are coming from
member states and Chris will provide for a
very concrete example in that regard, but
also coming from the European Commission
itself, mainly the proposed online service
content regulation, for instance, or we
all remember the story of copyright that
Chris will discuss a little bit further.
We saw how not only the member states, but
also European Commission or European
Union, in order to actually establish some
order in the digital space, they started
pushing the state's obligation and
especially states positive obligation to
protect the users' human rights in the
hands of online private platforms that
started replacing state actors and public
authorities within the online space. They
started assessing the content, the
legality of the content, deciding under a
very short time frames what should stay
online and what should go offline with no
public scrutiny or transparency about
practices that they kept deploying. And
they keep deploying to this day. Of
course, platforms under the threat of
legal liability often had to rely and
still have to rely on the content
recognition technologies for removing user
generated content. A typical example could
be also Avia law, which will be still
mentioned today during the presentation.
And the typical time frames are usually
those that extend from one hour to 24
hours, which is extremely short,
especially if any users would like to
appeal such a decision or seek an
effective remedy. At the same time, due to
the lack of harmonization and lack of
proper set of responsibilities that should
lie in the hands of these online
platforms. There was a lack of legal
certainty which would only reinforce the
vicious circle of removing more and more
of online content in order to escape any
possible liability. And at the end, to
this day, due to the lack of transparency,
we lack any evidence or research based
policy making, because platforms do not
want to share or inform the public
authorities what they actually do with the
content, how they moderate and those
transparency information that we receive
within their transparency reports are
usually quantity oriented instead of
quality. So they focus on how much content
is actually being removed and how fast,
which is not enough in order to create
laws that can actually provide any more
sustainable solutions. And ultimately, as
we all agree, the core issue doesn't lie
that much with how the content is being
moderated, but how content is being
distributed across platforms within the
core of their business models that
actually stands on a attention economy and
on the way, how sensational content is
often being amplified in order to actually
prolong that attention span of users that
visit platforms on regular basis. And I'm
packing quite a few issues here. And this
supposed to be just like a quick
introductory remark. I will now hand over
to Chris that will elaborate on all these
points a little bit further, and then we
take a look and unpack a few quite
important parts of the essay that we feel
should be prioritized in this debate at
the moment. Chris, over to you.
Christoph Schmon: Hi everybody. I'm
quite sure that many of you have noticed
that there's a growing appetite from the
side of the European Union to regulate the
Internet by using online platforms, as the
helping hands to monitor and censor what
users can say/share/do online. As you
see on the slide, the first highlight of
this growing appetite was corporate upload
filters, which are supposed to stop
copyright protected content online.
Thousands of people, old and young, went
on the streets to demonstrate for free
Internet, to demonstrate against technical
measures to turn the Internet into some
sort of censorship machine. We've made a
point then, and we continue making the
point now that upload filters are prone to
error, that upload filters cannot
understand context, that they are
unaffordable by all but the largest tech
companies, which happen to be all based in
the United States. But as you well know,
policymakers would not listen and
Article 13 of the copyright directive was
adopted by a small margin in the European
Parliament, also because some members of
European Parliament had troubles to press
the right buttons, but I think it's
important for us to understand that the
fight is far from over. The member states
must now implement the directive in the
way that is not at odds with fundamental
rights. And we argue that mandated
automated removal technologies are always
in conflict with fundamental rights. And
this includes data protection rights. It
is a data protection right not to be made
subject to Automated Decision-Making
online, if it involves your personal data
and if such decision making has a negative
effect. So we believe, all those legal
arguments aside, I think the most worrying
experience with upload filters is that it
has spillover effects to other
initiatives. Sure, if it works for
copyright protected companies, it may well
work for other types of content, right?
Except that it doesn't. Many considered
now a clever idea that web forums should
proactively monitor and check all sorts of
user content. May this be communication,
pictures or videos, and they should use
filters to take it down or the filters to
prevent the re-upload of such content. An
example of such spillover effect that Eliska
had mentioned is the draft regulation of
terrorist related content. It took a huge
joint effort of civil society groups and
some members of Parliament to reject the
worst of all text. We had recent
negotiations going on and at least we
managed to get out the requirement to use
uploads filters, but still a twenty four
hours removal obligation that may nudge
platforms to employ those filters
nevertheless. And we see that particularly
in national states, they are very fond of
the idea that platforms rather than
judges should be the new law
enforcers. There are now several states in
the European Union that have adopted laws
that would either oblige or nudge
platforms to monitor users speech
online. First up was the German NetzDG,
which set out systematic duties for
platforms. Then we had the French law
Avia, which copy-pasted the NetzDG and
made it worse. And last we have the
Austrian Hate Speech bill, which is a mix
of both the German and the French
proposal. They all go much beyond
copyright content, but focus on hate
speech and all sorts of content that may
be considered problematic and in those
respective countries, not necessarily in
other countries. And this brings me to the
next problem. How do we deal with content
that is illegal in one country, but legal
in another? A recent Court of Justice
ruling had confirmed that a court of a
small state like Austria can order
platforms not only to take down defamatory
content globally, but also to take down
identical or equivalent material using
automated technologies. For us this is a
terrible outcome, that will lead to a race
to the bottom, where the countries with
the least freedom of speech friendly laws
can superimpose their laws on every other
state in the world. We really believe that
all this nonsense has to stop. It's time
to acknowledge that the Internet is a
global space, a place of exchange of
creativity and the place where civil
liberties are suppose to exist. So we are
fighting now against all those national
initiatives. We had the great first
victory, when we have to bring down the
French law Avia - the Avia bill that had
imposed the duty for platforms to check
and remove potentially illegal content
within 24 hours. Before the Conseil
constitutionnel, the French Supreme Court,
we had argued that this would push
platforms to constantly check what users
post. And if platforms face high
fines... of course, they would be rather
motivated to block as much context of the
content as possible. We've made a point
that this would be against the Charter of
Fundamental Rights, including freedom of
information and freedom of expression. And
it was a great victory for us that the
French Supreme Court has struck down the
French Avia bill and followed our
argument, as you see on the slide. We also
see now that there's a push back at least
against the update of the German NetzDG,
which would have provided new access
rights for law enforcement authorities.
This and other provisions are considered
unconstitutional. And as far as I
understand and, perhaps listeners can
correct me, the German president has
refused to sign the bill and the Austrian
bill goes the similar path way - got
recently a red light from Brussels. The
commission considers it in conflict with
EU principles. Also, thanks to joint
effort by epicenter.works. And this shows
that something positive is going on, it's
a positive development, the pushback
against automated filter technologies.
But it's important to understand that
those national initiatives are not just
purely national attempts to regulate hate
speech. It's an attempt of an EU member
state to make their own bills, as badly as
they are, some sort of a prototype for EU-
wide legislation, a prototype for the
Digital Services Act. And as you know,
national member states have a say in EU
lawmaking: their voices are represented in
the council of the EU and the European
Commission, which will be disincentivized
to propose anything that would be voted
down by the council. I think that's a nice
takeaway from today, that lawmaking in
national member states is not an isolated
event. It's always political, it's always
Netzpolitik. The good news is that as
far as the European Commission proposal
for a Digital Services Act is concerned,
that it has not followed the footsteps of
those bad, badly designed and misguided
bills. It has respected our input, the
input from Access Now, from EFF, from the
EDRi network, from academics and many
others, that some key principles should
not be removed, like that liability for
speech should rest with the speaker. The
DSA, it's also a red light to channel
monitoring of users content. And there are
no sure badlines in there to remove
content that might be illegal. Instead,
the commission gives more slack to
platforms to take down posts in good faith,
which we call the EU style
Good Samaritan clause. Looking
through the global lenses of law making
it's very fascinating to see that while
the United States is flirting with the
idea to move away from the Good Samaritan
principle in Section 230 of the
Communications Decency Act, so the idea is
that platforms can voluntarily remove
content without being held liable for it,
the European Union flirts with the idea to
introduce it, to give more options to
platforms to act. That being said, the
major differences between the US and the
EU is that in Europe, platforms could be
held liable the moment they become aware
of the illegality of content. That's an
issue because the Digital Services Act has
now introduced a relatively sophisticated
system for user notification, complaint
mechanism, dispute resolution options,
which all leads to such awareness about
illegality or could lead to such
awareness. It's not quite clear for us how
platforms will make use of voluntary
measures to remove content. That being
said, we think that the commission's
proposal could have been much worse. And
the Parliament's reports on Digital
Services Act have demonstrated that the
new parliament is a bit better than the
old one. They have a lot of respect for
fundamental rights. So many members of
parliament that are quite fond of the idea
to protect civil liberties online. But we
know that this was only the start and we
know that we need another joint effort to
ensure that users are not monitored, they
are not at the mercy of algorithmic decision
making. And I think Eliska is now going to
explain a bit more about all this.
Eliska: Thank you. Thank you very much,
Chris. So, we can actually move now
further and unpack a few relevant
provisions for everything that has already
been mentioned, mainly in the realm of
content moderation and content creation,
which ultimately lies in the core of
Digital Services Act. And maybe not to
make it all so abstract... I also have the
printed version of the law here with me.
And if you look at it, it's quite an
impressive piece of work that the European
Commission did there. And I have to say
that even though it's a great start, it
still contains a lot of imperfections. And
I will try to summarize those now for you,
especially in the light of our end positioning
and as I mean civil societies in general,
because I believe that on all those
points, we have a pretty solid agreement
among each other. And what we were
actually hoping that Digital Services Act
will do, what it actually does and where
we see that we will need to actually
continue working very closely, especially
with the members of the European
Parliament in the future, once the draft
will actually enter the European
Parliament, which should happen relatively
soon. So as it already, as I already
mentioned at the beginning, quite briefly,
is how actually DSA distinguishes between
online platforms, which are defined within
the scope of the law and between very
large online platforms, which is exactly
that scope where all large online
gatekeepers fall into. DSA specificly then
distinguishes between obligations or
responsibilities of these actors, some
assigning to all of them, including online
platforms and some being extended
specifically due to the dominance and
power of these online gatekeepers hold.
This is mainly then the case when we
discuss the requirements for
transparency. There is a set of
requirements for transparency that apply
to online platforms, but then there is
still specific additional set of
transparency requirements for larger
online platforms. What DSA does especially
and this is the bed which is extremely
relevant for the content moderation
practices - it attempts to establish a
harmonized model for notice and action
procedure for allegedly illegal content.
Whatever alert red lines before I go into
the details on this was that DSA will be
touching only, or trying to regulate only
allegedly illegal content and stay away
from vaguely defined content categories
such as potentially harmful, but legal
content. There are other ways how the
legislation can tackle this content,
mainly through the meaningful transparency
requirements, accountability, tackling
issues within the open content recommender
systems and algorithmic curation. But we
didn't want the specific category of the
content to be included within the scope of
DSA. This is due to the fact that vaguely
defined terms that find their way into
legislation always lead to human rights
abuse in the future. I could give you
examples from Europe, such as the concept
of online harms within the UK, but also as
a global organizations. Both of us, we
actually often see how weak terminology
can quickly lead to even over
criminalization of speech or suppressing
the decent. Now, if we go back to
harmonize notice and action procedure,
what that actually means in practice, as
Christoph already mentioned, Europe has
so-called conditional model of
intermediate reliability, which is being
provided already and established by the
initial legal regime, which is the
e-Commerce Directive under Article 14 of
the e-Commerce Directive, which actually
states that unless the platform holds the
actual knowledge and according to the
wording of DSA now it's the actual
knowledge or awareness about the presence
of illegal content on their platform, they
cannot be held liable for such a content.
Now, we were asking for a harmonized
procedure regarding notice and action across
the EU for a while, precisely because we
wanted to see reinforced legal certainty.
Lack of legal certainty often translated
into overremoval of even legitimate
content from the platforms with no public
scrutiny. DSA does a good job, it's a good
starting point that actually tries to
attempt, to establish such a harmonized
procedure, but it's still lacking behind
on many aspects that we consider important
in order to strengthen protection of
fundamental rights of users. One of them
is, for instance, that harmonized notice
and action procedure, as envisioned by
DSA, is not specifically tailored to
different types of categories of user
generated content. And as we know, there
were some or many categories of content
that are deeply context dependent, linked
to the historical and sociopolitical
context of member state in question. And
due to their reliance on the automated
measures, that usually context blind, we
are worried that if notice and action
doesn't reflect this aspect in any further
ways we will end up again with over
removals of the content. What is probably
another huge issue that we are currently
lacking in the draft, even though DSA is
trying to create a proper appeal and
enforcement mechanisms and also appeal
mechanisms for users and different
alternative dispute settlement of the law
draft currently contains, there is no
possibility for content providers, for the
user that uploads the filter.. Sorry.
laughs That was a nice Freudian slip
there. For
a user that actually uploaded content to
appeal to directly actually use the
counter notification about that notified
content that belongs to that user. Nor
platforms are obliged to actually send the
notification to a user prior to any action
that is being taken against that
particular notified content. These are for
us a procedural safeguards for fairness
that users should have, and currently they
are not being reflected in the draft.
However, this is a good start and it's
something that we were pushing for. But I
think there are many more aspects that
these notice and action procedures will
need to contain in order to truly put
users at first. Now the notice and action
procedure is mainly focusing on the
illegal content. But there are ways in the
draft where potentially harmful content,
which is still legal - so the content that
actually violates the terms of service of
platforms is being mentioned throughout
the draft. So for us, it's now at the
moment exactly clear how that will work in
practice. So that's why we often use this
phrase that is also put on the slide: good
intention with imperfect solutions.
However, I want to emphasize again that
this is just the beginning and we will
still have time and space to work very
hard on this. Another kind of novel aspect
that DSA actually brings about is already
mentioned Good Samaritan Clause, and I
tend to call it the EU model or EU version
of Good Samaritan Clause. Good Samaritan
Clause originates in Section 230 of
Communication Decency Act, as Cristoph
already mentioned. But within the European
realm, it goes hand in hand with this
conditional model of liability which is
being preserved within the DSA legal
draft. That was also one of our main ask
to preserve this conditional model of
liability and it's great that this time
European Commission really listened. Why
we consider the Good Samaritan Clause
being important? In the past, when such a
security wasn't enshrined in the law, but
it was just somehow vaguely promised to
the commission that if the platform will
proactively deploy measures to fight
against the spread of illegal content,
they won't be held liable without
acknowledging that through such a use of
so-called proactive measures, the platform
could in theory gain the actual knowledge
about the existence of such a type of
content on its platform, which would
immediately trigger legal liability. This
threat of liability often pushed platforms
to the corner so they would rather remove
the content very quickly then to face more
serious consequences later on. That's why
we see the importance within the Good
Samaritan Clause or the European model of
Good Samaritan Clause, and we are glad
that it's currently being part of the
draft. One of the biggest downfalls or one
of the biggest disappointments when DSA
finally came out on the 15th of December
for us was to see that it's still online
platforms that will remain in charge when
it comes to assessing the legality of the
content and deciding what content should
be actually restricted and removed from a
platform and what should be available. We
often emphasize that it's very important
that the legality of the content is being
assessed by the independent judicial
authorities as in line with the rule of
law principles. We also do understand that
such a solution creates a big burden on
the judicial structure of member states.
Many member states see that as a very
expensive solutions, they don't always
want to create a special network, of
courts, or e-courts or other forms of
judicial review of the illegal or
allegedly illegal content. But we still
wanted to see more public scrutiny,
because for us this is truly just the
reaffirmation of already existing status
quo, as at the moment and there are many
jurisdictions within the EU and in the EU
itself, it's still online platform that
will call the final shots. What, on the
other hand, is a positive outcome that
we were also hardly pushing for are the
requirements for meaningful transparency.
So to understand better what platforms
actually do with the individual pieces of
content that are being shared on these
platforms and how actually transparency
can then ultimately empower user. Now, I
want to emphasize this because this is
still ongoing debate and we will touch
upon those issues in a minute. But we
don't see transparency as a silver bullet
to the issues such as amplification of
potentially harmful content or in general
that transparency will be enough to
actually hold platforms accountable.
Absolutely not. It will never be enough,
but it's a precondition for us to actually
seek such solutions in the future. DSA
contains specific requirements for
transparency, as I already mentioned, a
set of requirements that will be
applicable largely to all online platforms
and then still specific set of
requirements on the top of it. That will
be applicable only to very large online
platforms so the online gatekeepers. We
appreciate the effort, we see that the
list is very promising, but we still think
it could be more ambitious. Both EFF and
Access Now put forward a specific set of
requirements for meaningful transparency
that are in our positions. And so did EDRi
and other civil society or digital rights
activists in this space. And final point
that I'm going to make is the so-called
Pandora box of online targeting and
recommender systems. Why do I refer to
this as to Pandora Box? When a European
Parliament published its initiative
reports on DSA, there are two reports, one
being tabled by JURI Committee and then
another one by ENCO, especially the JURI
report contained paragraph 17, which calls
out for a better regulation of online
targeting and online advertisements, and
specifically calling for a ban of online
targeting and including the Phase-Out that
will then lead to a ban. We supported this
paragraph, which at the end was voted for
and is the part of the final report.
Nevertheless, we also do understand that
these wording of the article has to be
more nuanced in the future. Before I go
into the details there, I just want to say
that this part has never made it to DSA.
So there is no ban on online targeting or
online advertisement of any sort, which to
us, to some extent, it was certainly
disappointing too, we specifically would
call for a much more stricter approach
when it comes to behavioral targeting as
well as crossside tracking of online
users, but unfortunately, and as we
eventually also heard from Commissioner
Vestager, that was simply lag of will or,
maybe, too much pressure from other
lobbies in Brussels. And this provision
never found its way to the final draft of
DSA. That's the current state of art, we
will see what we will manage to achieve
once the DSA will enter the European
Parliament. And finally, the law also
contains a specific provision on
recommender systems. So the way how the
content is being distributed across
platform and how the data of users are
being abused for such a distribution and
personalization of user generated content.
In both cases, whether it's online
targeting and recommender systems within
the DSA, DSA goes as far as the
transparency requirements, the
explainability, but it does very little
for returning that control and empowerment
back to the user. So whether user can
obtain or opt out from these algorithmic
curation models, how it can actually be
optimized if they decide to optimize it?
All of that is at the moment very much
left outside of the scope of DSA. And so
that's the issue of interoperability,
which is definitely one of the key
issues being currently discussed and made
kind of possible hopes in the future for
returning that control and empowerment
back to the user. And I keep repeating
this as a mantra, but it's truly the main
driving force behind all our initiatives
and the work we do in these fields. So the
user and their fundamental rights. And on
that note, I would like to hand over back
to Chris, who will explain the issue of
interoperability and how to actually
empower you as a user and to strengthen
the protection of fundamental rights
further. Chris, it's yours now. Christoph:
Thank you. I think we all know or feel
that the Internet has seen better times.
If you look back over the last 20 years,
we have seen that transformation was going
on from an open Internet towards a more
closed one - monopolization. Big platforms
have built entire ecosystems and it seems
that they alone decide who gets to use
them. Those platforms have strong network
effects that have pushed platforms or
those platforms into gatekeeper position
which made it so easy for them to avoid
any real competition. This is especially
true when we think of social media
platforms. This year we celebrate the 20th
birthday of the e-Commerce Directive that
Eliska mentioned. The Internet bill that
will now be replaced by the Digital
Services Act. We believe it's a very good
time now to think and make a choice:
should we give even more power to the big
platforms that have created a lot of the
mess in the first place; or should we give
the power to the users, give the power
back to the people? For us, the answer is
clear. Big tech companies already employ a
wide array of technical measures. They
monitor, they remove, they disrespect user
privacy and the idea to turn them into
the Internet Police, with a special
license of censoring the speech of users,
will only solidify their dominance. So we
wouldn't like that. What we like is to
put users in charge over their online
experience. Users should, if we had a say,
choose for themselves which kind of
content they can see, what services they
can use to talk to their friends and
families. And we believe it's perhaps time
to break up those silos, those big
platforms have become to end the dominance
over data. One element to achieve this
would be to tackle the targeted ads
industry, as Eliska mentioned it, perhaps
to give an actual right to users not to be
subject to targeted ads or to give more
choice to use to decide, which content
they would like to see or not to see. In
the Digital Services Act, the Commission
went for transparency when it comes to ads
and better option for users to decide on
the recommended content, which is a start,
we can work with that. Another important
element to achieve user autonomy over data
is interoperability. If the European Union
really wants to break the power of those
data driven platforms that monopolize the
Internet, it needs regulations that
enables users to be in control over the
data. We believe that users should be able
to access data, to download data, to move,
manipulate their data as they see fit. And
part of that control is to port data
from one place to another. But data
portability, which we have under the
GDPR is not good enough. And we
see from the GDPR that it's not
working in practice. Users should be able
to communicate with friends across
platform boundaries, to be able to follow
their favorite content across different
platforms without having to create several
accounts. But to put it in other terms, if
you upset with the absence of privacy on
Facebook or how the content is moderated
on Facebook, you should be able to just
take your data with you using portability
options and move to an alternative
platforms, that is a better fit and this
without losing touch with your friends who
stay behind, who have not left the
incumbent big platform. So what we did for
Digital Services Act is to argue for
mandatory interoperability options that
would force Facebook to maintain APIs that
let users on other platforms exchange
messages and content with Facebook users.
However, if you look in the DSA, we see
that the commission completely missed the
mark on interoperability, which is
supposed to be dealt with by related legal
act, now it gets complicated. It's the
Digital Markets Act, the DMA, another
beautiful acronym. The Digital Markets Act
wants to tackle certain harmful business
practices by those gatekeeper platforms,
the very large tech companies that control
what is called core services. The core
service is a search engine, a social
networking service, a messaging service,
its operating systems and online
intermediation services. Like think of how
Amazon controls access to customers for
merchants that sell on its platforms or
how the Android and iPhone app stores as
chokepoints in delivering mobile software.
And many things we like in the new
proposal, the proposal of the Digital
Markets Act, for example there's a ban on
mixing data in there that you may wants to
ban gatekeeper's from mixing data from
data progress with the data they collect
on the customers. Another rule is to ban
cross tying - sort of practices that end
users must sign up for ancillary services.
So you should be able to use Android
without having to get a Google account for
example. You believe that this is all
good, but the DMA like the DSA is very
weak on interoperability. What it does is
to focus on real time data portability
instead. So instead of having
interoperable services, users will only be
able to send the data from one service to
another like from Facebook to Diaspora,
meaning that you would end up having two
accounts instead of one or to quote Cory
Doctorow who spoke yesterday already:
"Users would still be subject to the
sprawling garbage novela of abusive legalese
Facebook lovably calls its terms of
service." We believe that this is not
good enough. And the last slide, you see a
quote from the Margrethe Vestager who made
a very good statement last month, that we
need trustworthy services, fair use of
data and free speech and an interoperable
internet; we fully agree on that. And in
the next months and years, we will work on
this to actually happen. However, you can
imagine, it will not be easy. We already
see that European Union member states
follow the trend of platforms should
systematically check undesirable and
insightful content and share those data
with enforcement authorities, which is
even worse. We see an international trend
going on to move away from the immunity of
platform for use of content towards a more
active stance of those platforms. And we
see that recent terror attacks have fueled
ideas that monitoring is a good idea and
end to end encryption is a problem. So
whatever will be the result, you can bet
that European Union will want to make the
Digital Services Act and the Digital
Markets Act another export model. So this
time we want the numbers right in
parliament and the council, we want to
help members of parliament to press the
right buttons. And for all this we will
need your help, even if it means to learn
yet another acronym or several acronyms
after the GDPR. That's it from
our side - we are looking forward to the
discussion.Thank you.
Herald: OK, thank you Eliska and
Christoph. There are questions from the
internet and the first one is basically
we just have and as you mentioned in your
slides, Christoph, the copyright in the
digital single market with both
accountability and liability provisions,
you also briefly mentioned, I think even
the e-evidence proposal also. How do all
these proposals relate to each other? And
especially for a layperson, that is not
into all the Brussels jargon.
Christoph: I think Eliska, you raised
your hand, don't you?
Eliska: ...more or less unintentionally,
but yeah, kind of that. I can start and
then let you Christoph to step in. Yeah,
that's a very, very good question. And
this is specifically due to the fact that
when you mention especially online
terrorist content regulation, but also
recently proposed interim regulation on
child sexual abuse, they.. all these - we
call them sectoral legislation, so kind of
a little bit of parting from this
horizontal approach, meaning an approach
that tackles all categories of illegal
content in one way, instead of going after
specific categories such as online
terrorist content in the separate ways. So
it's a little bit paradoxical saying what
is currently also happening at the EU
level, because on one hand, we were
promised this systemic regulation that
will once for all establish harmonized
approach to combating illegal content
online and at the same time, which is
specifically DSA, the Digital Services
Act, and at the same time we still see
European Commission allowing for these
fundamental rights harmful legislative
proposals happening in these specific
sectors such as proposed online
terrorist content regulation or other
legislative acts seeking to somehow
regulate specific categories of user
generated content. This is quite puzzling
for us as a digital rights activists too,
and very often, actually, so I would maybe
separate DSA from this for a moment and
say that all of these sectoral
legislations what they have in common is:
first of all, continuing these negative
legislative trends that we already
described and that we constantly observe
in practice, such as shifting more and
more responsibility on online platforms.
And at the same time, what is also very
interesting, what they have in common is
the legal basis that they stand on, and
that's the legal basis that is rather
connected to the cooperation within the
digital single market, even though they
seek to tackle a very particular type of
category of content category, which is
manifestly illegal. So logically, if they
should have that appropriate legal ground,
it should be something more close to
police and judicial cooperation, which we
don't see happening in practice,
specifically because there is this idea
that platforms are the best suited to
decide how the illegal content will be
tackled in the online space. They can be
the fastest, they can be the most
effective. So they should actually have
that main decision making powers and
forced into taking those responsibilities
which have ever ultimately, according to
the rule of law principle, should and have
to be in the hands of the state and public
authorities, preferably judicial
authorities. So I would say they are all
bad news for fundamental rights protection
of online users, civil rights
organizations, all of us that are on this
call today. We're fighting very hard also
against the online service content
regulation. There was a lot of damage
control done, especially with the first
report that was tabled by the European
Parliament and also now during the last
trialogue since the negotiations seems to
be concluded and the outcome is not great.
It's far from ideal. And I'm worried that
with other sectoral legislative attempts
coming from the European Commission, we
might see the same outcome. It will be
very interesting to see how that will
actually then play together with the
Digital Services Act, which is trying to
do the exact opposite to actually fix this
negative legislative efforts that we see
at the EU level with these sectoral
legislation, but also with the member
states at the national level. I could also
mention the European Commission reaction
to some national legislative proposals.
But Christoph, I would leave that to you
and please step in.
Christoph: I think you explained it
perfectly, and the only thing I can
supplement here is that if you look at
this move from sectoral legislation,
asylum legislation to horizontal
legislation, now back to sectoral
legislation - it's a problem, it's a mess.
First, the two sides not very good
coordinated which brings troubles for
legal certainty. It makes it very
troublesome for platforms to follow up.
And it's problematic for us, for us in the
space. We are some sort of lobbyist as
well, just for public interest. But you
will have to have to deal with copyright,
with CSAM, with TERREG, with end to
end encryption, DSA, DMA and 15 other
parties to pop up content by content. It's
very hard to manage to have the capacity
ready to be early in the debate, and it's
so important to be early in the debate to
prevent that from happening. And I think
that's a huge challenge for us, to have
something for us to reflect to in the next
days. How can we join forces better in a
more systematic way in order to really
follow up on all those initiatives? That's
for me, a very problematic development.
Herald: So in summary it's a mess. So it
is related, but we can't explain how,
because it's such a mess. Fair enough.
I have another question for you, Eliska.
Someone was asking how the proposed
Good Samaritan clause works compared to..
how it currently works in Germany. But I
think it's a bit unreasonable to expect
everyone to know how it works in Germany.
I would rephrase it this as: how does this
proposed Good Samaritan clause work
compared to how it is now under the
e-Commerce Directive?
Eliska: Thank you very much. Yeah, so a
great question again, I think the first if
we put it into the context of the EU law
and apologies that I cannot really answer
how - you know, compare the German context
- I really don't dare to, I'm not a
German lawyer, so I wouldn't like to step
in those waters. But first of all, there
is no Good Samaritan clause per se within
the scope of e-Commerce Directive. It did
not really exist within the law. And I'm
using the pass sentence now because DSA is
trying to change that. So that level of
legal certainty was not really, really
there for the platforms. There was the
conditional model of the liability, which
is still preserved within the regulation.
But if you think of a Good Samaritan
clause as we know it from the section
230, or let's use that Good Samaritan
clause as an example, because also
e-Commerce Directive was actually drafted
as a response to Communication Decency Act
that was the legislation that puts things
into motion. So that's the first
ultimate point. I explain at the
beginning in my presentation what was then
happening in the space of combating
illegal content at the EU level, and
especially I would refer to the
communication that the European Commission
published, I think, back in 2018, where it
actually encouraged and called on online
platforms to proactively engage with
illegal content and use these proactive
measures to actually seek an adequate
response to illegal content. Now, to mix
that with this conditional model of
liability, which is of course defined by
the obtaining actual knowledge by the
platform that created a perfect storm that
I already explained. So the platforms knew
that they are kind of pushed by the
legislature to actually seek these active
responses to illegal content, often
deploying automated measures. But they
didn't have any legal certainty or
security on their side that if they do so,
they won't end up ultimately being held
legally liable and face legal consequences
as a result of obtaining actual knowledge
through those proactive measures that were
kind of the tool, how they could possibly
actually obtain that knowledge. Now, what
DSA does, it specifically actually simply
states and I think it's Article 6 in the
Digital Services Act, if I'm not mistaken,
and I can even open it, it specifically
basically says that platforms can use
these proactive measures or, you know,
continue using some tools that actually
seek to provide some responses to this
type of content without the fear of being
held liable. So it's it's an article which
has approximately, I think, two
paragraphs, but it's finally in the
legislation and that means that it will
help to reinforce the level of legal
certainty. I would also emphasize that
very often in Europe, when we discuss Good
Samaritan clause, and Good Samaritan is
actually very unfortunate term, because
it's very much connected to the American
legal tradition. But when it's being mixed
up with the conditional model of liability
and with the prohibition of general
monitoring, which is still upheld, these
are the main principles of the European
intermediary reliability law and the
regime that is applicable within the EU,
such a safeguard can be actually
beneficial and it won't lead hopefully to
these blanket immunity for online
platforms or to this idea that platforms
will be able to do whatever they want with
the illegal content without any public
scrutiny, because there are other
measures, safeguards and principles in
place as a part of conditional model of
liability that we have here in Europe. So
I'm sorry, maybe that was too complicated.
Legalistic explanation there. But this is
how these provisions should work in
practice. We, of course, have to wait for
the implementation of the law and see how
that will turn out. But the main purpose
is that this legal certainty that was
lacking until now can finally come to its
existence, which should help us to prevent
over removal of legitimate speech from
online platforms.
Herald: OK, thank you. I have two other
questions from the Internet about
interoperability, and I suppose I should
look at Christoph for them. The last one
I'm going to ask first is: would such
interoperability make it much more
difficult to combat harassment and
stalking on the Internet? How do you
police that kind of misbehavior if it's
across different platforms who are forced
to interoperate and also be conduits for
such bad behavior. And I'll come to the
earlier question if you've answered this
question Christoph.
Christoph: It's a pretty good question.
First, to understand our vision on
interoperability is to understand that we
would like to have it between platforms
that empower large platforms and the right
of smaller platforms, actually, to make
use of interoperability. So it should not
be among the big platforms. So small
platforms should be able to connect to the
big platforms. And second, we believe it
will help and not make it worse because we
have now a problem of hate speech, we have
now a problem of a lack of privacy, we
have now a problem of the attention
industry that works with, you know,
certain pictures put in certain frames to
trigger the attention of users, because users
don't have a choice of the content
moderation practices, users don't have a
choice to see which kind of content should
be shown. And users don't have options to
regulate the privacy. The idea of more
competitors would be exactly that I can
move to a space, where I'm not harassed
and not be made subject to certain content
that hurt my feelings. Right. And that
moment I get control. I can choose a
provider that gives me those options and
we would like even to go a step further.
Back end interoperability was a start. We
believe if users want to, they should be
able to delegate a third party company or
piece of a third party software to
interact with the platform on their
behalf. So users would have the option to
see a news feed in different order,
calibrate their own filters on
misinformation. So in this sense,
interoperability can be a great tool,
actually, to tackle hate speech and to
sort of negative developments. Of course,
there is a risk to it. I think the risk
comes rather from the data industry side
again, that we need to take care not to
place one or another data selling
industry on the one that we already face.
But for this, we have options as well to
avoid that from happening. But to answer
the question, we believe interoperability
is a tool actually to escape from the
negative developments you had mentioned.
Herald: Critical counter question for me
then, aren't you actually advocating for
just roll your own recommendation engines
to be able to do so? Can't you achieve
that without interoperability?
Christoph: Sure. Recounter question: Do
you think an average user can accomplish
that quite easily? You know, like when we
look at the Internet through the lenses of
market competition then we see that it is
the dominance of platforms over data that
have created those spaces, those developed
gardens where users have the feeling they are
trapped and cannot escape from. And there
are so many alternative options that can
not get off the ground because users feel
trapped, don't want to leave their friends
behind and don't have options, actually to
have a better moderation system. Of course,
you can be creative and, you know, use
plugins and whatever you see fit, but you
need to stay within the platform barriers.
But we would like to enable users to actually
leave developed garden, go to another
place, but still stay in touch with
friends who have made the choice to remain
there. And I think that's perhaps the
difference to what you had in mind.
Herald: I have a follow up question. Well,
another question from the Internet,
regardless of interoperability, and that
is, historically speaking, as soon as the
big players get involved in certain
standards, they tend to also shape policy
by being involved in that. How would that
be different in the case of
interoperability and specifically
mentioned by the person who ask the
question. That Mastodon probably
flourishes because nobody else was
involved in setting that standard.
Christoph: Ah it's an excellent question.
And we struggled with the question of
standards ourselves in our policy paper,
which is our recommendations for European
Union to enact certain provisions in the
new Digital Services Act. We abstain from
asking to establish new standards like API
standards. We believe it's a bad idea to
regulate technology like that. What we
want to do is that big platforms just
offer interoperability however they see fit.
We don't want to have a standard that can
be either again monopolized or lobbied by
the big platforms. Because then we end up
with the standards we already see
which we don't like. But it's a good
question. And what we did is with our
policy principles on interoperability to
give kind of a food for thought, how we
believe the end version should look like,
but the many questions that remain and we
don't know exactly the way how to go
there.
Herald: Yeah, I'm sorry of sticking to the
topic of interoperability, because most
questions are actually about that. One of
the other questions is how do we prevent
this from getting messed up like it
happens with PSD2? And for the audience
that don't know about PSD2 - PSD2 is a
Directive that forced banks to open up
APIs to other financial service providers,
which is also interoperability between
platforms, in this case for banking
platforms, which comes with all sorts of
privacy questions that weren't completely
thought through when that legislation came
about. Sorry for having this long winded
intoduction, Christoph, but I think it was
needed for people that don't know what
PSD2 means.
Christoph: It's a good question.
Interestingly, we never used PSD2 or the
Telecommunications Act because both have
interoperability options as those negative
examples we always used as examples that
hey, it's already possible. So you don't
have an excuse to say it's impossible to put
it in the law. What is true is that there
is a lot of mess around it. The question
of how to avoid the mess, it's a question
of.. Netzpolitik again. So the question
of whether policy makers are actually
listening to us or listening to industry
lobbyists. So the ones who raised the
question is absolutely right - there's a
huge risk for every topic we talk about.
Whether interoperability, whether it's use
of control about content, targeted ads,
liability, everything that we believe
should be the law, of course, could be
hijacked, could be redesigned in a way
that it will lead to more problems than
fewer problems. So, indeed, for every
policy question we raise, we need to ask
ourselves, is it worth the fight to risk
opening the box of the Pandora? Do we make
it worse? What we said is on that front -
we are happy to make a pressure and what
we need to do in the next year is to
convince them that we are the right person
to talk to. And that's perhaps a challenge
how to make that explicable to policy
makers. So those who ask the questions, I
think those should help us to come to
Brussels to the parliament and tell MEPs
how it's going to work.
Herald: On that note a question to both of
you. Citizen enforcement, I prefer the
term citizen overuses. Would it be helpful
to push for amendments in the parliament
for at least the targeting points you both
mentioned before? And if so, how?
Eliska: So I guess that I will start just
so Christoph can rest a little. So the
question was whether it would be useful to
push for those amendments? Was that's
right?
Herald: For amendments that cover the
targeting of citizens.
Eliska: Absolutely. So there is, of
course, short and long answer as to every
question. And so the short answer would be
yes, but given that the wording of such an
amendment will be precise and nuanced. We
are still working out our positioning on
online targeting and I think we all.. no
one can name those practices that we don't
want to see being deployed by platforms
and where we can actually imagine a proper
ban on such practices. We have recently
published one of our blog posts where we
actually unfold the way of thinking that
Access Now currently, you know, how we
are brainstorming about this whole issue.
And as I said, especially those.. that's
targeting that uses behavioral data of
users, citizens, then maybe let's go for
individuals because they are also obliged
to protect the rights of individuals that
are not their citizens. So that's
definitely one form where we can
definitely see and will be supporting the
ban and possible phase out that will
actually lead to a ban. The same goes for
the cross side tracking of users due to
the fact how users data are being abused
again as being the integral part of the
business models of these platforms and so
on and so forth. So that's one of the
direction we will be definitely taking.
And again, we are inviting all of you to
help us out, to brainstorm together with
us, to assess different options,
directions that we should take into
consideration and not forget about. But I
personally think that this will be one of
the main battles when it comes to DSA,
where we will definitely need to be on the
same page and harmonize and joint the
forces, because DSA gives us a good ground
at the moment, but it doesn't go far
enough. So, yes, definitely the answer is
yes, but given that, we will have a very
nuanced position so we know what we are
asking for and we are taking into
consideration also those other aspects
that could eventually play out badly in
practice. So good intentions are not
enough when it comes to DSA.
Herald: Thank you. I think we are slightly
over time, but I've been told beforehand
it's OK to do so for a few minutes and
there's three questions that are open. One
of one of them, I will answer myself. That
is basically: have the member states
responded? And the answer to that is no.
The member states have not taken any
position. And two others, I think are
quite interesting and important from
a techy perspective. One is: is there
anything you see that might affect current
decentralized platforms like the
Fediverse, Mastodon? And the other is:
will any review of the data protection,
sorry, the Database Protection Directive
affect make the search engines and
interact of these again?
Christoph: Perhaps I jump in and Eliska,
you take over. First, member states have
given an opinion actually on DSA, there
have been one-two official submissions,
plus joint letters, plus discussions in
council on whether the DSA was presented.
I have some nice protocols, which showed
the different attitude of member states
towards it. So for us, it also means we
need to work straight away with the
council to ensure that the package would
be good. What was the question? Ah, yes, I
think the answer to the question depends
on the what lawyers call materials scope
applications whether it would apply to
those platform models at all. But
Eliska you can help me out here. You have
always criticized for the e-Commerce
Directive that it was not quite clear how
it would relate to first non profit
platforms. And many of these alternative
platforms are like that, because there was
this issue of providing a service against
remuneration and what's not quite clear
what it means. Would it apply to Wikipedia
if you get like donations would have
applied to a blogger if you have, like,
pop ups, ads or something like that. So
that's, I think, one huge question. And
the second question is, in as much those
new due diligence obligation would force
alternative platforms, governance models
to redesign the interfaces. And those are
open question for us. We have not analyzed
that in detail, but we see that.. we have
worried that it would not only impact the
large platforms, but many others as well.
What do you think Eliska?
Eliska: Yeah, I can only agree and
especially regarding the non profit
question. Yeah, this is also or was and
always been one of our main asks for
nonprofit organizations. And actually it's
not quite clear how that will play out now
in practice. By the way, how DSA standing,
because at the moment it actually speaks
about online platforms and then it speaks
about very large online platforms, but
what it will be and how it will impact
nonprofit organizations, whether these are
bloggers or organizations like us civil
rights organizations that remain to be
seen. I also know that the European Court
of Human Rights and its jurisprudence try
to establish some principles for the
nonprofit since Delfi AS versus Estonia
and then with the empty e-decision and the
Swedish case that followed afterwards. But
I'm not sure how well it was actually
elaborated later on. But this is something
we will be definitely looking at and
working on further and regarding the
impact on the smaller players and also at
the interface idea, this is still
something we are actually also wondering
about and thinking how this will turn out
in practice. And we are hoping to actually
develop our positioning further on these
issues as well, because we actually
started working, all of us on DSA and on
our recommendations already. I think a
year ago, maybe a year and a half ago when
we were working just with the leagues that
Politico or other media platforms in
Brussels were actually sharing with us.
And we were working with the bits and
pieces and trying to put our thinking
together. Now we have the draft and I
think we need to do another round of very
detailed thinking. What will be ultimately
our position or what will be those
recommendations and amendments in the
European Parliament we will be supporting
and pushing for. So it's a period of
hard work for all of us. Not mentioning
that I always say that, you know, we stand
against a huge lobby power that is going
to be and is constantly being exercised by
these private actors. But I also have to
say that we had a very good cooperation
with the European Commission throughout
the process. And I think that I can say
that on behalf of all of us, that we feel
that the European Commission really listen
this time. So, yeah, more question marks
than answers here from me, I think.
Herald: This is fine with me, not knowing
something is fine. I think we're
definitely run out of time. Thank you both
for being here. Well, enjoy the 2D online
world and the Congress. Thank you. That
wraps the session up for us. Thank you all.
Christoph: Thank you.
Eliska: Thank you very much. Bye.
music
Subtitles created by c3subtitles.de
in the year 2021. Join, and help us!