- 
Musik
 
- 
Herald: Who of you is using Facebook? Twitter? 
Diaspora?
 
- 
concerned noise And all of that data
you enter there
 
- 
gets to server, gets into the hand of somebody
who's using it
 
- 
and the next talk
is especially about that,
 
- 
because there's also intelligent machines
and intelligent algorithms
 
- 
that try to make something
out of that data.
 
- 
So the post-doc researcher Jennifer Helsby
 
- 
of the University of Chicago,
which works in this
 
- 
intersection between policy and 
technology,
 
- 
will now ask you the question:
To who would we give that power?
 
- 
Dr. Helsby: Thanks.
applause
 
- 
Okay, so, today I'm gonna do a brief tour
of intelligent systems
 
- 
and how they're currently used
 
- 
and then we're gonna look at some examples
with respect
 
- 
to the properties that we might care about
 
- 
these systems having,
and I'll talk a little bit about
 
- 
some of the work that's been done in academia
 
- 
on these topics.
 
- 
And then we'll talk about some
promising paths forward.
 
- 
So, I wanna start with this:
Kranzberg's First Law of Technology
 
- 
So, it's not good or bad,
but it also isn't neutral.
 
- 
Technology shapes our world,
and it can act as
 
- 
a liberating force-- or an oppressive and
controlling force.
 
- 
So, in this talk, I'm gonna go
towards some of the aspects
 
- 
of intelligent systems that might be more
controlling in nature.
 
- 
So, as we all know,
 
- 
because of the rapidly decreasing cost
of storage and computation,
 
- 
along with the rise of new sensor technologies,
 
- 
data collection devices
are being pushed into every
 
- 
aspect of our lives: in our homes, our cars,
 
- 
in our pockets, on our wrists.
 
- 
And data collection systems act as intermediaries
 
- 
for a huge amount of human communication.
 
- 
And much of this data sits in government
 
- 
and corporate databases.
 
- 
So, in order to make use of this data,
 
- 
we need to be able to make some inferences.
 
- 
So, one way of approaching this is I can hire
 
- 
a lot of humans, and I can have these humans
 
- 
manually examine the data, and they can acquire
 
- 
expert knowledge of the domain, and then
 
- 
perhaps they can make some decisions
 
- 
or at least some recommendations
based on it.
 
- 
However, there's some problems with this.
 
- 
One is that it's slow, and thus expensive.
 
- 
It's also biased. We know that humans have
 
- 
all sorts of biases, both conscious and unconscious,
 
- 
and it would be nice to have a system
that did not have
 
- 
these inaccuracies.
 
- 
It's also not very transparent: I might
 
- 
not really know the factors that led to
 
- 
some decisions being made.
 
- 
Even humans themselves
often don't really understand
 
- 
why they came to a given decision, because
 
- 
of their being emotional in nature.
 
- 
And, thus, these human decision making systems
 
- 
are often difficult to audit.
 
- 
So, another way to proceed is maybe instead
 
- 
I study the system and the data carefully
 
- 
and I write down the best rules
for making a decision
 
- 
or, I can have a machine
dynamically figure out
 
- 
the best rules, as in machine learning.
 
- 
So, maybe this is a better approach.
 
- 
It's certainly fast, and thus cheap.
 
- 
And maybe I can construct
the system in such a way
 
- 
that it doesn't have the biases that are inherent
 
- 
in human decision making.
 
- 
And, since I've written these rules down,
 
- 
or a computer has learned these rules,
 
- 
then I can just show them to somebody, right?
 
- 
And then they can audit it.
 
- 
So, more and more decision making is being
 
- 
done in this way.
 
- 
And so, in this model, we take data
 
- 
we make an inference based on that data
 
- 
using these algorithms, and then
 
- 
we can take actions.
 
- 
And, when we take this more scientific approach
 
- 
to making decisions and optimizing for
 
- 
a desired outcome,
we can take an experimental approach
 
- 
so we can determine
which actions are most effective
 
- 
in achieving a desired outcome.
 
- 
Maybe there are some types of communication
 
- 
styles that are most effective
with certain people.
 
- 
I can perhaps deploy some individualized incentives
 
- 
to get the outcome that I desire.
 
- 
And, maybe even if I carefully design an experiment
 
- 
with the environment in which people make
 
- 
these decisions, perhaps even very small changes
 
- 
can introduce significant changes
in peoples' behavior.
 
- 
So, through these mechanisms,
and this experimental approach,
 
- 
I can maximize the probability
that humans do
 
- 
what I want.
 
- 
So, algorithmic decision making is being used
 
- 
in industry, and is used
in lots of other areas,
 
- 
from astrophysics to medicine, and is now
 
- 
moving into new domains, including
 
- 
government applications.
 
- 
So, we have recommendation engines like
Netflix, Yelp, SoundCloud,
 
- 
that direct our attention to what we should
 
- 
watch and listen to.
 
- 
Since 2009, Google uses
personalized searched results,
 
- 
including if you're not logged in
into your Google account.
 
- 
And we also have algorithm curation and filtering,
 
- 
as in the case of Facebook News Feed,
 
- 
Google News, Yahoo News,
 
- 
which shows you what news articles, for example,
 
- 
you should be looking at.
 
- 
And this is important, because a lot of people
 
- 
get news from these media.
 
- 
We even have algorithmic journalists!
 
- 
So, automatic systems generate articles
 
- 
about weather, traffic, or sports
 
- 
instead of a human.
 
- 
And, another application that's more recent
 
- 
is the use of predictive systems
 
- 
in political campaigns.
 
- 
So, political campaigns also now take this
 
- 
approach to predict on an individual basis
 
- 
which candidate voters
are likely to vote for.
 
- 
And then they can target,
on an individual basis,
 
- 
those that can be persuaded otherwise.
 
- 
And, finally, in the public sector,
 
- 
we're starting to use predictive systems
 
- 
in areas from policing, to health,
to education and energy.
 
- 
So, there are some advantages to this.
 
- 
So, one thing is that we can automate
 
- 
aspects of our lives
that we consider to be mundane
 
- 
using systems that are intelligent
 
- 
and adaptive enough.
 
- 
We can make use of all the data
 
- 
and really get the pieces of information we
 
- 
really care about.
 
- 
We can spend money in the most effective way,
 
- 
and we can do this with this experimental
 
- 
approach to optimize actions to produce
 
- 
desired outcomes.
 
- 
So, we can embed intelligence
 
- 
into all of these mundane objects
 
- 
and enable them to make decisions for us,
 
- 
and so that's what we're doing more and more,
 
- 
and we can have an object
that decides for us
 
- 
what temperature we should set our house,
 
- 
what we should be doing, etc.
 
- 
So, there might be some implications here.
 
- 
We want these systems
that do work on this data
 
- 
to increase the opportunities
available to us.
 
- 
But it might be that there are some implications
 
- 
that we have not carefully thought through.
 
- 
This is a new area, and people are only
 
- 
starting to scratch the surface of what the
 
- 
problems might be.
 
- 
In some cases, they might narrow the options
 
- 
available to people,
 
- 
and this approach subjects people to
 
- 
suggestive messaging intended to nudge them
 
- 
to a desired outcome.
 
- 
Some people may have a problem with that.
 
- 
Values we care about are not gonna be
 
- 
baked into these systems by default.
 
- 
It's also the case that some algorithmic systems
 
- 
facilitate work that we do not like.
 
- 
For example, in the case of mass surveillance.
 
- 
And even the same systems,
 
- 
used by different people or organizations,
 
- 
have very different consequences.
 
- 
For example, if I can predict
 
- 
with high accuracy, based on say search queries,
 
- 
who's gonna be admitted to a hospital,
 
- 
some people would be interested
in knowing that.
 
- 
You might be interested
in having your doctor know that.
 
- 
But that same predictive model
in the hands of
 
- 
an insurance company
has a very different implication.
 
- 
So, the point here is that these systems
 
- 
structure and influence how humans interact
 
- 
with each other, how they interact with society,
 
- 
and how they interact with government.
 
- 
And if they constrain what people can do,
 
- 
we should really care about this.
 
- 
So now I'm gonna go to
sort of an extreme case,
 
- 
just as an example, and that's this
Chinese Social Credit System.
 
- 
And so this is probably one of the more
 
- 
ambitious uses of data,
 
- 
that is used to rank each citizen
 
- 
based on their behavior, in China.
 
- 
So right now, there are various pilot systems
 
- 
deployed by various companies doing this in
China.
 
- 
They're currently voluntary, and by 2020
 
- 
this system is gonna be decided on,
 
- 
or a combination of the systems,
 
- 
that is gonna be mandatory for everyone.
 
- 
And so, in this system, there are some citizens,
 
- 
and a huge range of data sources are used.
 
- 
So, some of the data sources are
 
- 
your financial data,
 
- 
your criminal history,
 
- 
how many points you have
on your driver's license,
 
- 
medical information-- for example,
if you take birth control pills,
 
- 
that's incorporated.
 
- 
Your purchase history-- for example,
if you purchase games,
 
- 
you are down-ranked in the system.
 
- 
Some of the systems, not all of them,
 
- 
incorporate social media monitoring,
 
- 
which makes sense if you're a state like China,
 
- 
you probably want to know about
 
- 
political statements that people
are saying on social media.
 
- 
And, one of the more interesting parts is
 
- 
social network analysis:
looking at the relationships between people.
 
- 
So, if you have a close relationship with
somebody
 
- 
and they have a low credit score,
 
- 
that can have implications on your credit
score.
 
- 
So, the way that these scores
are generated is secret.
 
- 
And, according to the call for these systems
 
- 
put out by the government,
 
- 
the goal is to
"carry forward the sincerity and
 
- 
traditional virtues" and
establish the idea of a
 
- 
"sincerity culture."
 
- 
But wait, it gets better:
 
- 
so, there's a portal that enables citizens
 
- 
to look up the citizen score of anyone.
 
- 
And many people like this system,
 
- 
they think it's a fun game.
 
- 
They boast about it on social media,
 
- 
they put their score in their dating profile,
 
- 
because if you're ranked highly you're
 
- 
part of an exclusive club.
 
- 
You can get VIP treatment
at hotels and other companies.
 
- 
But the downside is that, if you're excluded
 
- 
from that club, your weak score
may have other implications,
 
- 
like being unable to get access
to credit, housing, jobs.
 
- 
There is some reporting that even travel visas
 
- 
might be restricted
if your score is particularly low.
 
- 
So, a system like this, for a state, is really
 
- 
the optimal solution
to the problem of the public.
 
- 
It constitutes a very subtle and insiduous
 
- 
mechanism of social control.
 
- 
You don't need to spend a lot of money on
 
- 
police or prisons if you can set up a system
 
- 
where people discourage one another from
 
- 
anti-social acts like political action
in exchange for
 
- 
a coupon for a free Uber ride.
 
- 
So, there are a lot of
legitimate questions here:
 
- 
What protections does
user data have in this scheme?
 
- 
Do any safeguards exist to prevent tampering?
 
- 
What mechanism, if any, is there to prevent
 
- 
false input data from creating erroneous inferences?
 
- 
Is there any way that people can fix
 
- 
their score once they're ranked poorly?
 
- 
Or does it end up becoming a
 
- 
self-fulfilling prophecy?
 
- 
Your weak score means you have less access
 
- 
to jobs and credit, and now you will have
 
- 
limited access to opportunity.
 
- 
So, let's take a step back.
 
- 
So, what do we want?
 
- 
So, we probably don't want that,
 
- 
but as advocates we really wanna
 
- 
understand what questions we should be asking
 
- 
of these systems. Right now there's
 
- 
very little oversight,
 
- 
and we wanna make sure that we don't
 
- 
sort of sleepwalk our way to a situation
 
- 
where we've lost even more power
 
- 
to these centralized systems of control.
 
- 
And if you're an implementer, we wanna understand
 
- 
what can we be doing better.
 
- 
Are there better ways that we can be implementing
 
- 
these systems?
 
- 
Are there values that, as humans,
 
- 
we care about that we should make sure
 
- 
these systems have?
 
- 
So, the first thing
that most people in the room
 
- 
might think about is privacy.
 
- 
Which is, of course, of the utmost importance.
 
- 
We need privacy, and there is a good discussion
 
- 
on the importance of protecting
user data where possible.
 
- 
So, in this talk, I'm gonna focus
on the other aspects of
 
- 
algorithmic decision making,
 
- 
that I think have got less attention.
 
- 
Because it's not just privacy
that we need to worry about here.
 
- 
We also want systems that are fair and equitable.
 
- 
We want transparent systems,
 
- 
we don't want opaque decisions
to be made about us,
 
- 
decisions that might have serious impacts
 
- 
on our lives.
 
- 
And we need some accountability mechanisms.
 
- 
So, for the rest of this talk
 
- 
we're gonna go through each one of these things
 
- 
and look at some examples.
 
- 
So, the first thing is fairness.
 
- 
And so, as I said in the beginning,
this is one area
 
- 
where there might be an advantage
 
- 
to making decisions by machine,
 
- 
especially in areas where there have
 
- 
historically been fairness issues with
 
- 
decision making, such as law enforcement.
 
- 
So, this is one way that police departments
 
- 
use predictive models.
 
- 
The idea here is police would like to
 
- 
allocate resources in a more effective way,
 
- 
and they would also like to enable
 
- 
proactive policing.
 
- 
So, if you can predict where crimes
are going to occur,
 
- 
or who is going to commit crimes,
 
- 
then you can put cops in those places,
 
- 
or perhaps following these people,
 
- 
and then the crimes will not occur.
 
- 
So, it's sort of the pre-crime approach.
 
- 
So, there are a few ways of going about this.
 
- 
One way is doing this individual-level prediction.
 
- 
So you take each citizen
and estimate the risk
 
- 
that each citizen will participate,
say, in violence
 
- 
based on some data.
 
- 
And then you can flag those people that are
 
- 
considered particularly violent.
 
- 
So, this is currently done.
 
- 
This is done in the U.S.
 
- 
It's done in Chicago,
by the Chicago Police Department.
 
- 
And they maintain a heat list of individuals
 
- 
that are considered most likely to commit,
 
- 
or be the victim of, violence.
 
- 
And this is done using data
that the police maintain.
 
- 
So, the features that are used
in this predictive model
 
- 
include things that are derived from
 
- 
individuals' criminal history.
 
- 
So, for example, have they been involved in
 
- 
gun violence in the past?
 
- 
Do they have narcotics arrests? And so on.
 
- 
But another thing that's incorporated
 
- 
in the Chicago Police Department model is
 
- 
information derived from
social media network analysis.
 
- 
So, who you interact with,
 
- 
as noted in police data.
 
- 
So, for example, your co-arrestees.
 
- 
When officers conduct field interviews,
 
- 
who are people interacting with?
 
- 
And then this is all incorporated
into this risk score.
 
- 
So another way to proceed,
 
- 
which is the method that most companies
 
- 
that sell products like this
to the police have taken,
 
- 
is instead predicting which areas
 
- 
are likely to have crimes committed in them.
 
- 
So, take my city, I put a grid down,
 
- 
and then I use crime statistics
 
- 
and maybe some ancillary data sources,
 
- 
to determine which areas have
 
- 
the highest risk of crimes occurring in them,
 
- 
and I can flag those areas and send
 
- 
police officers to them.
 
- 
So now, let's look at some of the tools
 
- 
that are used for this geographic-level prediction.
 
- 
So, here are 3 companies that sell these
 
- 
geographic-level predictive policing systems.
 
- 
So, PredPol has a system that uses
 
- 
primarily crime statistics:
 
- 
only the time, place, and type of crime
 
- 
to predict where crimes will occur.
 
- 
HunchLab uses a wider range of data sources
 
- 
including, for example, weather
 
- 
and then Hitachi is a newer system
 
- 
that has a predictive crime analytics tool
 
- 
that also incorporates social media.
 
- 
The first one, to my knowledge, to do so.
 
- 
And these systems are in use
 
- 
in 50+ cities in the U.S.
 
- 
So, why do police departments buy this?
 
- 
Some police departments are interesting in
 
- 
buying systems like this, because they're marketed
 
- 
as impartial systems,
 
- 
so it's a way to police in an unbiased way.
 
- 
And so, these companies make
 
- 
statements like this--
 
- 
by the way, the references
will all be at the end,
 
- 
and they'll be on the slides--
 
- 
So, for example
 
- 
the predictive crime analytics from Hitachi
 
- 
claims that the system is anonymous,
 
- 
because it shows you an area,
 
- 
it doesn't show you
to look for a particular person.
 
- 
and PredPol reassures people that
 
- 
it eliminates any liberties or profiling concerns.
 
- 
And HunchLab notes that the system
 
- 
fairly represents priorities for public safety
 
- 
and is unbiased by race
or ethnicity, for example.
 
- 
So, let's take a minute
to describe in more detail
 
- 
what we mean when we talk about fairness.
 
- 
So, when we talk about fairness,
 
- 
we mean a few things.
 
- 
So, one is fairness with respect to individuals:
 
- 
so if I'm very similar to somebody
 
- 
and we go through some process
 
- 
and there is two very different
outcomes to that process
 
- 
we would consider that to be unfair.
 
- 
So, we want similar people to be treated
 
- 
in a similar way.
 
- 
But, there are certain protected attributes
 
- 
that we wouldn't want someone
 
- 
to discriminate based on.
 
- 
And so, there's this other property,
Group Fairness.
 
- 
So, we can look at the statistical parity
 
- 
between groups, based on gender, race, etc.
 
- 
and see if they're treated in a similar way.
 
- 
And we might not expect that in some cases,
 
- 
for example if the base rates in each group
 
- 
are very different.
 
- 
And then there's also Fairness in Errors.
 
- 
All predictive systems are gonna make errors,
 
- 
and if the errors are concentrated,
 
- 
then that may also represent unfairness.
 
- 
And so this concern arose recently with Facebook
 
- 
because people with Native American names
 
- 
had their profiles flagged as fraudulent
 
- 
far more often than those
with White American names.
 
- 
So these are the sorts of things
that we worry about
 
- 
and each of these are metrics,
 
- 
and if you're interested more you should
 
- 
check those 2 papers out.
 
- 
So, how can potential issues
with predictive policing
 
- 
have implications for these principles?
 
- 
So, one problem is
the training data that's used.
 
- 
Some of these systems only use crime statistics,
 
- 
other systems-- all of them use crime statistics
 
- 
in some way.
 
- 
So, one problem is that crime databases
 
- 
contain only crimes that've been detected.
 
- 
Right? So, the police are only gonna detect
 
- 
crimes that they know are happening,
 
- 
either through patrol and their own investigation
 
- 
or because they've been alerted to crime,
 
- 
for example by a citizen calling the police.
 
- 
So, a citizen has to feel like
they can call the police,
 
- 
like that's a good idea.
 
- 
So, some crimes suffer
from this problem less than others:
 
- 
for example, gun violence
is much easier to detect
 
- 
relative to fraud, for example,
 
- 
which is very difficult to detect.
 
- 
Now the racial profiling aspect
of this might come in
 
- 
because of biased policing in the past.
 
- 
So, for example, for marijuana arrests,
 
- 
black people are arrested in the U.S. at rates
 
- 
4 times that of white people,
 
- 
even though there is statistical parity
 
- 
with these 2 groups, to within a few percent.
 
- 
So, this is where problems can arise.
 
- 
So, let's go back to this
 
- 
geographic-level predictive policing.
 
- 
So the danger here is that, unless this system
 
- 
is very carefully constructed,
 
- 
this sort of crime area ranking might
 
- 
again become a self-fulling prophecy.
 
- 
If you send police officers to these areas,
 
- 
you further scrutinize them,
 
- 
and then again you're only detecting a subset
 
- 
of crimes, and the cycle continues.
 
- 
So, one obvious issue is that
 
- 
this statement about geographic-based
crime prediction
 
- 
being anonymous is not true,
 
- 
because race and location are very strongly
 
- 
correlated in the U.S.
 
- 
And this is something that machine-learning
systems
 
- 
can potentially learn.
 
- 
Another issue is that, for example,
 
- 
for individual fairness, one of my homes
 
- 
sits within one of these boxes.
 
- 
Some of these boxes
in these systems are very small,
 
- 
for example PredPol is 500ft x 500ft,
 
- 
so it's maybe only a few houses.
 
- 
So, the implications of this system are that
 
- 
you have police officers maybe sitting
 
- 
in a police cruiser outside your home
 
- 
and a few doors down someone
 
- 
may not be within that box,
 
- 
and doesn't have this.
 
- 
So, that may represent unfairness.
 
- 
So, there are real questions here,
 
- 
especially because there's no opt-out.
 
- 
There's no way to opt-out of this system:
 
- 
if you live in a city that has this,
 
- 
then you have to deal with it.
 
- 
So, it's quite difficult to find out
 
- 
what's really going on
 
- 
because the algorithm is secret.
 
- 
And, in most cases, we don't know
 
- 
the full details of the inputs.
 
- 
We have some idea
about what features are used,
 
- 
but that's about it.
 
- 
We also don't know the output.
 
- 
That would be knowing police allocation,
 
- 
police strategies,
 
- 
and in order to nail down
what's really going on here
 
- 
in order to verify the validity of
 
- 
these companies' claims,
 
- 
it may be necessary
to have a 3rd party come in,
 
- 
examine the inputs and outputs of the system,
 
- 
and say concretely what's going on.
 
- 
And if everything is fine and dandy
 
- 
then this shouldn't be a problem.
 
- 
So, that's potentially one role that
 
- 
advocates can play.
 
- 
Maybe we should start pushing for audits
 
- 
of systems that are used in this way.
 
- 
These could have serious implications
 
- 
for peoples' lives.
 
- 
So, we'll return
to this idea a little bit later,
 
- 
but for now this leads us
nicely to Transparency.
 
- 
So, we wanna know
 
- 
what these systems are doing.
 
- 
But it's very hard,
for the reasons described earlier,
 
- 
but even in the case of something like
 
- 
trying to understand Google's search algorithm,
 
- 
it's difficult because it's personalized.
 
- 
So, by construction, each user is
 
- 
only seeing one endpoint.
 
- 
So, it's a very isolating system.
 
- 
What do other people see?
 
- 
And one reason it's difficult to make
 
- 
some of these systems transparent
 
- 
is because of, simply, the complexity
 
- 
of the algorithms.
 
- 
So, an algorithm can become so complex that
 
- 
it's difficult to comprehend,
 
- 
even for the designer of the system,
 
- 
or the implementer of the system.
 
- 
The designed might know that this algorithm
 
- 
maximizes some metric-- say, accuracy,
 
- 
but they may not always have a solid
 
- 
understanding of what the algorithm is doing
 
- 
for all inputs.
 
- 
Certainly with respect to fairness.
 
- 
So, in some cases,
it might not be appropriate to use
 
- 
an extremely complex model.
 
- 
It might be better to use a simpler system
 
- 
with human-interpretable features.
 
- 
Another issue that arises
 
- 
from the opacity of these systems
 
- 
and the centralized control
 
- 
is that it makes them very influential.
 
- 
And thus, an excellent target
 
- 
for manipulation or tampering.
 
- 
So, this might be tampering that is done
 
- 
from an organization that controls the system,
 
- 
or an insider at one of the organizations,
 
- 
or anyone who's able to compromise their security.
 
- 
So, this is an interesting academic work
 
- 
that looked at the possibility of
 
- 
slightly modifying search rankings
 
- 
to shift people's political views.
 
- 
So, since people are most likely to
 
- 
click on the top search results,
 
- 
so 90% of clicks go to the
first page of search results,
 
- 
then perhaps by reshuffling
things a little bit,
 
- 
or maybe dropping some search results,
 
- 
you can influence people's views
 
- 
in a coherent way,
 
- 
and maybe you can make it so subtle
 
- 
that no one is able to notice.
 
- 
So in this academic study,
 
- 
they did an experiment
 
- 
in the 2014 Indian election.
 
- 
So they used real voters,
 
- 
and they kept the size
of the experiment small enough
 
- 
that it was not going to influence the outcome
 
- 
of the election.
 
- 
So the researchers took people,
 
- 
they determined their political leaning,
 
- 
and they segmented them into
control and treatment groups,
 
- 
where the treatment was manipulation
 
- 
of the search ranking results,
 
- 
And then they had these people
browse the web.
 
- 
And what they found, is that
 
- 
this mechanism is very effective at shifting
 
- 
people's voter preferences.
 
- 
So, in this study, they were able to introduce
 
- 
a 20% shift in voter preferences.
 
- 
Even alerting users to the fact that this
 
- 
was going to be done, telling them
 
- 
"we are going to manipulate your search results,"
 
- 
"really pay attention,"
 
- 
they were totally unable to decrease
 
- 
the magnitude of the effect.
 
- 
So, the margins of error in many elections
 
- 
is incredibly small,
 
- 
and the authors estimate that this shift
 
- 
could change the outcome of about
 
- 
25% of elections worldwide, if this were done.
 
- 
And the bias is so small that no one can tell.
 
- 
So, all humans, no matter how smart
 
- 
and resistant to manipulation
we think we are,
 
- 
all of us are subject to this sort of manipulation,
 
- 
and we really can't tell.
 
- 
So, I'm not saying that this is occurring,
 
- 
but right now there is no
regulation to stop this,
 
- 
there is no way we could reliably detect this,
 
- 
so there's a huge amount of power here.
 
- 
So, something to think about.
 
- 
But it's not only corporations that are interested
 
- 
in this sort of behavioral manipulation.
 
- 
In 2010, UK Prime Minister David Cameron
 
- 
created this UK Behavioural Insights Team,
 
- 
which is informally called the Nudge Unit.
 
- 
And so what they do is
they use behavioral science
 
- 
and this predictive analytics approach,
 
- 
with experimentation,
 
- 
to have people make better decisions
 
- 
for themselves and society--
 
- 
as determined by the UK government.
 
- 
And as of a few months ago,
 
- 
after an executive order signed by Obama
 
- 
in September, the United States now has
 
- 
its own Nudge Unit.
 
- 
So, to be clear, I don't think that this is
 
- 
some sort of malicious plot.
 
- 
I think that there can be huge value
 
- 
in these sorts of initiatives,
 
- 
positively impacting people's lives,
 
- 
but when this sort of behavioral manipulation
 
- 
is being done, in part openly,
 
- 
oversight is pretty important,
 
- 
and we really need to consider
 
- 
what these systems are optimizing for.
 
- 
And that's something that we might
 
- 
not always know, or at least understand,
 
- 
so for example, for industry,
 
- 
we do have a pretty good understanding there:
 
- 
industry cares about optimizing for
 
- 
the time spent on the website,
 
- 
Facebook wants you to spend more time on Facebook,
 
- 
they want you to click on ads,
 
- 
click on newsfeed items,
 
- 
they want you to like things.
 
- 
And, fundamentally: profit.
 
- 
So, already this has some serious implications,
 
- 
and this had pretty serious implications
 
- 
in the last 10 years, in media for example.
 
- 
The optimizing for click-through rate in journalism
 
- 
has produced a race to the bottom
 
- 
in terms of quality.
 
- 
And another issue is that optimizing
 
- 
for what people like might not always be
 
- 
the best approach.
 
- 
So, Facebook officials have said publicly
 
- 
about how Facebook's goal is to make you happy,
 
- 
they want you to open that newsfeed
 
- 
and just feel great.
 
- 
But, there's an issue there, right?
 
- 
Because people get their news,
 
- 
like 40% of people according to Pew Research,
 
- 
get their news from Facebook.
 
- 
So, if people don't want to see
 
- 
war and corpses,
because it makes them feel sad,
 
- 
so this is not a system that is gonna optimize
 
- 
for an informed population.
 
- 
It's not gonna produce a population that is
 
- 
ready to engage in civic life.
 
- 
It's gonna produce an amused populations
 
- 
whose time is occupied by cat pictures.
 
- 
So, in politics, we have a similar
 
- 
optimization problem that's occurring.
 
- 
So, these political campaigns that use
 
- 
these predictive systems,
 
- 
are optimizing for votes for the desired candidate,
 
- 
of course.
 
- 
So, instead of a political campaign being
 
- 
--well, maybe this is a naive view, but--
 
- 
being an open discussion of the issues
 
- 
facing the country,
 
- 
it becomes this micro-targeted
persuasion game,
 
- 
and the people that get targeted
 
- 
are a very small subset of all people,
 
- 
and it's only gonna be people that are
 
- 
you know, on the edge, maybe disinterested,
 
- 
those are the people that are gonna get attention
 
- 
from political candidates.
 
- 
In policy, as with these Nudge Units,
 
- 
they're being used to enable
 
- 
better use of government services.
 
- 
There are some good projects that have
 
- 
come out of this:
 
- 
increasing voter registration,
 
- 
improving health outcomes,
 
- 
improving education outcomes.
 
- 
But some of these predictive systems
 
- 
that we're starting to see in government
 
- 
are optimizing for compliance,
 
- 
as is the case with predictive policing.
 
- 
So this is something that we need to
 
- 
watch carefully.
 
- 
I think this is a nice quote that
 
- 
sort of describes the problem.
 
- 
In some ways me might be narrowing
 
- 
our horizon, and the danger is that
 
- 
these tools are separating people.
 
- 
And this is particularly bad
 
- 
for political action, because political action
 
- 
requires people to have shared experience,
 
- 
and thus are able to collectively act
 
- 
to exert pressure to fix problems.
 
- 
So, finally: accountability.
 
- 
So, we need some oversight mechanisms.
 
- 
For example, in the case of errors--
 
- 
so this is particularly important for
 
- 
civil or bureaucratic systems.
 
- 
So, when an algorithm produces some decision,
 
- 
we don't always want humans to just
 
- 
defer to the machine,
 
- 
and that might represent one of the problems.
 
- 
So, there are starting to be some cases
 
- 
of computer algorithms yielding a decision,
 
- 
and then humans being unable to correct
 
- 
an obvious error.
 
- 
So there's this case in Georgia,
in the United States,
 
- 
where 2 young people went to
 
- 
the Department of Motor Vehicles,
 
- 
they're twins, and they went
 
- 
to get their driver's license.
 
- 
However, they were both flagged by
 
- 
a fraud algorithm that uses facial recognition
 
- 
to look for similar faces,
 
- 
and I guess the people that designed the system
 
- 
didn't think of the possibility of twins.
 
- 
Yeah.
So, they just left
 
- 
without their driver's licenses.
 
- 
The people in the Department of Motor Vehicles
 
- 
were unable to correct this.
 
- 
So, this is one implication--
 
- 
it's like something out of Kafka.
 
- 
But there are also cases of errors being made,
 
- 
and people not noticing until
 
- 
after actions have been taken,
 
- 
some of them very serious--
 
- 
because people simply deferred
 
- 
to the machine.
 
- 
So, this is an example from San Francisco.
 
- 
So, an ALPR-- an Automated License Plate Reader--
 
- 
is a device that uses image recognition
 
- 
to detect and read license plates,
 
- 
and usually to compare license plates
 
- 
with a known list of plates of interest.
 
- 
And, so, San Francisco uses these
 
- 
and they're mounted on police cars.
 
- 
So, in this case, San Francisco ALPR
 
- 
got a hit on a car,
 
- 
and it was the car of a 47-year-old woman,
 
- 
with no criminal history.
 
- 
And so it was a false hit
 
- 
because it was a blurry image,
 
- 
and it matched erroneously with
 
- 
one of the plates of interest
 
- 
that happened to be a stolen vehicle.
 
- 
So, they conducted a traffic stop on her,
 
- 
and they take her out of the vehicle,
 
- 
they search her and the vehicle,
 
- 
she gets a pat-down,
 
- 
and they have her kneel
 
- 
at gunpoint, in the street.
 
- 
So, how much oversight should be present
 
- 
depends on the implications of the system.
 
- 
It's certainly the case that
 
- 
for some of these decision-making systems,
 
- 
an error might not be that important,
 
- 
it could be relatively harmless,
 
- 
but in this case,
an error in this algorithmic decision
 
- 
led to this totally innocent person
 
- 
literally having a gun pointed at her.
 
- 
So, that brings us to: we need some way of
 
- 
getting some information about
 
- 
what is going on here.
 
- 
We don't wanna have to wait for these events
 
- 
before we are able to determine
 
- 
some information about the system.
 
- 
So, auditing is one option:
 
- 
to independently verify the statements
 
- 
of companies, in situations where we have
 
- 
inputs and outputs.
 
- 
So, for example, this could be done with
 
- 
Google, Facebook.
 
- 
If you have the inputs of a system,
 
- 
say you have test accounts,
 
- 
or real accounts,
 
- 
maybe you can collect
people's information together.
 
- 
So that was something that was done
 
- 
during the 2012 Obama campaign
 
- 
by ProPublica.
 
- 
People noticed that they were getting
 
- 
different emails from the Obama campaign,
 
- 
and were interested to see
 
- 
based on what factors
 
- 
the emails were changing.
 
- 
So, I think about 200 people submitted emails
 
- 
and they were able to determine some information
 
- 
about what the emails
were being varied based on.
 
- 
So there have been some successful
 
- 
attempts at this.
 
- 
So, compare inputs and then look at
 
- 
why one item was shown to one user
 
- 
and not another, and see if there's
 
- 
any statistical differences.
 
- 
So, there's some potential legal issues
 
- 
with the test accounts, so that's something
 
- 
to think about-- I'm not a lawyer.
 
- 
So, for example, if you wanna examine
 
- 
ad-targeting algorithms,
 
- 
one way to proceed is to construct
 
- 
a browsing profile, and then examine
 
- 
what ads are served back to you.
 
- 
And so this is something that
 
- 
academic researchers have looked at,
 
- 
because, at the time at least,
 
- 
you didn't need to make an account to do this.
 
- 
So, this was a study that was presented at
 
- 
Privacy Enhancing Technologies last year,
 
- 
and in this study, the researchers
 
- 
generate some browsing profiles
 
- 
that differ only by one characteristic,
 
- 
so they're basically identical in every way
 
- 
except for one thing.
 
- 
And that is denoted by Treatment 1 and 2.
 
- 
So this is a randomized, controlled trial,
 
- 
but I left out the randomization part
 
- 
for simplicity.
 
- 
So, in one study,
they applied a treatment of gender.
 
- 
So, they had the browsing profiles
 
- 
in Treatment 1 be male browsing profiles,
 
- 
and the browsing profiles in Treatment 2
be female.
 
- 
And they wanted to see: is there any difference
 
- 
in the way that ads are targeted
 
- 
if browsing profiles are effectively identical
 
- 
except for gender?
 
- 
So, it turns out that there was.
 
- 
So, a 3rd-party site was showing Google ads
 
- 
for senior executive positions
 
- 
at a rate 6 times higher to the fake men
 
- 
than for the fake women in this study.
 
- 
So, this sort of auditing is not going to
 
- 
be able to determine everything
 
- 
that algorithms are doing, but they can
 
- 
sometimes uncover interesting,
 
- 
at least statistical differences.
 
- 
So, this leads us to the fundamental issue:
 
- 
Right now, we're really not in control
 
- 
of some of these systems,
 
- 
and we really need these predictive systems
 
- 
to be controlled by us,
 
- 
in order for them not to be used
 
- 
as a system of control.
 
- 
So there are some technologies that I'd like
 
- 
to point you all to.
 
- 
We need tools in the digital commons
 
- 
that can help address some of these concerns.
 
- 
So, the first thing is that of course
 
- 
we known that minimizing the amount of
 
- 
data available can help in some contexts,
 
- 
which we can do by making systems
 
- 
that are private by design, and by default.
 
- 
Another thing is that these audit tools
 
- 
might be useful.
 
- 
And, so, these 2 nice examples in academia...
 
- 
the ad experiment that I just showed was done
 
- 
using AdFisher.
 
- 
So, these are 2 toolkits that you can use
 
- 
to start doing this sort of auditing.
 
- 
Another technology that is generally useful,
 
- 
but particularly in the case of prediction
 
- 
it's useful to maintain access to
 
- 
as many sites as possible,
 
- 
through anonymity systems like Tor,
 
- 
because it's impossible to personalize
 
- 
when everyone looks the same.
 
- 
So this is a very important technology.
 
- 
Something that doesn't really exist,
 
- 
but that I think is pretty important,
 
- 
is having some tool to view the landscape.
 
- 
So, as we know from these few studies
 
- 
that have been done,
 
- 
different people are not seeing the internet
 
- 
in the same way.
 
- 
This is one reason why we don't like censorship.
 
- 
But, rich and poor people,
 
- 
from academic research we know that
 
- 
there is widespread price discrimination
on the internet,
 
- 
so rich and poor people see a different view
 
- 
of the Internet,
 
- 
men and women see a different view
 
- 
of the Internet.
 
- 
We wanna know how different people
 
- 
see the same site,
 
- 
and this could be the beginning of
 
- 
a defense system for this sort of
 
- 
manipulation/tampering that I showed earlier.
 
- 
Another interesting approach is obfuscation:
 
- 
injecting noise into the system.
 
- 
So there's an interesting browser extension
 
- 
called Adnauseum, that's for Firefox,
 
- 
which clicks on every single ad you're served,
 
- 
to inject noise.
 
- 
So that's, I think, an interesting approach
 
- 
that people haven't looked at too much.
 
- 
So in terms of policy,
 
- 
Facebook and Google, these internet giants,
 
- 
have billions of users,
 
- 
and sometimes they like to call themselves
 
- 
new public utilities,
 
- 
and if that's the case then
 
- 
it might be necessary to subject them
 
- 
to additional regulation.
 
- 
Another problem that's come up,
 
- 
for example with some of the studies
 
- 
that Facebook has done,
 
- 
is sometimes a lack of ethics review.
 
- 
So, for example, in academia,
 
- 
if you're gonna do research involving humans,
 
- 
there's an Institutional Review Board
 
- 
that you go to that verifies that
 
- 
you're doing things in an ethical manner.
 
- 
And some companies do have internal
 
- 
review processes like this, but it might
 
- 
be important to have an independent
 
- 
ethics board that does this sort of thing.
 
- 
And we really need 3rd-party auditing.
 
- 
So, for example, some companies
 
- 
don't want auditing to be done
 
- 
because of IP concerns,
 
- 
and if that's the concern
 
- 
maybe having a set of people
 
- 
that are not paid by the company
 
- 
to check how some of these systems
 
- 
are being implemented,
 
- 
could help give us confidence that
 
- 
things are being done in a reasonable way.
 
- 
So, in closing,
 
- 
algorithmic decision making is here,
 
- 
and it's barreling forward
at a very fast rate,
 
- 
and we need to figure out what
 
- 
the guide rails should be,
 
- 
and how to install them
 
- 
to handle some of the potential threats.
 
- 
There's a huge amount of power here.
 
- 
We need more openness in these systems.
 
- 
And, right now,
 
- 
with the intelligent systems that do exist,
 
- 
we don't know what's occurring really,
 
- 
and we need to watch carefully
 
- 
where and how these systems are being used.
 
- 
And I think this community has
 
- 
an important role to play in this fight,
 
- 
to study what's being done,
 
- 
to show people what's being done,
 
- 
to raise the debate and advocate,
 
- 
and, where necessary, to resist.
 
- 
Thanks.
 
- 
applause
 
- 
Herald: So, let's have a question and answer.
 
- 
Microphone 2, please.
 
- 
Mic 2: Hi there.
 
- 
Thanks for the talk.
 
- 
Since these pre-crime softwares also
 
- 
arrived here in Germany
 
- 
with the start of the so-called CopWatch system
 
- 
in southern Germany,
and Bavaria and Nuremberg especially,
 
- 
where they try to predict burglary crime
 
- 
using that criminal record
 
- 
geographical analysis, like you explained,
 
- 
leads me to a 2-fold question:
 
- 
first, have you heard of any research
 
- 
that measures the effectiveness
 
- 
of such measures, at all?
 
- 
And, second:
 
- 
What do you think of the game theory
 
- 
if the thieves or the bad guys
 
- 
know the system, and when they
game the system,
 
- 
they will probably win,
 
- 
since one police officer in an interview said
 
- 
this system is used to reduce
 
- 
the personal costs of policing,
 
- 
so they just send the guys
where the red flags are,
 
- 
and the others take the day off.
 
- 
Dr. Helsby: Yup.
 
- 
Um, so, with respect to
 
- 
testing the effectiveness of predictive policing,
 
- 
the companies,
 
- 
some of them do randomized, controlled trials
 
- 
and claim a reduction in policing.
 
- 
The best independent study that I've seen
 
- 
is by this RAND Corporation
 
- 
that did a study in, I think,
 
- 
Shreveport, Louisiana,
 
- 
and in their report they claim
 
- 
that there was no statistically significant
 
- 
difference, they didn't find any reduction.
 
- 
And it was specifically looking at
 
- 
property crime, which I think you mentioned.
 
- 
So, I think right now there's sort of
 
- 
conflicting reports between
 
- 
the independent auditors
and these company claims.
 
- 
So there definitely needs to be more study.
 
- 
And then, the 2nd thing...sorry,
remind me what it was?
 
- 
Mic 2: What about the guys gaming the system?
 
- 
Dr. Helsby: Oh, yeah.
 
- 
I think it's a legitimate concern.
 
- 
Like, if all the outputs
were just immediately public,
 
- 
then, yes, everyone knows the location
 
- 
of all police officers,
 
- 
and I imagine that people would have
 
- 
a problem with that.
 
- 
Yup.
 
- 
Heraldl: Microphone #4, please.
 
- 
Mic 4: Yeah, this is not actually a question,
 
- 
but just a comment.
 
- 
I've enjoyed your talk very much,
 
- 
in particular after watching
 
- 
the talk in Hall 1 earlier in the afternoon.
 
- 
The "Say Hi to Your New Boss", about
 
- 
algorithms that are trained with big data,
 
- 
and finally make decisions.
 
- 
And I think these 2 talks are kind of complementary,
 
- 
and if people are interested in the topic
 
- 
they might want to check out the other talk
 
- 
and watch it later, because these
 
- 
fit very well together.
 
- 
Dr. Helsby: Yeah, it was a great talk.
 
- 
Herald: Microphone #2, please.
 
- 
Mic 2: Um, yeah, you mentioned
 
- 
the need to have some kind of 3rd-party auditing
 
- 
or some kind of way to
 
- 
peek into these algorithms
 
- 
and to see what they're doing,
 
- 
and to see if they're being fair.
 
- 
Can you talk a little bit more about that?
 
- 
Like, going forward,
 
- 
some kind of regulatory structures
 
- 
would probably have to emerge
 
- 
to analyze and to look at
 
- 
these black boxes that are just sort of
 
- 
popping up everywhere and, you know,
 
- 
controlling more and more of the things
 
- 
in our lives, and important decisions.
 
- 
So, just, what kind of discussions
 
- 
are there for that?
 
- 
And what kind of possibility
is there for that?
 
- 
And, I'm sure that companies would be
 
- 
very, very resistant to
 
- 
any kind of attempt to look into
 
- 
algorithms, and to...
 
- 
Dr. Helsby: Yeah, I mean, definitely
 
- 
companies would be very resistant to
 
- 
having people look into their algorithms.
 
- 
So, if you wanna do a very rigorous
 
- 
audit of what's going on
 
- 
then it's probably necessary to have
 
- 
a few people come in
 
- 
and sign NDAs, and then
 
- 
look through the systems.
 
- 
So, that's one way to proceed.
 
- 
But, another way to proceed that--
 
- 
so, these academic researchers have done
 
- 
a few experiments
 
- 
and found some interesting things,
 
- 
and that's sort all the attempts at auditing
 
- 
that we've seen:
 
- 
there was 1 attempt in 2012
for the Obama campaign,
 
- 
but there's really not been any
 
- 
sort of systematic attempt--
 
- 
you know, like, in censorship
 
- 
we see a systematic attempt to
 
- 
do measurement as often as possible,
 
- 
check what's going on,
 
- 
and that itself, you know,
 
- 
can act as an oversight mechanism.
 
- 
But, right now,
 
- 
I think many of these companies
 
- 
realize no one is watching,
 
- 
so there's no real push to have
 
- 
people verify: are you being fair when you
 
- 
implement this system?
 
- 
Because no one's really checking.
 
- 
Mic 2: Do you think that,
 
- 
at some point, it would be like
 
- 
an FDA or SEC, to give some American examples...
 
- 
an actual government regulatory agency
 
- 
that has the power and ability to
 
- 
not just sort of look and try to
 
- 
reverse engineer some of these algorithms,
 
- 
but actually peek in there and make sure
 
- 
that things are fair, because it seems like
 
- 
there's just-- it's so important now
 
- 
that, again, it could be the difference between
 
- 
life and death, between
 
- 
getting a job, not getting a job,
 
- 
being pulled over,
not being pulled over,
 
- 
being racially profiled,
not racially profiled,
 
- 
things like that.
Dr. Helsby: Right.
 
- 
Mic 2: Is it moving in that direction?
 
- 
Or is it way too early for it?
 
- 
Dr. Helsby: I mean, so some people have...
 
- 
someone has called for, like,
 
- 
a Federal Search Commission,
 
- 
or like a Federal Algorithms Commission,
 
- 
that would do this sort of oversight work,
 
- 
but it's in such early stages right now
 
- 
that there's no real push for that.
 
- 
But I think it's a good idea.
 
- 
Herald: And again, #2 please.
 
- 
Mic 2: Thank you again for your talk.
 
- 
I was just curious if you can point
 
- 
to any examples of
 
- 
either current producers or consumers
 
- 
of these algorithmic systems
 
- 
who are actively and publicly trying
 
- 
to do so in a responsible manner
 
- 
by describing what they're trying to do
 
- 
and how they're going about it?
 
- 
Dr. Helsby: So, yeah, there are some companies,
 
- 
for example, like DataKind,
 
- 
that try to deploy algorithmic systems
 
- 
in as responsible a way as possible,
 
- 
for like public policy.
 
- 
Like, I actually also implement systems
 
- 
for public policy in a transparent way.
 
- 
Like, all the code is in GitHub, etc.
 
- 
And it is also the case to give credit to
 
- 
Google, and these giants,
 
- 
they're trying to implement transparency systems
 
- 
that help you understand.
 
- 
This has been done with respect to
 
- 
how your data is being collected,
 
- 
but for example if you go on Amazon.com
 
- 
you can see a recommendation has been made,
 
- 
and that is pretty transparent.
 
- 
You can see "this item
was recommended to me,"
 
- 
so you know that prediction
is being used in this case,
 
- 
and it will say why prediction is being used:
 
- 
because you purchased some item.
 
- 
And Google has a similar thing,
 
- 
if you go to like Google Ad Settings,
 
- 
you can even turn off personalization of ads
 
- 
if you want,
 
- 
and you can also see some of the inferences
 
- 
that have been learned about you.
 
- 
A subset of the inferences that have been
 
- 
learned about you.
 
- 
So, like, what interests...
 
- 
Herald: A question from the internet, please?
 
- 
Signal Angel: Yes, billetQ is asking
 
- 
how do you avoid biases in machine learning?
 
- 
I asume analysis system, for example,
 
- 
could be biased against women and minorities,
 
- 
if used for hiring decisions
based on known data.
 
- 
Dr. Helsby: Yeah, so one thing is to
 
- 
just explicitly check.
 
- 
So, you can check to see how
 
- 
positive outcomes are being distributed
 
- 
among those protected classes.
 
- 
You could also incorporate these sort of
 
- 
fairness constraints in the function
 
- 
that you optimize when you train the system,
 
- 
and so, if you're interested in reading more
 
- 
about this, the 2 papers--
 
- 
let me go to References--
 
- 
there's a good paper called
 
- 
Fairness Through Awareness that describes
 
- 
how to go about doing this,
 
- 
so I recommend this person read that.
 
- 
It's good.
 
- 
Herald: Microphone 2, please.
 
- 
Mic2: Thanks again for your talk.
 
- 
Umm, hello?
 
- 
Okay.
 
- 
Umm, I see of course a problem with
 
- 
all the black boxes that you describe
 
- 
with regards for the crime systems,
 
- 
but when we look at the advertising systems
 
- 
in many cases they are very networked.
 
- 
There are many different systems collaborating
 
- 
and exchanging data via open APIs:
 
- 
RESTful APIs, and various
 
- 
demand-side platforms
and audience-exchange platforms,
 
- 
and everything.
 
- 
So, can that help to at least
 
- 
increase awareness on where targeting, personalization
 
- 
might be happening?
 
- 
I mean, I'm looking at systems like
 
- 
BuiltWith, that surface what kind of
 
- 
JavaScript libraries are used elsewhere.
 
- 
So, is that something that could help
 
- 
at least to give a better awareness
 
- 
and listing all the points where
 
- 
you might be targeted...
 
- 
Dr. Helsby: So, like, with respect to
 
- 
advertising, the fact that
there is behind the scenes
 
- 
this like complicated auction process
 
- 
that's occurring, just makes things
 
- 
a lot more complicated.
 
- 
So, for example, I said briefly
 
- 
that they found that there's this
statistical difference
 
- 
between how men and women are treated,
 
- 
but it doesn't necessarily mean that
 
- 
"Oh, the algorithm is definitely biased."
 
- 
It could be because of this auction process,
 
- 
it could be that women are considered
 
- 
more valuable when it comes to advertising,
 
- 
and so these executive ads are getting
 
- 
outbid by some other ads,
 
- 
and so there's a lot of potential
 
- 
causes for that.
 
- 
So, I think it just makes things
a lot more complicated.
 
- 
I don't know if it helps
with the bias at all.
 
- 
Mic 2: Well, the question was more
 
- 
a direction... can it help to surface
 
- 
and make people aware of that fact?
 
- 
I mean, I can talk to my kids probably,
 
- 
and they will probably understand,
 
- 
but I can't explain that to my grandma,
 
- 
who's also, umm, looking at an iPad.
 
- 
Dr. Helsby: So, the fact that
 
- 
the systems are...
 
- 
I don't know if I understand.
 
- 
Mic 2: OK. I think that the main problem
 
- 
is that we are behind the industry efforts
 
- 
to being targeted at, and many people
 
- 
do know, but a lot more people don't know,
 
- 
and making them aware of the fact
 
- 
that they are a target, in a way,
 
- 
is something that can only be shown
 
- 
by a 3rd party that disposed that data,
 
- 
and make audits in a way--
 
- 
maybe in an automated way.
 
- 
Dr. Helsby: Right.
 
- 
Yeah, I think it certainly
could help with advocacy
 
- 
if that's the point, yeah.
 
- 
Herald: Another question
from the internet, please.
 
- 
Signal Angel: Yes, on IRC they are asking
 
- 
if we know that prediction in some cases
 
- 
provides an influence that cannot be controlled.
 
- 
So, r4v5 would like to know from you
 
- 
if there are some cases or areas where
 
- 
machine learning simply shouldn't go?
 
- 
Dr. Helsby: Umm, so I think...
 
- 
I mean, yes, I think that it is the case
 
- 
that in some cases machine learning
 
- 
might not be appropriate.
 
- 
For example, if you use machine learning
 
- 
to decide who should be searched.
 
- 
I don't think it should be the case that
 
- 
machine learning algorithms should
 
- 
ever be used to determine
 
- 
probable cause, or something like that.
 
- 
So, if it's just one piece of evidence
 
- 
that you consider,
 
- 
and there's human oversight always,
 
- 
maybe it's fine, but
 
- 
we should be very suspicious and hesitant
 
- 
in certain contexts where
 
- 
the ramifications are very serious.
 
- 
Like the No Fly List, and so on.
 
- 
Herald: And #2 again.
 
- 
Mic 2: A second question
 
- 
that just occurred to me, if you don't mind.
 
- 
Umm, until the advent of
 
- 
algorithmic systems,
 
- 
when there've been cases of serious harm
 
- 
that's been resulted in individuals or groups,
 
- 
and it's been demonstrated that
 
- 
it's occurred because of
 
- 
an individual or a system of people
 
- 
being systematically biased, then often
 
- 
one of the actions that's taken is
 
- 
pressure's applied, and then
 
- 
people are required to change,
 
- 
and hopely be held responsible,
 
- 
and then change the way that they do things
 
- 
to try to remove bias from that system.
 
- 
What's the current thinking about
 
- 
how we can go about doing that
 
- 
when the systems that are doing that
 
- 
are algorithmic?
 
- 
Is it just going to be human oversight,
 
- 
and humans are gonna have to be
 
- 
held responsible for the oversight?
 
- 
Dr. Helsby: So, in terms of bias,
 
- 
if we're concerned about bias towards
 
- 
particular types of people,
 
- 
that's something that we can optimize for.
 
- 
So, we can train systems that are unbiased
 
- 
in this way.
 
- 
So that's one way to deal with it.
 
- 
But there's always gonna be errors,
 
- 
so that's sort of a separate issue
 
- 
from the bias, and in the case
 
- 
where there are errors,
 
- 
there must be oversight.
 
- 
So, one way that one could improve
 
- 
the way that this is done
 
- 
is by making sure that you're
 
- 
keeping track of confidence of decisions.
 
- 
So, if you have a low confidence prediction,
 
- 
then maybe a human
should come in and check things.
 
- 
So, that might be one way to proceed.
 
- 
Herald: So, there's no more question.
 
- 
I close this talk now,
 
- 
and thank you very much
 
- 
and a big applause to
 
- 
Jennifer Helsby!
 
- 
roaring applause
 
- 
subtitles created by c3subtitles.de
Join, and help us!