-
rC3 postroll music
-
Herald: Now, imagine a stage with an
artist performing in front of a crowd.
-
Is there a way to measure and even quantify
the shows impact on the spectators?
-
Kai Kunze is going to address this
question in his talk Boiling Minds now.
-
Kai, up to you.
-
Kai: Thanks a lot for the introduction,
but we have a short video. I hope
-
that can be played right now.
-
intense electronic staccato music
-
music shifts to include softer piano tones
-
music shifts again to include harp-like tones
-
music keeps gently shifting
-
longer drawn out, slowly decreasing pitch
-
shift towards slow, guitar-like sounds
-
with light crackling noises
-
music getting quieter, softer
-
and fades away
-
inaudible talking
-
Kai: So thanks a lot for the intro and
this is the Boiling Mind talks or linking
-
physiology and choreography. I just started
off with this short video, that could
-
give you an overview over the experience
of this dance performance that we
-
staged in Tokyo beginning of the year,
just before the lockdown, actually.
-
And the idea behind this was: we wanted to
put the audience on a stage. So breaking the
-
fourth wall. Trying to use physiological
sensing in the audience. And that change
-
then is reflected on stage over the
projection, sound and also audio to
-
influence the dancers and performers and
then, of course, feed them back again to
-
the audience. So creating an augmented
feedback loop. In his talk today, I just
-
want to give you a small overview, a
little bit about the motivation, why I
-
thought it's a nice topic for the remote
experience from the Chaos Computer Club
-
and also a little bit more about the
concept, the set up and the design
-
iterations, as well as the lessons
learned. So for me to give this talk,
-
I thought it's a good way to exchange
expertise and get a couple of people that
-
might be interested for the next
iterations, because I think we are still
-
not done with this work. So it's still
kind of work in progress. And also a way
-
to share data. So to do some explorative
data analysis on the recorded performances
-
that we have. And then most important: I
wanted to create a more creative way to
-
use physiological data and explore it,
because also for me as a researcher
-
working on variable computing or activity
recognition, often we just look into
-
recognizing or predicting certain motions
or certain mental states.
-
And that kind of, at least for simple things,
feeds back into this very - I think -
-
idiotic or stupid ideas of surveillance and
applications cases and that.
-
So can we create more intuitive ways
to use physiological data?
-
So from a concept perspective, I think the
-
video gave a good overview of what we
tried to create. However,
-
what we did in 3 performances was: We used
physiological sensors on all audience
-
members. So for us, it was important that
we are not singling out individual people
-
to just get feedback from them, but have
the whole response, the whole physiological
-
state of the audience as an input to the
performance. In that case, we actually
-
used heart rate variability and also
galvanic skin response as inputs.
-
And these inputs then changed the projection
that you could see. The lights, especially
-
the intensity of the lights and also
the sound. And that, again, then led to
-
changes in the dancing behavior of the
performers.
-
For the sensing, we went with a variable
set up,
-
so in this case a fully wireless wristband,
because we wanted to do something that is
-
easy to wear and easy to put on and to put
off. We had a couple of iterations on that
-
and we decided then for electrodermal
activity and also heart activity
-
to sense, because there's some
related work that link these sensors to
-
engagement stress and also excitement
measures. And the question then was also
-
where to sense it first. We went with a
couple of wrist bands and also kind of
-
commercial approaches or half-commercial
approaches. However, the sensing quality
-
was just not good enough, especially from
the wrist. You cannot really get a good
-
electrodermal activity, so galvanic skin
response. It's more or less a sweat
-
sensor. So that means that you can detect
if somebody is sweating and some of the
-
sweat is actually then related to a stress
response. And in that case, there are a
-
couple of ways to measure that. So it
could be on the lower part of your hand or
-
also on the fingers. These are usually the
best positions. So we used the fingers.
-
Over the fingers we can also get heartrate
activity. And in addition to that, there's
-
also a small motion sensor, so a gyro and an
accelerometer in the wristband. We haven't
-
used that for the performance right now, but
we still have the recordings also from the
-
audience for that. When I say we, I mean
George especially and also Dingding,
-
2 researchers that work with me, did
actually took care of the designs.
-
So then the question was also how to
map it to the environment or the staging.
-
In this case, actually, this was done
by a different team,
-
this was done by the embodied media team
also at KMD.
-
So I know a little bit about it,
but I'm definitely not an expert.
-
And for the initial design we
thought we use the EDA for the movement
-
speed of the projection. So the EDA rate
of change is matched to movement of these
-
blobs that you could see or also the meshs
that you can see and the color represents
-
the heart rate. We went for the LFHF
feature that's low frequency, high
-
frequency ratio and should give you,
according to related work, some indication
-
about excitement. For the lights: the
lights were also bound to the heart rate,
-
in this case, the beats per minute, and
they were matched to intensity. So if the
-
beats per minute of the audience go
collectively up, the light gets brighter,
-
otherwise, it's dimmer. For the audio: we
had an audio designer that cared about
-
sounds and faded in and faded out specific
sounds also related to the EDA to the
-
relative rate of change of the electro-
dermal activity. All this happened while
-
the sensors were connected over sensing
server in QT to touch designer software
-
that generated this type of projections.
And also the music got fed into and that
-
was then controlling the feedback
to the dancers. If you want to
-
have a bit more of detail, I uploaded the
work in progress preprint paper, a draft
-
of an accepted TI paper. So in case you are
interested in the mappings and the design
-
decisions for the projections, there is
a little bit more information there.
-
I'm also happy later on to answer those
questions. However, I will probably just
-
forward them to the designers, that worked
on them. And then, for the overall
-
performance, what happened was, we started
out with an explanation of the experience.
-
So it was already advertised as a performance
that would take in electrodermal
-
activity and heartbeat activity.
So, people that bought tickets or came to
-
the event already had a little bit of
background information. We, of course,
-
made also sure that we explained at the
beginning what type of sensing we will be
-
using. Also what the risks and problems
with these type of sensors and data
-
collection is and then audience could decide,
with informed consent, if they just want to
-
stream the data, don't want to do
anything, or they want to stream and also
-
contribute the data anonymously to our
research. And then when the performance
-
started, we had a couple of pieces and
parts, that is something that you can see in
-
B, where we showed the live data feed from
all of the audiences in individual tiles. We
-
had that in before just for debugging, but
actually the audience liked that. And so
-
we made it a part of the performance, also
deciding with the choreographers to
-
include that. And then for the rest, as
you see in C, we have the individual
-
objects, these blob objects that move
according to the EDA data and change colour
-
based on the heart rate information. So
the low to high frequency. In B, you see
-
also these clouds. And yet similarly, the
size is related to the heart rate data.
-
And the movement again is EDA. And there's
also one scene in E where the dancers pick
-
one person in the audience and ask them to
come on stage. And then we will display
-
that audience members data at large in the
back of the projection. And for the rest,
-
again, we're using this excitement data
from the heart rate and from the
-
electrodermal activity to change sizes and
colours. So, to come up with this design, we
-
went the co-design route, discussing with
the researchers, dancers, visual
-
designers, audio designers a couple of
times. And actually that's also how I got
-
involved first, because the initial idea is
also from Moe, the primary designer of this
-
piece, were to combine somehow perception
and motion. And I worked a bit in research
-
with the eye tracking. So you see on the
screen the pupil website eye tracker it is
-
and open source eye tracking solution and
also EOG electro-oculography glasses, that
-
use the capacitance of your eyeballs to
detect something. Rough about eye emotion.
-
And we thought at the beginning, we want
to combine this, a person seeing the play
-
with the motions of the dancers and
understand that better. So that's kind of
-
how it started. The second inspiration for
this idea in the theatre came from a
-
visiting scholar, Jamie. Jamie Ward came
over and his work with the flood theater
-
in London. That's an inclusive theatre
that also does workshops or Shakespeare
-
workshops. And he did some sensing just
with the accelerometers and gyroscopes or
-
inertial motion wristbands to detect
interpersonal synchrony between
-
participants in these workshops. And then
we thought, when he came over, we had a
-
small piece where we looked into this
interpersonal synchrony again in face to
-
face communications. I mean, now we are
remote and I'm just talking into a camera
-
and I cannot see anybody. But usually, if
you would have a face to face conversation,
-
doesn't happen too often anymore,
unfortunately. We would show some type of
-
synchronies or, you know, kind of eyeblink,
head nod and so on would synchronize with
-
the other person, if you're talking to
them. And we also showed, that in small
-
recordings also we showed that we
can recognize this in a variable sensing
-
setup. So again, using some glasses and we
thought, why don't we try to scale that
-
up? Why don't we try and see what happens
in a theatre performance or in another
-
dance performance and see if we can
recognize also some type of synchrony. And
-
with a couple of ideation sessions, a
couple of also test performances, also
-
including dancers trying out glasses,
trying out other headwear. And that was
-
not really possible to use for the dancers
during the performance. We came up with an
-
initial prototype and that we tried out,
so in, I think November 2018 or so, where
-
we used a couple of pupil-labs and also
pupil-invisible. These are nicer eye tracking
-
glasses, they are optical eye tracking
glasses, so they have small cameras in
-
them, distributed in the audience. A couple
of those Yoji glasses, they have also
-
initial motion sensors in them. So
accelerometer and gyroscope. And we had at
-
the time heart rate sensors. However, they
were fixed and wired to the system. And
-
also the dancers wore some wristbands
where we could record the motion data. And
-
then what we did in these cases, then we
had projections on three frames on top
-
of the dancers. One was showing the blink
and the headnod synchronization of the
-
audience. The other one showed heart rate
and variability. And the third one just
-
showed raw feed from one of the eye
trackers. And it looked more or less like
-
this. And from a technical perspective, we
were surprised because it actually worked.
-
So we could stream around 10 glasses,
three eye trackers and four, five, I think
-
heart rate sensors at the same time and the surfer
worked. However, from an audience
-
perspective, a lot of the feedback was the
audience didn't like that just some people
-
got singled out and got the device by
themselves and others could not really
-
contribute and could not also see the
data. And then also from a performance
-
perspective, the dancers didn't really
like that they couldn't interact with the
-
data. The dance piece also in this case
was pre-choreographed. So there was no
-
possibility for the dancers to really
interact with the data. And then also,
-
again, from an esthetic perspective, we
really didn't like that the screens were
-
on top because either you would
concentrate on the screens or you would
-
concentrate on the dance performance. And
you had to kind of make a decision also on
-
what type of visualization you would focus
on. So overall, you know, kind of partly
-
okay, but still there were some troubles.
So one was definitely we wanted to include
-
all of the audience. Meaning we wanted to
have everybody participate. Then the
-
problem with that part was then also
having enough eye trackers or having
-
enough head worn devices was an issue. In
addition to that, you know, kind of, if
-
it's head worn some people might not like
it. The pandemic hadn't started yet. When
-
we did the recordings, however, there was
already the information, some information
-
about the virus going around. So we didn't
really want as, putting everybody,
-
giving everybody some eyeglasses. So then
we moved to the heart rate and, galvanic
-
skin response solution and the set up
where the projection is now part of the
-
stage. So we used the two walls, but we
also used, it's a little bit hard to see
-
in the images, but we also used the floor
as another projection surface for the
-
dancers to interact with and the main
interaction, actually came then over the
-
sound. So then moving over to the lessons
learned. So what did we take away from
-
from that experience? And the first part
was talking with the dancers and talking
-
with the audience often, if you saw,
especially the more intricate, the more
-
abstract visualizations, it was sometimes
hard to interpret also how their own data
-
would feed into that visualization. So,
you know, kind of some audience members
-
mentioned to some point in time they were
not sure if they're influencing anything
-
or if it had an effect on other parts,
especially if they saw the live data. It
-
was kind of obvious. But for future work,
we really want to play more with the
-
agency and also perceived agency of the
audiences and the performers. And we also
-
really wonder how can we measure this type
of feedback loops? Because now we have
-
these recordings. We looked also a little
bit more into the data, but it's hard to
-
understand. Were we successful? I think in
some extent maybe yes, because the
-
experience was fun. It was enjoyable. But
on this level of, did we really create
-
feedback loops and how do you evaluate
feedback loops, that's something that we
-
want to address in future work. On the
other hand, what was surprising I
-
mentioned before the raw data was
something that the dancers as well as the
-
audience really liked. And that was
surprising for me because I thought we had
-
to hide that more or less. But we had it
on. As I said, there's kind of a debug at
-
the beginning of some test screenings and
audience members were interested in it and
-
could see and were talking about: "Oh, see
your heart rate is going up or your EDA is
-
going up." And the dancers also like that.
And we used that then in the performance
-
in the three performances that we then
successfully made for especially scenes
-
where the dancers would interact directly
with parts of the audience. At the
-
beginning of the play is a scene where the
dancers give out business cards to some
-
audience members. And it was fun to see
that some audience members could identify
-
themselves, other audience members would
identify somebody else that was sitting
-
next to them. And then this member had a
spike in EDA because of the surprise. So
-
there was really, you know, kind of some
interaction going on. So maybe staying if
-
you're planning to do a similar event,
staying close to the raw data and also low
-
latency. So I think it's quite important
for some types of these interactions. From
-
the dancers there was a big interest, on
the one side, they wanted to use the data
-
for reflection. So they really liked that
they had the printouts of the effects of
-
the audience later on. However, they also
wanted to dance more with biometric data
-
and also use that for their rehearsals
more. So, of course, you know, we had to
-
co-design, so we worked directly. We
showed the dancers the sensors and the
-
possibilities and then worked with them to
figure out what can work and what cannot
-
work and what might have an effect, what
might not have an effect. And then we did
-
some, as you saw, also some prototype
screenings and also some internal
-
rehearsals where we used some recorded
data. We used some, a couple of people of
-
us were sitting in the audience. We got a
couple of other researchers and also
-
students involved to sit in the audience
to stream data. And we also worked a
-
little bit with prerecorded experiences
and also synthetic experiences, how we
-
envisioned that the data would move. But
still, it was not enough in terms of
-
providing an intuitive way to understand
what is going on, especially also for the
-
visualizations and the projections. They
were harder to interpret than the sound in
-
the sound sphere. So and then the next and
the biggest point, maybe as well is, the
-
sensors and the feature best practices. So
we're still wondering, you know, what to
-
use. We're still searching. What kind of,
sensing equipment can we use to relay
-
this, in this invisible link between
audience and performers? How can we
-
augment that? We started out with the
perception and eye tracking part, we then
-
went to wrist one device because it's
easier to maintain and it's also wireless.
-
And it worked quite well to stream 50 to
60 audience members for one of those
-
events to a wireless router and do the
recording, as well as the life
-
visualization with it. However, the
features might have not been.
-
Audio Failure
-
Okay. Sorry for the short part where it was
offline. So, we were talking about a sense
-
of features and best practices. So in this
case, we are still searching for the right
-
type of sensors and features to use for
this type of audience, performer
-
interaction. And we were using, the, yeah,
the low frequency, high frequency ratio of
-
the heart rate values and also the
relative changes of the EDA. And that was
-
working, I would say not that well,
compared to other features that we now
-
found while looking into the performances
and the recorded data of the around, 98
-
participants that agreed to share the data
with us, for these performances. And from
-
the preliminary analysis that Karen Han,
one of our researchers working on and
-
looking into what type of features are
indicative of changes in the performance.
-
It seems that a feature called PNN that's
related to heart rate variability to the
-
R-R intervals is, seems to be quite good. And
also the peak detection per minute using
-
the EDA data. So we're just counting the
relative changes, the relative up and
-
down, for the EDA. If you're interested
I'm happy to share the data with you. So
-
we have three performances each
around an hour and 98 participants in
-
total. And we have the heart rate data,
the EDA data, from the two fingers as well
-
as, the motion data as well. We haven't
used the motion data at all except for
-
filtering out a little bit the EDA and
heart rate data because if you're moving a
-
lot, you will have some errors and some
problems, some motion artifacts in it. But
-
what do I mean with why is the PNN or why
is the EDA peak detection so nice? Let's
-
look a little bit closer into the data.
And here you see I just highlighted
-
performance three from the previous plots.
You see PNN50 on the left side is the scale, the
-
blue line gives you the average of the
PNN50 value. So this is the R-R interval
-
related heart rate variability feature and
that feature is especially related to
-
relaxation and also to stress. So usually
a lower PNN50 value means you're more
-
relaxed and a higher value means that
you're. No, higher value means that you're
-
more relaxed, sorry. Lower value means
that you are more stressed out. So what happens
-
now in the performance is something that
fits very, very well and correlates with
-
the intention of the choreographer.
Because the first half of the performance,
-
you see section one, two, three, four,
five and six on the bottom. The first half
-
of the performance is to create a conflict
in the audience and to stir them up a
-
little. So, for example, also the business
card scene is part of that part, or also
-
the scene where somebody gets brought from
the audience to the stage and joins the
-
performance is also part of that versus
the latter part is more about reflection
-
and also relaxation. So taking in what you
experienced at the first part, and that's
-
something that you see actually quite nice
in the PNN50. So at the beginning it's
-
rather low. That means the audience is
slightly tense versus in the latter part
-
they more relaxed. Similarly, the EDA in
the bottom as a bar chart gives you an
-
indication of a lot of peaks happening at
specific points. And these points
-
correlate very well to memorable scenes in
the performance. So seeing the one scene,
-
where, actually section four, the red one,
is the one where somebody from the
-
audience gets brought onto the stage.
Where is this? I think around minute
-
twelve there is a scene where the dancers
handout business cards. And that's
-
also something, I think. So it's
promising, we're not there yet definitely
-
from the data analysis part, but there are
some interesting things to see. And that
-
kind of brings me back to the starting
point. So I think, it was an amazing
-
experience actually, working with a lot of
talented people on that and the
-
performance was a lot of fun, but we are
slowly moving towards putting the audience
-
on stage and trying to break the fourth
wall, I think, with these type of setups.
-
And that leads me then also to the end of
the talk where I just have to do a shout
-
out for the people who did the actual
work. So all of the talented performers
-
and the project lead, especially Moe who
organized and was also the link between
-
the artistic side and the dancers with
Mademoiselle Cinema and us, as well as the
-
choreographer Ito-san. And yeah, I hope I
didn't miss anybody. So that's it. So
-
thanks a lot for this opportunity to
introduce this work to you. And now I'm
-
open for a couple of questions, remarks. I
wanted to also host a self organized
-
session sometime. I haven't really gotten
the link or anything, but I'll probably
-
just post something on Twitter or in one
of the chats if you want to stay in
-
contact. I'll try to get two or three
researchers also to join. I know George,
-
who was working on the hardware, and
Karen, who worked on the visualizations,
-
the data analysis might be available. And
if you interested in that, just send me an
-
email or check, maybe, I just also add it
to the blog post or so if I get the link
-
later. So, yeah. Thanks a
lot for the attention.
-
Herald: Thanks, Kai, for this nice talk.
For the audience, please excuse us for the
-
small disruption of service we had here.
We're a little bit late already, but I
-
think we still have time for a question or
so. Unfortunately, I don't see anything
-
here online at the moment. So if
somebody tried to pose a question and
-
there was also disruption of service, I
apologize beforehand for that. On the
-
other hand now, Kai, you talked about data
sharing. So how can the data be accessed?
-
Do people need to access you or drop to
you a mail or personal message?
-
Kai: Yeah, we're on the,
so right now, no publication is
-
still accepted and there's also some
issues actually, a little bit of some
-
rights issues or so on. So the
easiest part is just to send me a mail.
-
It will be posted sometime next year
on a more public website. But the easiest
-
is just to post me a mail. There're already
a couple of people working on it and we
-
have the rights to share it. It's just a little
bit of a question of setting it up.
-
I wanted to have the website also online
before the talk, but yeah, as with the
-
technical difficulties and so on, everything
is a little bit harder this year.
-
Herald: Indeed. Indeed. Thanks,
guys. Yes, I'd say that's it for this
-
session. Thank you very much again for
your presentation. And I'll switch back to
-
the others.
-
postroll music
-
Subtitles created by c3subtitles.de
in the year 2020. Join, and help us!