35C3 Intro music Herald Angel: We at the Congress, we not only talk about technology, we also talk about social and ethical responsibility. About how we can change the world for good. The Good Technology Collective supports the development guidelines... sorry, it supports the development process of new technology with ethical engineering guidelines, that offer a practical way to take ethic and social impact into account. Yannick Leretaille - and I hope this was okay - will tell you more about it. Please welcome on stage with a very warm applause Yann Leretaille. applause Yannick Leretaille: Hi, thanks for the introduction. So before we start, can you kind of show me your hand if you, like, work in tech building products as designers, engineers, coders, product management? OK, so it's like 95 percent, 90 percent. Great. Yeah. So, today we kind of try to answer the question: What is good technology and how can we build better technology. Before that, shortly something of me. So I am Yann. I'm French- German. Kind of a hacker, among the CCC for a long time, entrepreneur, like, co- founder of a startup in Berlin. And I'm also founding member of the Good Technology Collective. The Good Technology Collective was founded about a year ago or almost over a year now actually by a very diverse expert council and we kinda have like 3 areas of work. The first one is trying to educate the public about current issues with technology, then, to educate engineers or to build better technology, and then long-term hopefully one day we'll be able to work like in legislation as well. Here, it's a bit of what we achieved so far. We've like 27 council members now. We have several media partnerships and published around 20 articles, that's kind of the public education part. Then we organized or participated in roughly 15 events already. And we are now publishing one standard, well, kind of today actually, and applause and if you're interested in what we do, then, yeah, sign up for the newsletter and we keep you up to date and you can join events. So as I said the Expert Council is really, really diverse. We have everything from people in academia, to people in government, to technology makers, to philosophers, all sorts, journalists. And the reason that is the case is that a year ago we kind of noticed that in our own circles, like, as technology makers or academics, we were all talking about a lot of, kind of, voice on development and technology, but no one was really kind of getting together and looking at it from all angles. And there have been a lot of very weird and troublesome developments in the last two years. I think we really finally feel, you know like, the impact of filter bubbles. Something we have talked for like five years, but now it's like, really like, you know, deciding over elections and people become politically radicalized and society is, kind of, polarized more because they only see a certain opinion. And we have situations that we only knew, like, from science fiction, just kind of, you know, pre-crime, like, governments, kind of, over-arching and trying to use machine learning to make decisions on whether or not you should go to jail. We have more and more machine learning and big data and automization going into basically every single aspect of our lives and not all of it been has been positive. You know, like, literally everything from e-commerce to banking to navigating to moving to the vault now goes through these interfaces. That present us the data and a slice of the world at a time. And then at the same time we have really positive developments. Right? We have things like this, you know, like space travel, finally something's happening. And we have huge advances in medicine. Maybe soon we'll have, like, self-driving cars and great renewable technology. And it kind of begs the question: How can it be that good and bad use of technology are kind of showing up at such an increasing rate in this, like, on such extremes, right? And maybe the reason is just that everything got so complicated, right? Data is basically doubling every couple of years, so no human can possibly process anymore. So we had to build more and more complex algorithms to process it, connecting more and more parts together. And no one really seems to understand it anymore, it seems. And that leads to unintended consequences. I've an example here: So, Google Photos – this is actually only two years ago – launched a classifier to automatically go through all of your pictures and tell you what it is. You could say "Show me the picture of the bird in summer at this location" and it would find it for you. Kind of really cool technology, and they released it to, like, a planetary user base until someone figured out that people of color were always marked as gorillas. Of course it was a huge PR disaster, why somehow no one found out about this before it came out... But now the interesting thing is: In two years they didn't even manage to fix it! Their solution was to just block all kind of apes, so they're just not found anymore. And that's how they solved it, right? But if even Google can't solve this... what does it mean? And then, at the same time, you know, sometimes we seem to have, kind of, intended consequences? I have an example... another example here: Uber Greyball. I don't know if anyone heard about it. So Uber was very eager to change regulation and push the services globally as much as possible, and kind of starting a fight with, you know, all the taxi laws and regulation, and taxi drivers in the various countries around the world. And what they realized, of course, is that they didn't really want people to be able to, like, investigate what they were doing or, like, finding individual drivers. So they built this absolutely massive operation which was like following data in social media profiles, linking, like, your credit card and location data to find out if you were working for the government. And if you did, you would just never find a car. It would just not show up, right? And that was clearly intentional, all right. So at the same time they were pushing, like, on, like, the lobbyism, political side to change regulation, while heavily manipulating the people that were pushing to change the regulation, right? Which is really not a very nice thing to do, I would say. And... The thing that I find, kind of... worrisome about this: No matter if it's intended or unintended, is that it actually gets worse, right? The more and more systems we interconnect, the worse these consequences can get. And I've an example here: So this is a screenshot I took of Google Maps yesterday and you notice there are, like, certain locations... So they're kind of highlighted on this map and I don't know if you knew it but this map and the locations that Google highlight look different for every single person. Actually, I went again and looked today and it looked different again. So, Google is already heavily filtering and kind of highlighting certain places, like, maybe this restaurant over there, if you can see it. And I would say, like, from just opening the map, that's not obvious to you that it's doing that. Or that it's trying to decide for you which place is interesting for you. However, that's probably not such a big issue. But the same company, Google with Waymo, is also developing this – and they just started deploying them: self-driving cars. They're... ...still a good couple of years away from actually making it reality, but they are really – in terms of, like, all the others trying it at the moment – the farthest, I would say, and in some cities they started deploying self-driving cars. So now, just think like 5, 10 years into the future and you have signed up in your Google... ...self-driving car. Probably you don't have your own car, right? And you go in the car and you are like: "Hey, Yann, where do you want to go?" Do you want to go to work? Because, I mean obviously that's why I probably go most of the time. Do you want to go to your favorite Asian restaurant, like the one we just saw on the map? Which is actually not my favorite, but the first one I went to. So Google just assumed it was. Do you want to go to another Asian restaurant? Because, obviously, that's all I like. And then McDonald's. Because, everyone goes there. And maybe the fifth entry is an advertisement. And you would say: Well, Yann, you know, that's still kind of fine, but it's OK because I can still click on: 'No, I don't want these 5 options, give me, like, the full map.' But now, we went back here. So, even though you are seeing the map, you're not actually not seeing all the choices, right? Google is actually filtering for you where it thinks you want to go. So now we have, you know, the car like this symbol of mobility and freedom. It enables so much change in our society that it's actually reducing the part of the world that you see. And because, I mean these days they call it AI, I think it's just machine learning, because these machine learning algorithms all do pattern matching and basically just can recognize similarities. When you open the map and you zoom in and you select a random place, it would only suggest places to you where other people have been before. So now the restaurant that opened around the corner you'll probably not even discover it anymore. And no one will. And it will probably close. And the only ones that will stay are the ones that are already established now. And all of that without being really obvious to anyone who would use the technology. Because it has become like kind of a black box. So, I do want self-driving cars, I really do. I don't want a future like this. Right. And if we want to prevent that future, I think we have to first ask a very simple question, which is: Who is responsible for designing these products? So, do you know the answer? audience: inaudible Yann: Say it louder. audience: We are. Yann: Yeah, we are. Right. That's a really frustrating thing about it that actually gets us, right, as engineers and developers. You know we are always driven by perfection. We want to create, like, the perfect code sources. One problem, really, really nice. You know. Chasing the next challenge over and over trying to be first. But we have to realize that at the same time we are kind of working on frontier technologies, right, on things, technology, that are really kind of on the edge of values and norms we have in society. And if we are not careful and just, like, focus on our small problem and don't look at the big picture, then we have no say in on which side of the coin the technology will fall. And probably it will take a couple of years, or by that time we alreaday moved on, I guess. So. It's just that technology has become so powerful and interconnected and impactful, because we are not building stuff that it's not affecting like 10 or 100 people or a city but literally millions of people, that we really have to take a step back and not only look at the individual problem as the challenge but also the big picture. And I think if you want to do that we have to start by asking the right questions. And the first question of course is: What is good technology? So, that's also the name of the talk. Unfortunately, I don't have a perfect answer for that. And probably we will never find a perfect answer for that. So, what I would like to propose is to establish some guidelines and engineering processes that help us to build better technology. To kind of ensure the same where we have quality insurance and project management systems and processes to, like, kind of, this you were tasked with. And companies that what we build is actually, has a net positive outcome for society. And we call it the good technology standard. We've kind of been working that over, the last year, and we really wanted to make it really practical. And what we kind of realized is that if you want to make it practical you have to make it very easy to use and also mostly, actually what was surprising, just ask the right questions. So, what is important though, is that if you adapt the standard, it has to be in all project phases. It has to involve everyone. So, from, like, the CTO to, like, the product managers to actually legal. Today, legal has this interesting role, where you develop something and then you're like: Okay, now, legal, make sure that we can actually ship it. And that's what usually happens. And, yeah, down to the individual engineer. And if it's not applied globally and people start making exceptions then of course it won't be worth very much. Generally, we kind of identified four main areas that we think are important, kind of defining, kind of an abstract way, if a product is good. And the first one is empowerment. A good product should empower its users. And that's kind of a tricky thing. So, as humans we have very limited decision power. Right? And we are faced with, as I said before, like, this huge amount of data and choices. So it seems very natural to build machines and interfaces that try to make a lot of decisions for us. Like the Google Maps one we saw before. But we have to be careful because if we do that too much then the machine ends up making all decisions for us. So often, when you develop something you should really ask yourself, like, in the end if I take everything together am I actually empowering users, or am I taking responsibility away from them? Do I respect the individual choice? Why does he say: I don't want this, or they give you their preference, do we actually respect it or do we still try to, you know, just figure out what is better for them. Do my users actually feel like they benefit from using the product? So, I couldn't, actually not a lot of people ask themselves, because usually you think like in terms of: Are you benefiting your company? And I think what's really pressing in that aspect: does it help the users, the humans behind it, to grow in any way. If it helps them to be more effective or faster or do more things or be more relaxed or more healthy, right, then it's probably positive. But if you can't identify any of these, then you really have to think about it. And then, in terms of AI, in machine learning, are we actually kind of impacting their own reasoning so that they can't make proper decisions anymore. The second one is Purposeful Product Design. That one is one that, it's been kind of a pet peeve for me for a really long time. So these days we have a lot of products that are kind of like this. I don't have something specifically against Philips Hue, but there seems to be, like, this trend that is kind of, making smart things, right? You take a product, put a Wi-Fi chip on it, just slap it on there. Label it "smart", and then you make tons of profit, right? And a lot of these new products we've been seeing around us, like, everyone is saying, like, oh yeah, we will have this great interconnected feature, but most of them are actually not changing the actual product, right, like, the Wi-Fi connected washing machine today is still a boring washing machine that breaks down after two years. But it has Wi-Fi, so you can see what it's doing when you're in the park. And we think we should really think more in terms of intelligent design. How can we design it in the first place so it's intelligent, not smart. That the different components interact in a way, that it serves a purpose well, and the kind of intelligent by design philosophy is, when you start using your product you kind of try to identify the core purpose of it. And based on that, you just use all the technologies available to rebuild it from scratch. So, instead of building a Wi-Fi connect washing machine would actually try to build a better washing machine. And if it ends up having Wi-Fi, then that's good, but it doesn't has to. And along each step actually try to ask yourself: Am I actually improving washing machines here? Or am I just creating another data point? And yeah, a good example for that is, kind of, a watch. Of course it's very old and old technology, it was invented a long time ago. But back when it was invented it was for something you could have on your arm or in your pocket in the beginning and it was kind of a natural extension of yourself, right, that kind of enhances your senses because it's never there, you don't really feel it. But when you need it it's always there and then you can just look at it and you know the time. And that profoundly changed how, like, we humans actually worked in society because we couldn't meet in the same place at the same time. So, when you build a new product try to ask yourself what is the purpose of the product, who is it for. Often I talk to people and they talk to me for one hour, what like, literally the details of how they solved the problem but they can't tell me who their customer is. Then does this product actually make sense? Do I have features, and these distract my users, that I maybe just don't need. And can I find more intelligent solutions by kind of thinking outside of the box and focusing on the purpose of it. And then of course what is the long term product vision like, where do we want this to go? This kind of technology I'm developing in the next years. The next one is kind of, Societal Impact, that goes into what I talked about in the beginning with all the negative consequences we have seen. A lot of people these days don't realize that even if you're, like, in a small start up and you're working on, I don't know, a technology, or robots, or whatever. You don't know if your algorithm, or your mechanism, or whatever you build, will be used by 100 million people in five years. Because this has happened a lot, right? So, only when starting to build it you have to think: If this product would be used by 10 million, 100, maybe even a billion people, like Facebook, would it have negative consequences? Right, because then you get completely different effects in society, completely different engagement cycles and so on. Then, are we taking advantage of human weaknesses? So this is arguably something that's just their technology. A lot of products these days kind of try to hack your brain, what, we understand really well how, like, engagement works and addiction. So a lot of things, like social networks, actually have been focusing, you know, and also built by engineers, you know, trying to get a little number from 0.1% to 0.2%, can mean that you just do extensive A/B testing, create an interface that no one can stop looking at. You just continue scrolling, right? You just continue, and two hours have passed and you haven't actually talked to anyone. And this attention grabbing is kind of an issue and we can see that Apple actually now implemented screen time and they actually tell you how much time you spend on your phone. So there's definitely ways to build technology that even helps you to get away from these. And then for everything that involves AI and machine learning, you really have to take a really deep look at your data sets and your algorithms because it's very, very easy to build in biases and discrimination. And again, if you it applied to all of society many people who are less fortunate, or more fortunate, or they're just different, you know they just do different things, kind of fall out of the grid and now suddenly they can't, like, [unintelligible] anymore. Or use Uber, or Air B'n'B, or just live a normal life, or do financial transactions. And then, kind of what I said in the beginning, not only look at your product but also, if you combine it with other technologies that are upcoming, are there certain combinations that are dangerous? And for that I kind of recommend to do, like, some techno or litmus test to just try to come up with the craziest scenario that your technology could entail. And if it's not too bad then, probably good. The next thing is, kind of, sustainability. I think in today's world it really should be part of a good product, right. The first question is of course kind of obvious. Are we limiting product lifetime? Do we maybe have planned obsolescence, or if we build something that is so dependent on so many services and we're not only going to support it for one year anyways, that basically it will have to be thrown in the trash afterwards. Maybe it would be possible to add a standalone node or a very basic fallback feature so that at least the products continues to work. Especially if you talk about things like home appliances. Then, what is the environmental impact? A good example here would be, you know, crypto currencies who are now using as much energy as certain countries. And when you consider that just think like is there maybe an alternative solution that doesn't have such a big impact. And of course we are still capitalism, it has to be economically viable, but often there aren't, often it's again just really small tweaks. Then of course: Which other services are you working with? But for example I would say, like, as european companies, we're in Europe here, maybe try to work mostly with suppliers from Europe, right, because you know they follow GDPR and strict rules, and in a sense the US. Or check your supply chain if you build hardware. And then for hardware specifically that's because also I have, like, we also do hardware in my company, I always found that interesting. We're kind of in a world where everyone tries to save, like, the last little bit of money out of every device that is built and often makes the difference between plastic and metal screws like half a cent, right. And at that point it doesn't really change your margins much. And maybe as an engineer, you know, just say no and say: You know, we don't have to do that. The savings are too small to redesign everything and it will impact upon our quality so much that it just breaks earlier. These are kind of the main four points. I hope that makes sense. Then we have two more, kind of, additional checklists. The first one is data collection. So really, just if, especially like in terms of like IOT, you know, everyone focuses on kind of collecting as much data as possible without actually having an application. And I think we really have to start seeing that as a liability. And instead try to really define the application first, define which data we need for it, and then really just collect that. And we can start collecting more data later on. And that can really prevent a lot of these negative cycles we have seen. By just having machine learning organisms run on of it kind of unsupervised and seeing what comes out. Then also kind of really interesting I found that, many times, like, a lot of people are so fascinated by the amount of data, right, just try to have as many data points as possible. But very often you can realize exactly the same application with a fraction of data points. Because what you really need is, like, trends. And that usually also makes the product more efficient. Then how privacy intrusive is the data we collect? Right. There's a big difference between, let's say, the temperature in this building and everyone's individual movements here. And if it is privacy intrusive then we should really, really think hard if we want to collect it. Because we don't know how it might be used at a later point. And then, are we actually collecting data without people realizing that they do it, right, especially if we look at Facebook and Google. They're collecting a lot of data without really implicit consent. But of course at some point you like all agreed to the privacy policy. But it's often not clear to you when and which data is collected. And that's kind of dangerous and kind of in the same way if you kind of build dark patterns into your app. They kind of fool you into sharing even more data. I had, like, an example that someone told me yesterday. I don't if you know Venmo which is this American system where you pay each other with your smartphone. Basically to split the bill in a restaurant. By default, all transactions are public. So, like 200 million public transactions which everyone can see, including the description of it. So for some of the more maybe not so legal payments that was also very obvious, right? And it's totally un-obvious when you use the app that that is happening. So that's definitely a dark pattern that they're employing here. And then the next point is User Product Education and Transparency. Is a user able to understand how the product works? And, of course, we can't really ever have a perfect explanation of all the intricacies of the technology. But these days for most people almost all of the apps, the interfaces, the building technology and tech. This is a complete black box and no one is really doing an effort to explain it to them why most companies advertise it like this magical thing. But that just leads to kind of this immunization where you just look at it and you don't even try to understand it. I'm pretty sure that no one ever, like, these days is still opening up a PC and trying looking at the components, right, because everything is in tablet and it's integrated and it's sold to us like this magical media consumption machine. Then, are users informed when decisions are made for them? So we had that in Empowerment, that we should try to reduce the amount of decisions we make for the user. But sometimes, that's a good thing to do. But then, is it transparently communicated? I would be totally fine with Google Maps filtering out for me the points of interest if it would actually tell me that it's doing that. And if you can't understand why it made that decision and why it showed me this place. And maybe also have a way to switch it off if I want. But today we seem to kind of assume that we know better for the people why it's, so we found the perfect algorithm that has a perfect answer. So we don't even have to explain how it works, right? We just do it and people will be happy. But then we end up with is very negative consequences. And then, that's more like a marketing thing, how is it actually advertised? I find it, for example, quite worrisome that things like Siri and Alexa and Google home are, like, sold as these magical AI machines that make your life better, and are you personal assistant. When in reality they are actually still pretty dumb, pattern matching. And that also creates a big disconnect. Because now we have children growing up who actually think that Alexa is a person. And that's kind of dangerous. And I think we should try to prevent that because for these children, basically, it kind of creates this veil and it's humanized. And that's especially dangerous if then the machine starts to make decisions for them. And suggestions because they will take them as if a human did it for them. So, what is that? So, these are kind of the main areas. Of course it's a bit more complicated. So we just published the standard today in the first draft version. And it's basically three parts of science introduction, kind of the questions and checklists that you just saw. And then actually how to implement it in your company, which processes to have, at which point you basically should have kind of a feature gate. And I would kind of ask everyone to go there, look at it, contribute, shared it with people. We hope that we'll have a final version ready kind of in Q1 and that by then people can start to implement it. Oh, yeah. So, even though we have this standard, right, I want to make it clear having such a standard and implementing it in your organization or for yourself or your product is great. It actually doesn't remove your responsibility, right? This can only be successful if we actually all accept that we are responsible. Right? If today I build a bridge as a structural engineer and the bridge breaks down because I miscalculated, I am responsible. And I think, equally, we have to accept that if we build technology like this we also have to, kind of, assume that responsibility. And before we kind of move to Q&A, I'd like to kind of take the citations. This is Chamath Palihapitiya, former Facebook executive, from the really early times. And also, around a year ago when we actually saw the GTC he said this in a conference: "I feel tremendous guilt. I think in the back in the deep restlessness of our mind we knew something bad could happen. But I think the way we defined it is not like this. It is now literally at a point where I think we have created tools that are ripping apart the social fabric of how society works." And personally, and I hope the same for you, I do not want to be that person that five years down the line realizes that they built that technology. So if there is one take-away that you can take home from this talk, then to just start asking yourself: What is good technology, what does it mean for you? What does it mean for the products you build and what does it mean for your organization? Thanks. applause Herald: Thank you. Yann Leretaille. Do we have questions in the room? There are microphones, microphones number 1, 2, 3, 4, 5. If you have a question please speak loud into the microphone, as the people in the stream want to hear you as well. I think microphone number 1 was the fastest. So please. Question: Thank you for your talk. I just want to make a short comment first and then ask a question. I think this last thing you mentioned about offering users the options to have more control of the interface there is also a problem that users don't want it. Because when you look at the statistics of how people use online web tools, only maybe 5 percent of them actually use that option. So companies remove them because for them it seems like it's something not so efficient for user experience. This was just one thing to mention and maybe you can respond to that. But what I wanted to ask you was, that all these principles that you presented, they seem to be very sound and interesting and good. We can all accept them as developers. But how would you propose to actually sell them to companies. Because if you adopt a principle like this as an individual based on your ideology or the way that you think, okay, it's great it will work, but how would you convince a company which is driven by profits to adopt these practices? Have you thought of this and what's your idea about this? Thank you. Yann: Yeah. Maybe to the first part. First, that giving people choice is something that people do not want and that's why companies removed it. I think if you look at the development process it's basically like a huge cycle of optimization and user testing geared towards a very specific goal, right, which is usually set by leadership which is, like, bringing engagement up or increase user amount by 200 percent. So I would say the goals were, or are today, mostly misaligned. And that's why we end up with interfaces that are in a very certain way, right? If we set the goals differently, and I mean that's why we have like UI and UX research. I'm very sure we can find ways to build interfaces that are just different. And still engaging, but also give that choice. To the second question. I mean it's kind of interesting. So I wouldn't expect a company like Google to implement something like this, because it's a bit against that. This is more by that point probably but I've met a lot of, like, also high level executives already, who were actually very aware of kind of the issues of technology that they built. And there is definitely interest there. Also, more like industrial side, and so on, especially, it seems like self-driving cars to actually adopt that. And in the end I think, you know, if everyone actually demands it, then there's a pretty high probability that it might actually happen. Especially, as workers in the tech field, we are quite flexible in the selection of our employer. So I think if you give it some time, that's definitely something that's very possible. The second aspect is that, actually, if we looked at something like Facebook, I think they overdid it. Say, optimize that so far and push the engagement machine and kind of triggering like your brain cells to never stop being on the site and keeps scrolling, that people got too much of it. And now they're leaving the platform in droves. And of course Facebook would not go down, they own all these other social networks. But for the product itself. as you can see, that, long term it's not even necessarily a positive business outcome. And everything we are advertising here still also to have very profitable businesses, right, just tweaking the right screws. Herald: Thank you. We have a question from the interweb. Signal Angel: Yes. Fly asks a question that goes into a similar direction. In recent months we had numerous reports about social media executives forbidding their children to use the products they create at work. I think these people know that their products are made addictive deliberately. Do you think your work is somewhat superfluous because big companies are doing the opposite on purpose. Yann: Right. I think that's why you have to draw the line between intentional and unintentional. If we go to intentional things like what Uber did and so on. At some point it should probably become a legal issue. Unfortunately we are not there yet and usually regulation is kind of lagging way behind. So I think for now we should focus on, you know, the more unintentional consequences, of which there are plentiful and kind of appeal to the good in humans. Herald: Okay. Microphone number 2 please. Q: Thank you for sharing your ideas about educating the engineer. What about educating the customer, the consumer who purchases the product. Yann: Yeah. So that's a really valid point. Right. As I said I think [unintelligible] like part of your product development. And the way you build a product should also be how you educate your users on how it works. Generally, we have a really big kind of technology illiteracy problem. Things have been moving so fast in the last year that most people haven't really caught up and they just don't understand things anymore. And I think again that's like a shared responsibility, right? You can't just do that in the tech field. You have to talk to your relatives, to people. That's why we're doing, like, this series of articles and media partnerships to kind of explain and make these things transparent. One thing we just started working on is a children's book. Because for children, like, the entire world just exists with this shiny glass surfaces and they don't understand at all what is happening. But it's also primetime to explain to them, like, really simple machine learning algorithms. How they work, how like, filterbubbles work, how decisions are made. And if you understand that from an early age on, then maybe you'll be able to deal with what is happening. In a way better, an educated way. But I do think that is a very long process and so only if we start and the more work we invest in that, the earlier people will be better educated. Herald: Thank you. Microphone number 1 please. Q: Thanks for sharing your insights. I feel like, while you presented these rules along with their meaning, the specific selection might seem a bit arbitrary. And for my personal acceptance and willingness to implement them it would be interesting to know the reasoning, besides common sense, that justifies this specific selection of rules. So, it would be interesting to know if you looked at examples from history, or if you just sat down and discussed things, or if you just grabbed some rules out of the air. And so my question is: What influenced you for the development of these specific rules? Yann: It's a very complicated question. So how did we come up this specific selection of rules and also, like, the main building blocks of what we think should good technology be. Well, let's say first what we didn't want to do, right. We didn't want to create like a value framework and say, like, this is good, this is bad, don't do this kind of research or technology. Because this would also be outdated, it doesn't apply to everyone. We probably couldn't even agree in the expert council on that because it's very diverse. Generally, we try to get everyone on the table. And we talked about issues we had, like, for example me as an entrepreneur. And when I was, like, in developing products with our own engineers. Issues we've seen in terms of public perception. Issues we've seen, like, on a more governmental level. Then we also have, like, cryptologists in there. So we looked at that as well and then we made a really, really long list and kind of started clustering it. And a couple of things did get cut off. But generally, based on the clustering, these were kind of the main themes that we saw. And again, it's really more of a tool for yourself as a company that developers, designers and engineers to really understand the impact and evaluate it. Right. This is what these questions are aimed at. And we think that for that they do a very good job. From microphone 1: Thank you. Herald: Thank you. And I think. Microphone number 2 has a question again. Q: Hi. I was just wondering how you've gone about engaging with other standards bodies, that perhaps have a wider representation. It looks largely like from your team of the council currently that there's not necessarily a lot of engagement outside of Europe. So how do you go about getting representation from Asia. For example. Yann: No, at the moment you're correct the GTC is mostly a European initaitive. We are in talks with other organizations who work on similar issues and regularly exchange ideas. But, yeah, we thought we should probably start somewhere. In Europe is actually a really good place to start. Like a societal discourse about technology and the impact it has and also to to have change. But I think if for example compared to things like Asia or the US where is a very different perception of privacy and technology and progress and like the rights of the individual Europe is actually a really good place to do that. And we can also see things like GDPR regulation, that actually, ... It's kind of complicated. It's also kind of a big step forward in terms of protecting the individual from exactly these kind of consequences. Of course though, long term we would like to expand this globally. Herald: Thank you. Microphone number 1 again. Q: Hello. Just a short question. I couldn't find a donate button on your website. Do you accept donations? Is money a problem? Like, do you need it? Yann: Yes, we do need money. However it's a bit complicated because we want to stay as independent as possible. So we are not accepting project related money. So you can't like say we want to do certain research product with you, it has to be unconditional. And the second thing we do is for the events we organize. We usually have sponsors that provide, like, venue and food and logistics and things like that. But that's an, ... for the event. And again, I can't, like, change the program of it. So if you want to do that you can come into contact with us. We don't have a mechanism yet for individuals to donate. We might add that. Herald: Thank you. Did you think about Patreon or something like that? Yann: We thought about quite a few options. Yeah, but it's actually not so easy to not fall into the trap that, like, as organizations in space have been, like, Google at some point sweeps in and it's like: Hey, do you want all this cash. And then very quickly you have a big conflict of interest. Even if you don't want that to happen it starts happening. Herald: Yeah right. Number 1 please. Q: I was wondering how do you unite the second and third points in your checklist. Because the second one is intelligence by design. The third one is to take into account future technologies. But companies do not want to push back their technologies endlessly to take into account future technologies. And on the other hand they don't want to compromise their own design too much. Yann: Yeah. Okay. Okay. Got it. So you were saying if we should always stop these, like, future scenarios and the worst case and everything and incorporate every possible thing that might happen in the future we might end up doing nothing because everything looks horrible. For that I would say, like, we are not like technology haters. We are all from areas working in tech. So of course the idea is that you can just take a look at what is there today and try to make an assessment based on that. And the idea is if you look it up and meet the standards that over time actually you try to,... When you add new major features to look back at your assessment from before and see if it changed. So the idea is you kind of create a snapshot of how it is now. And this kind of document that you end up as part of your documentation kind of evolved over time as your product changes and the technology around it changes as well. Herald: Thank you. Microphone number 2. Q: So thanks for the talk and especially the effort. Just to echo back the question that was asked a bit before on starting with Europe. I do think it's a good option. What I'm a little bit worried is it might be the only option. It might become irrelevant rather quickly because it's easy to do, it's less hard to implement. Maybe in Europe now. Okay. The question is. It might work in Europe now but if Europe doesn't have the same economical power it cannot bog in as much politically with let's say China or the US in Silicon Valley. So will it still be possible and relevant if the economical balance shifts? Yann: Yes, I mean we have to start somewhere, right? Just saying "Oh, economical balance will shift anyway, Google will invent singularity, and that's why we shouldn't do anything" is, I think, one of the reasons why we actually got here, why it kind of is this assumption that there is like this really big picture that is kind of working against us, so we all do our small part to fulfill that kind of evil vision by not doing anything. I think we have to start somewhere and I think for having operated for one year, we have been actually quite successful so far and we have a good progress. And I'm totally looking forward to make it a bit more global and to start traveling more, I think that like one event outside Europe last year in the US and that will definitely increase over time, and we're also working on making kind of our ambassadors more mobile and kind of expand to other locations. So it's definitely on the roadmap but it's not like yeah, just staying here. But yeah, you have to start somewhere and that's what we did. Herald: Nice, thank you. Number 1 please. Mic 1: Yeah. One thing I haven't found was – all those general rules you formulated fit into the more general rules of society, like the constitutional rules. Have you considered that and it's just not clearly stated and will be stated, or did you develop them more from the bottom up? Yann: Yes, you are completely right. So we are defining the process and the questions to ask yourself, but we are actually not defining a value framework. The reason for that is that societies are different, as I said they are widely different expectations towards technology, privacy, how society should work, all the ones about. The second one is that every company is also different, right, every company has their own company culture and things they want to do and they don't want to do. If I would say, for example, we would have put in there "You should not build weapons or something like that", right, that would mean that all these companies that work in that field couldn't try to adapt it. And while I don't want them to build weapons maybe in their value framework that's OK and we don't want to impose that, right. That's why I said in the beginning we actually, we're called the Good Technology Collective, we are not defining what it is and I think that's really important. We are not trying to impose our opinion here. We want others to decide for themselves what is good and cannot support them and guide them in building products that they believe are good. Herald: Thank you. Number two . Mic 2: Hello, thanks for sharing. As engineer we always want users to spend more time to use our product, right? But I'm working at mobile game company. Yep. We are making, we are making a world that users love our product. So we want users spend more time in our game. So we may make a lot of money, yeah, but when users spend time to play our game they may lose something. Yeah. You know. So how do we think about the balance in a game, mobile game. Yeah. Yann: Hmm. It's a really difficult question. So the question was like specifically for mobile gaming. Where's kind of the balance between trying to engage people more and, yeah, basically making them addicted and having them spend all their money, I guess. I personally would say it's about intent, right? It's totally fine with a business model where you make money with a game. I mean that's kind of good and people do want entertainment. But if you actively use, like, research in how, like, you know, like the brain actually works and how it get super engaged, and if you basically build in, like, gamification and lotteries, which a lot of, I think, have done, where basically your game becomes a slot machine, right, you always want to see the next opening of a crate and see what you got. Kind of making it a luck based game, actually. I think if you go too far into that direction, at some point you cross the line. Where that line is you have to decide yourself, right, some of it could be a good game and dynamic but there definitely some games out there, I would say with quite a reason to say that they pushed to the limit quite a bit too far. And if you actually look how they did it because they wrote about it, they actually did use very modern research and very extensive testing to really find out these, all these patterns that make you addicted. And then it's not much better than an actual slot machine. And that probably we don't want. Herald: So it's also an ethical question for each and every one of us, right? Yann: Yes. Herald: I think there is a light and I think this light means the interwebs has a question. Signal angel: I, there's another question from ploy about practical usage, I guess. Are you putting your guidelines at work in your company? You said you're an entrepeneur. Yann: That's a great question. Yes, we will. So we kind of just completed some and there was kind of a lot of work to get there. Once they are finished and released we will definitely be one of the first adopters. Herald: Nice. And with this I think we're done for today. Yann: Perfect. Herald: Yann, people, warm applause! applause postroll music subtitles created by c3subtitles.de in the year 2020. Join, and help us!