Return to Video

36C3 - The Case for Scale in Cyber Security

  • 0:20 - 0:26
    Herald: Vincezo Izzo is entrepreneur and
    investor with a focus on cybersecurity. He
  • 0:26 - 0:32
    has started up, gotten bought, and
    repeated this a few times, and now he is
  • 0:32 - 0:37
    an advisor who advises people on starting
    up companies, getting bought, and
  • 0:37 - 0:45
    repeating that. He is also director at
    CrowdStrike and an associate at MIT Media
  • 0:45 - 0:54
    Lab.
    Just checking the time to make sure that
  • 0:54 - 1:00
    we start on time, and this is, start
    talking now. On the scale of cyber
  • 1:00 - 1:04
    security. Please give a warm welcome to
    Vincenzo.
  • 1:04 - 1:09
    Applause
  • 1:09 - 1:14
    Vincenzo Izzo: So hi, everyone, thanks for
    being here. As Karen said, I have made a
  • 1:14 - 1:19
    few changes to my career, but my
    background is originally technical, and
  • 1:19 - 1:25
    what I wanted to do today is to talk about
    a trend that I think we sort of take for
  • 1:25 - 1:32
    granted and it's to some extent obvious
    but also underappreciated. And that is the
  • 1:32 - 1:40
    cloud scale in security. Specifically,
    when I say cloud scale, what I mean is the
  • 1:40 - 1:47
    ability to process very large amounts of
    data as well as spawn computing power with
  • 1:47 - 1:55
    ease, and how that has played a role in
    our industry in the past decade or so. But
  • 1:55 - 2:01
    before I talk about that, I think some
    context is important. So I joined the
  • 2:01 - 2:06
    industry about 15 years ago and back in
    the days, even even a place like the
  • 2:06 - 2:13
    Congress was a much smaller place. It was
    to some extent cozier and the community
  • 2:13 - 2:20
    was tiny. The industry was fairly niche.
    And then something happened around 2010.
  • 2:20 - 2:24
    People realized that there were more and
    more state sponsored attacks being carried
  • 2:24 - 2:32
    out. From Operation Aurora against Google,
    to the Mandiant report APT1, that was the
  • 2:32 - 2:39
    first reported document how the Chinese
    PLA was hacking west- let's call it the
  • 2:39 - 2:47
    western world infrastructure for IP theft.
    And that changed a lot for the
  • 2:47 - 2:54
    industry. There have been two significant
    changes because of all of this attention.
  • 2:54 - 2:59
    The first one is notoriety. We went from
    being, as I said, a relatively unknown
  • 2:59 - 3:05
    industry to something that everyone talk
    about. If you open any kind of a
  • 3:05 - 3:10
    newspaper, there's almost always an
    article on cybersecurity, boardrooms talk
  • 3:10 - 3:15
    about cybersecurity... and in a sense,
    again, back when I joined, cybersecurity
  • 3:15 - 3:19
    wasn't a thing. It used to be called
    infosec. And now very few people know what
  • 3:19 - 3:25
    infosec even means. So notoriety is one
    thing, but notoriety is not the only thing
  • 3:25 - 3:29
    that changed. The other thing that changed
    is the amount of money deployed in the
  • 3:29 - 3:37
    sector. So, back in 2004, depending on the
    estimate you trust there, the total
  • 3:37 - 3:42
    spending for cybersecurity was between
    three point five to ten billion dollars.
  • 3:42 - 3:49
    Today's over 120 billion dollars. And so
    it kind of looks exponential. But the
  • 3:49 - 3:56
    spending came with a almost... Like, a
    very significant change in the type of
  • 3:56 - 4:01
    players there are in the industry today.
    So a lot of the traditional vendors that
  • 4:01 - 4:05
    used to sell security software have kind
    of disappeared. And what you have today
  • 4:05 - 4:10
    are two kinds of player largely. You have
    the big tech vendors. So you have
  • 4:10 - 4:14
    companies like Google, Amazon, Apple and
    so on, and so forth, that have sort of
  • 4:14 - 4:17
    decided to take security more seriously.
    Some of them are trying to monetize
  • 4:17 - 4:24
    security. Others are trying to use it as a
    sort of like slogan to sell more phones.
  • 4:24 - 4:30
    The other group of people or entities are
    large cloud-based security vendors. And
  • 4:30 - 4:35
    what both groups have in common is that
    they're using more and more sort of like
  • 4:35 - 4:43
    cloud-scale and cloud resources to try to
    tackle security problems. And so what I
  • 4:43 - 4:50
    want to discuss today is from a somewhat
    technical perspective, our scale has made
  • 4:50 - 4:56
    a significant impact in the way we
    approach problems, but also in the kind of
  • 4:56 - 5:01
    people that we have in the industry today.
    So what I'm gonna do is to give you a few
  • 5:01 - 5:08
    examples of the change that we've gone
    through. And one of the, I think one of
  • 5:08 - 5:13
    the important things to keep in mind is
    that what scale has done, at least in the
  • 5:13 - 5:20
    past decade, is it has given defense a
    significant edge over offense. It's not
  • 5:20 - 5:24
    necessarily here to stay, but I think it's
    an important trend that it's somewhat
  • 5:24 - 5:32
    overlooked. So let me start with endpoint
    security. So back in the 80s, a few people
  • 5:32 - 5:38
    started to toy with this idea of IDS
    systems. And the idea behind an IDS system
  • 5:38 - 5:43
    is pretty straightforward. You want to
    create a baseline benign behavior for a
  • 5:43 - 5:48
    machine, and then if that machine starts
    to exhibit anomalous behavior, you would
  • 5:48 - 5:55
    flag that as potentially malicious. This
    was the first paper published on a host
  • 5:55 - 6:00
    based IDS systems. Now, the problem with
    host based IDS systems is that they never
  • 6:00 - 6:07
    actually quite made it as a commercial
    product. And the reason for this... There
  • 6:07 - 6:11
    were largely two reasons for this: The
    first one is that it was really hard to
  • 6:11 - 6:17
    interpret results. So it was really hard
    to figure out: "Hey, here's an anomaly and
  • 6:17 - 6:22
    this is why this anomaly might actually be
    a security incident." The second problem
  • 6:22 - 6:27
    was, you had a lot of false positives and
    it was kind of hard to establish a benign
  • 6:27 - 6:32
    baseline on a single machine, because you
    had a lot of variance on how an individual
  • 6:32 - 6:37
    machine would behave. So what happened is
    that commercially we kind of got stuck
  • 6:37 - 6:44
    with antivirus, antivirus vendors, and
    signatures for a very long time. Now, fast
  • 6:44 - 6:56
    forward to 2013. As I mentioned, the APT1
    report came out and AV companies actually
  • 6:56 - 7:02
    admitted that they weren't that useful at
    detecting stuff like Stuxnet or Flame. And
  • 7:02 - 7:09
    so there was kind of like a new kid on the
    block, and the buzzword name for it was
  • 7:09 - 7:14
    EDR. So, endpoint detection and response.
    But when you strip EDR from like the
  • 7:14 - 7:20
    marketing fluff, what EDR really is, is
    effectively host-based intrusion detection
  • 7:20 - 7:27
    system at scale. So in other words, scale
    and ability to have cloud-scale has made
  • 7:27 - 7:34
    IDS systems possible in two ways. The
    first one is that because you actually now
  • 7:34 - 7:38
    have this sort of like data lake with a
    number of machines, you have much larger
  • 7:38 - 7:44
    datasets to train and test detections on.
    What that means is, it's much easier to
  • 7:44 - 7:50
    establish the benign baseline, and it's
    much easier to create proper detection, so
  • 7:50 - 7:55
    they don't detect just malware, but also
    sort of like malware-less attacks. The
  • 7:55 - 8:01
    other thing is that EDR vendors and also
    companies that have internal EDR systems
  • 8:01 - 8:06
    have -to a large extent- economy of scale.
    And what that means is you can actually
  • 8:06 - 8:11
    have a team of analysts that can create
    explanation and sort of an ontology to
  • 8:11 - 8:18
    explain why a given detection may actually
    represent a security incident. On top of
  • 8:18 - 8:22
    it, because you have those data lake, you
    are now able to mine that for a data to
  • 8:22 - 8:28
    figure out new attack patterns that you
    weren't aware of in the past. So this in
  • 8:28 - 8:32
    itself is a pretty significant
    achievement, because we finally managed to
  • 8:32 - 8:37
    move away from signatures to something
    that works much better and is able to
  • 8:37 - 8:42
    detect a broader range of attacks. But the
    other thing that EDR system solved, sort
  • 8:42 - 8:48
    of like as a side effect, is the data
    sharing problem. So, if you've been around
  • 8:48 - 8:53
    industry for a long time, there have been
    many attempts at sharing threat data
  • 8:53 - 8:58
    across different entities and they all
    kind of failed because it was really hard
  • 8:58 - 9:04
    to establish sort of like a protocol to
    share those data. But implicitly, what EDR
  • 9:04 - 9:13
    has done, is to force people to share and
    collect threat intelligence data and just
  • 9:13 - 9:19
    in general data from endpoints. And so now
    you have the vendors being the sort of
  • 9:19 - 9:24
    implicitly trusted third party that can
    use that data to write detections that can
  • 9:24 - 9:30
    be applied to all the systems, not just an
    individual company or any individual
  • 9:30 - 9:36
    machine. And the result of that, the
    implication of that is that the meme that
  • 9:36 - 9:40
    the attacker only needs to get it right
    once and the defender needs to get it
  • 9:40 - 9:44
    right all the time is actually not that
    true anymore, because in the past you were
  • 9:44 - 9:49
    in a situation where if you had an
    offensive infrastructure, whether it was
  • 9:49 - 9:54
    servers, whether it was exploit chains,
    you could more often than not reuse them
  • 9:54 - 9:59
    over and over again. Even if you had
    malware, all you had to do was to slightly
  • 9:59 - 10:05
    mutate the sample and you would pass any
    kind of detection. But today that is not
  • 10:05 - 10:11
    true anymore in most cases. If you get
    detected on one machine, all of the
  • 10:11 - 10:15
    sudden, all of your offensive
    infrastructure has to be scrapped and you
  • 10:15 - 10:20
    need to start from scratch. So this is the
    first example and I think in itself is
  • 10:20 - 10:27
    quite significant. The second example that
    I want to talk about is fuzzing. And
  • 10:27 - 10:31
    fuzzing is interesting also for another
    reason, which is it gives us a glimpse
  • 10:31 - 10:36
    into what I think the future might look
    like. So as you're probably familiar, if
  • 10:36 - 10:40
    you've done any apps like work in the
    past, Fuzzing has been sort of like a
  • 10:40 - 10:47
    staple in the apps like arsenal for a very
    long time. But in the past, probably five
  • 10:47 - 10:52
    years or so, fuzzing has gone through some
    kind of renaissance in the sense that two
  • 10:52 - 10:58
    things have changed. Two things have
    improved massively. The first one is that
  • 10:58 - 11:05
    we finally managed to find a better way to
    assess the fitness function that we use to
  • 11:05 - 11:10
    guide fuzzing. So a few years ago,
    somebody called Michal Zalewski release a
  • 11:10 - 11:18
    fuzzer called AFL, and one of the primary
    intuitions behind AFL was that you could
  • 11:18 - 11:23
    actually instead of using code coverage to
    drive the fuzzer, you could use path
  • 11:23 - 11:29
    coverage to drive the fuzzer and that
    turned fuzzing in a way more, you know,
  • 11:29 - 11:34
    much more effective instrument to find
    bugs. But the second intuition that I
  • 11:34 - 11:40
    think is even more important and that
    changed fuzzing significantly is the fact
  • 11:40 - 11:46
    that as far as fuzzing is concerned, speed
    is more important than smarts. You know,
  • 11:46 - 11:53
    in a way. And what I mean by this is that
    when you look at AFL, AFL as an example,
  • 11:53 - 11:58
    is an extremely dumb fuzzer. It does stuff
    like byte flipping, bit flipping. It has
  • 11:58 - 12:04
    very, very simple strategies for mutation.
    But what AFL does very well is, it's an
  • 12:04 - 12:10
    extremely optimized piece of C code and it
    scales very well. And so you are in a
  • 12:10 - 12:15
    situation where if you have a reasonably
    good server, where you can run AFL, you
  • 12:15 - 12:23
    can synthesize a very complex file formats
    in very few iterations. And what I find
  • 12:23 - 12:28
    that amazing is that this intuition
    doesn't apply just to file formats. This
  • 12:28 - 12:32
    intuition applies to much more complicated
    state machines. So the other example that
  • 12:32 - 12:38
    I want to talk about as far as fuzzing
    goes, is ClusterFuzz. ClusterFuzz is a
  • 12:38 - 12:46
    fuzzing harness used by the Chrome team to
    find bugs in Chrome and ClusterFuzz has
  • 12:46 - 12:51
    been around for about six years. In the
    span of six years ClusterFuzz found
  • 12:51 - 12:55
    sixteen thousand bugs in Chrome alone,
    plus another eleven thousand bugs in a
  • 12:55 - 13:00
    bunch of open source projects. If you
    compare ClusterFuzz with the second most
  • 13:00 - 13:07
    successful fuzzer are out there for
    JavaScript engines, you'll find that the
  • 13:07 - 13:13
    second fuzzer called jsfunfuzz found about
    six thousand bugs in the span of eight to
  • 13:13 - 13:19
    nine years. And if you look at the code,
    the main difference between the two is not
  • 13:19 - 13:23
    the mutation engine. The mutation engine
    is actually pretty similar. They don't...
  • 13:23 - 13:26
    ClusterFuzz doesn't do anything
    particularly fancy, but what ClusterFuzz
  • 13:26 - 13:33
    does very well is it scales massively. So
    ClusterFuzz today runs on about twenty
  • 13:33 - 13:39
    five thousand cores. And so with fuzzing
    we're now at a stage where the bug churn
  • 13:39 - 13:45
    is so high that defense again has an
    advantage compared to offense because it
  • 13:45 - 13:51
    becomes much quicker to fix bugs than it
    becomes to fix exploit chains, which would
  • 13:51 - 13:57
    have been unthinkable just a few years
    ago. The last example that I want to bring
  • 13:57 - 14:04
    up is a slightly different one. So, a few
    months ago, the TAG team at Google found
  • 14:04 - 14:12
    in the wild a server that was used for a
    watering hole attack, and it was thought
  • 14:12 - 14:17
    that the server was used against Chinese
    Muslim dissidents. But what's interesting
  • 14:17 - 14:21
    is that the way you would detect this kind
    of attack in the past was that you would
  • 14:21 - 14:27
    have a compromised device and you would
    sort of like work backwards from there.
  • 14:27 - 14:32
    You would try to figure out how the device
    got compromised. What's interesting is
  • 14:32 - 14:36
    that the way they found the server was
    effectively to mine their local copy of
  • 14:36 - 14:43
    the Internet. And so, again, this is
    another example of scale that gives them a
  • 14:43 - 14:50
    significant advantage to defense versus
    offense. So, in all of these examples
  • 14:50 - 14:56
    that I brought up, I think when you look
    deeper into them, you realise that it's
  • 14:56 - 15:00
    not that the state of security has
    improved because we've necessarily got
  • 15:00 - 15:06
    better at security. It's that it has
    improved because we got better at handling
  • 15:06 - 15:10
    large amounts of data, storing large
    amounts of data and spawning computing
  • 15:10 - 15:19
    power and resources quickly when needed.
    So, if that is true, then one of... the
  • 15:19 - 15:23
    other thing to realise is that in many of
    these cases, when you look back at the
  • 15:23 - 15:29
    examples that I brought up, it actually is
    the case that the problem at scale looks
  • 15:29 - 15:34
    very different from the problem at a much
    smaller scale, and the solution as a
  • 15:34 - 15:39
    result is very different. So I'm going to
    use a silly example to try to drive the
  • 15:39 - 15:45
    point home. Let's say that your job is to
    audit this function. And so you need to
  • 15:45 - 15:49
    find bugs and this function. In case
    you're not familiar with C code, the
  • 15:49 - 15:58
    problem here is that you can overflow or
    underflow that buffer at your pleasure
  • 15:58 - 16:05
    just by passing a random value for "pos".
    Now, if you were to manually audit this
  • 16:05 - 16:13
    thing, or if your job was to audit
    this function, well, you could use... You
  • 16:13 - 16:18
    would have many tools you could use. You
    could do manual code auditing. You could
  • 16:18 - 16:22
    use a symbolic execution engine. You could
    use a fuzzer. You could use static
  • 16:22 - 16:28
    analysis. And a lot of the solutions that
    are optimal for this case end up being
  • 16:28 - 16:32
    completely useless, if now your task
    becomes to audit this function and this is
  • 16:32 - 16:39
    because the state machine that this
    function implements is so complex that a
  • 16:39 - 16:45
    lot of those tools don't scale to get
    here. Now, for a lot of the problems I've
  • 16:45 - 16:51
    talked about it, we kind of face the same
    situation where the solution at scale and
  • 16:51 - 16:58
    a problem of scale looks very different.
    And so one thing, one realization is that
  • 16:58 - 17:03
    engineering skills today are actually more
    important than security skills in many
  • 17:03 - 17:09
    ways. So when you look... when you think
    back at fuzzers like ClusterFuzz, or AFL,
  • 17:09 - 17:15
    or again EDR tools, what matters there is
    not really any kind of security expertise.
  • 17:15 - 17:20
    What matters there is the ability to
    design systems that scale arbitrarily
  • 17:20 - 17:26
    well, in sort of like their backend, to
    design, to write code that is very
  • 17:26 - 17:33
    performant and none of this has really
    much to do with traditional security
  • 17:33 - 17:39
    skills. The other thing you realize is
    when you combine these two things is that
  • 17:39 - 17:47
    a lot of what we consider research is
    happening in a different world to some
  • 17:47 - 17:52
    extent. So, six years ago, about six years
    ago, I gave a talk at a conference called
  • 17:52 - 17:57
    CCS and it's an academic conference. And
    basically what I... my message there was
  • 17:57 - 18:02
    that if academia wanted to do research
    that was relevant to the industry, they
  • 18:02 - 18:07
    had to talk to the industry more. And I
    think we are now reached the point where
  • 18:07 - 18:13
    this is true for industry in the sense
    that if we want to still produce
  • 18:13 - 18:20
    significant research at places like CCC,
    we are kind of in a bad spot because a lot
  • 18:20 - 18:26
    of the innovation that is practical in the
    real world is happening very large... in
  • 18:26 - 18:31
    very large environments that few of us
    have access to. And I'm going to talk a
  • 18:31 - 18:35
    bit more about this in a second. But
    before I do, there is a question that I
  • 18:35 - 18:42
    think is important to digress on a bit.
    And this is the question of:
  • 18:42 - 18:46
    Have we changed
    significantly as an industry, are we are
  • 18:46 - 18:53
    in sort of like a new age of the industry?
    And I think that if you were to split the
  • 18:53 - 18:59
    industry in phases, we left the kind of
    like artisanal phase, the phase where what
  • 18:59 - 19:04
    mattered the most was security knowledge.
    And we're now in a phase where we have
  • 19:04 - 19:09
    this large scale expert systems that
    require significant more
  • 19:09 - 19:14
    engineering skills, that they require
    security skills, but they still take input
  • 19:14 - 19:19
    from kind of like security practitioners.
    And I think there is a question of: Is
  • 19:19 - 19:24
    this it? Or is this the kind of like where
    the industry is going to stay, or is there
  • 19:24 - 19:31
    more to come? I know better than to make
    predictions in security, 'cause most of
  • 19:31 - 19:36
    the times they tend to be wrong, but I
    want to draw a parallel. And that parallel
  • 19:36 - 19:42
    is with another industry, and it's Machine
    Learning. So, somebody called Rich Sutton
  • 19:42 - 19:46
    who is one of the godfather of machine
    learning, wrote an essay called "The
  • 19:46 - 19:52
    Bitter Truth". And in that essay, he
    reflects on many decades of machine
  • 19:52 - 19:58
    learning work and what he says in the
    essay is that people tried for a very long
  • 19:58 - 20:02
    time to embed knowledge in machine
    learning systems. The rationale was that
  • 20:02 - 20:06
    if you could embed knowledge, you would
    have a smart... you could build smarter
  • 20:06 - 20:12
    systems. But it turns out that what
    actually worked were things that scale
  • 20:12 - 20:18
    arbitrarily well with more computational
    power, more storage capabilities. And so,
  • 20:18 - 20:23
    what he realized was that what actually
    worked for machine learning was search and
  • 20:23 - 20:30
    learning. And when you look at stuff like
    AlphaGo today, AlphaGo works not really
  • 20:30 - 20:36
    because it has a lot of goal knowledge. It
    works because it has a lot of computing
  • 20:36 - 20:44
    power. It has the ability to train itself
    faster and faster. And so there is a
  • 20:44 - 20:49
    question of how much of this can
    potentially port to security. Obviously,
  • 20:49 - 20:53
    security is a bit different, it's more
    adversarial in nature, so it's not quite
  • 20:53 - 20:58
    the same thing. But I think we are... we
    have only scratched the surface of what
  • 20:58 - 21:05
    can be done as far as reaching a newer
    level of automation where security
  • 21:05 - 21:10
    knowledge will matter less and less. So I
    want to go back to the AFL example that I
  • 21:10 - 21:16
    brought up earlier, because one way to
    think about AFL is to think about it as a
  • 21:16 - 21:23
    reinforcement learning fuzzer. And what I
    mean by this... is in this slide, what AFL
  • 21:23 - 21:30
    capable to do, was to take one single JPEG
    file and in the span of about twelve
  • 21:30 - 21:36
    hundred days iteration, they were
    completely random dumb mutation. Go to
  • 21:36 - 21:41
    another well-formed JPEG file. And when
    you think about it, this is an amazing
  • 21:41 - 21:48
    achievement because there was no knowledge
    of the file format in AFL. And so we are
  • 21:48 - 21:53
    in... we are now more and more building
    systems that do not require any kind of
  • 21:53 - 21:57
    expert knowledge as far as security is
    concerned. The other example that I want
  • 21:57 - 22:02
    to talk about is the Cyber Grand
    Challenge. So DARPA ... a few years ago
  • 22:02 - 22:05
    started this competition called Cyber
    Grand Challenge,
  • 22:05 - 22:10
    and the Idea behind cyber grand challenge
    was to try to answer the question of can
  • 22:10 - 22:14
    you automagically do exploit generation
    and can you automatically do patch
  • 22:14 - 22:21
    generation. And obviously they did it on
    some well toy environments. But if you
  • 22:21 - 22:25
    talk today to anybody who does automatic
    export generation research, they'll tell
  • 22:25 - 22:31
    you that we are probably five years away
    from being able to automatically
  • 22:31 - 22:36
    synthesize non trivial exploits, which is
    an amazing achievement because if you
  • 22:36 - 22:41
    asked anybody five years ago, most people,
    myself included, would tell you that
  • 22:41 - 22:46
    time would not come anytime soon. The
    third example that I want to bring up is
  • 22:46 - 22:51
    something called Amazon Macie, which is a
    new sort of service released by Amazon.
  • 22:51 - 22:56
    And what it does is basically uses machine
    learning to try to automatically identify
  • 22:56 - 23:02
    PII information and intellectual property
    in the data. You started with a AWS and
  • 23:02 - 23:08
    then tried to give you a better sense of
    what happens to that data. So in all of
  • 23:08 - 23:12
    these cases, when you think about them,
    again, it's a scenario where there is very
  • 23:12 - 23:21
    little security expertise needed. What
    matters more is engineering skills. So
  • 23:21 - 23:28
    everything I've said so far is reasonably
    positive for scale. Is a positive scale,
  • 23:28 - 23:34
    it is a positive, sort of like case for
    scale. But I think that there is another
  • 23:34 - 23:42
    side of scale that is worth touching on.
    And I think especially to this audience is
  • 23:42 - 23:49
    important to think about. And the other
    side of scale is that scale breeds
  • 23:49 - 23:55
    centralization. And so to the point I was
    making earlier about where, where is
  • 23:55 - 24:00
    research happening, where is real word
    applicable research happening, and that
  • 24:00 - 24:08
    happens increasingly in places like Amazon
    or Google or large security vendors or
  • 24:08 - 24:14
    some intelligence agencies. And so what
    that means is the field, the barriers to
  • 24:14 - 24:20
    entry to the field are are significantly
    higher. So I said earlier that I tried to
  • 24:20 - 24:25
    join the industry about 15 years ago. Back
    then, I was still in high school. And one
  • 24:25 - 24:29
    of the things that was cool about the
    industry for me was that as long as you
  • 24:29 - 24:34
    had a reasonably decent internet
    connection and a laptop, you could
  • 24:34 - 24:40
    contribute to the top of the industry. You
    could see what everyone was up to. You
  • 24:40 - 24:44
    could do research that was relevant to
    what the what the industry was working on.
  • 24:44 - 24:49
    But today, the same sort of like 15, 16
    year old kid in high school would have a
  • 24:49 - 24:55
    much harder time contributing to the
    industry. And so we are in a situation
  • 24:55 - 25:01
    where... but because scale breeds
    centralization. We are in a situation
  • 25:01 - 25:06
    where we will likely increase the barrier
    of entry to a point where if you want to
  • 25:06 - 25:11
    contribute meaningfully to security, you
    will have to go through a very
  • 25:11 - 25:16
    standardized path where you probably do
    computer science and then you go work for
  • 25:16 - 25:26
    a big tech company. And that's not
    necessarily a positive. So I think the
  • 25:26 - 25:31
    same Kranzberg principle applies to scale
    in a sense, where it has done a lot of
  • 25:31 - 25:38
    positive things for the sector, but it
    also comes with some consequences. And if
  • 25:38 - 25:45
    there is one takeaway from this talk
    that I would like you to have is to think
  • 25:45 - 25:52
    about how much something that is pretty
    mundane that we take for granted in our
  • 25:52 - 25:57
    day to day has changed the industry and
    how much that will probably contribute to
  • 25:57 - 26:00
    the next phase of the industry. And not
    just from a technical standpoint. It's not
  • 26:00 - 26:04
    that the solutions we use today are
    much different from what we used to use,
  • 26:04 - 26:08
    but also from the kind of people that are
    part of the industry and the community.
  • 26:08 - 26:13
    And that's all I had. Thank you for
    listening.
  • 26:13 - 26:24
    Applause
  • 26:24 - 26:28
    Herald: Thank you very much. We have time
    for questions. So if you have any
  • 26:28 - 26:32
    questions for Vincenzo, please line up
    behind the microphones that are marked
  • 26:32 - 26:37
    with numbers and I will give you a signal
    if you can ask a question. We also have
  • 26:37 - 26:41
    our wonderful signal angels that have been
    keeping an eye on the Internet to see if
  • 26:41 - 26:47
    there are any questions from either
    Twitter, Mastodon or IRC. Are there any
  • 26:47 - 26:53
    questions from the Internet? We'll just
    have to mic fourth... microphone number
  • 26:53 - 26:58
    nine to be turned on and then we'll have a
    question from the Internet for Vincenzo.
  • 26:58 - 27:01
    And please don't be shy. Line up behind
    the microphone. Ask any questions.
  • 27:01 - 27:05
    Signal Angel: Now it's on. But actually
    there are no questions from the Internet
  • 27:05 - 27:09
    right now.
    Herald: There must be people in the room
  • 27:09 - 27:15
    that have some questions. I cannot see
    anybody lining up. Do you have any advice
  • 27:15 - 27:19
    for people that want to work on some
    security on scale?
  • 27:19 - 27:24
    Vincenzo: I mean, I just had to think a
    lot of the interesting research is
  • 27:24 - 27:28
    happening more and more like tech
    companies and similar. And so as much as
  • 27:28 - 27:34
    it pains me. It's probably the advice to
    think either whether you can find other
  • 27:34 - 27:40
    ways to get access to large amounts of
    data or and computational power or maybe
  • 27:40 - 27:45
    consideresting into one of those places.
    Herald: And we now actually have questions
  • 27:45 - 27:51
    at microphone number one.
    Microphone 1: Can you hear me? Yeah. Thank
  • 27:51 - 27:55
    you for the great talk. You're making a
    very strong case that information at scale
  • 27:55 - 28:00
    has benefited security, but is that also
    statistical evidence for that?
  • 28:00 - 28:05
    Vincenzo: So I think, well, it's a bit
    hard to answer the question because a lot
  • 28:05 - 28:10
    of the people that have an incentive to
    answer that question are also kind of
  • 28:10 - 28:17
    biased, but I think when you look at
    metrics like well, time in terms of how
  • 28:17 - 28:22
    much time people spend on attackers
    machine, that has decreased significantly
  • 28:22 - 28:29
    like it, it has statistically decreased
    significantly. As far as the other
  • 28:29 - 28:33
    examples I brought up, like fuzzing and
    similar. I don't think I as far as I'm
  • 28:33 - 28:43
    aware, there hasn't been any sort of
    rigorous study around where now we are.
  • 28:43 - 28:51
    We've reached the place where defense has
    kind of like an edge against offense. But
  • 28:51 - 28:57
    I think if I talk to anybody who has kind
    of like some offensive security knowledge
  • 28:57 - 29:05
    or they did work in offense, the overall
    feedback that I hear is that it's becoming
  • 29:05 - 29:12
    much harder to keep bug chains alive for a
    long time. And this is in large part not
  • 29:12 - 29:18
    really for for countermeasures. It's in
    large part because bugs keep churning.
  • 29:18 - 29:23
    So there isn't a lot of
    statistical evidence, but from what I can
  • 29:23 - 29:29
    gather, it seems to be the case.
    Herald: We have one more question from
  • 29:29 - 29:32
    microphone number one.
    Microphone 1: So thank you for the
  • 29:32 - 29:36
    interesting talk. My question goes in the
    direction of the centralization that you
  • 29:36 - 29:40
    mentioned, that the large like the
    hyperscalers are converging to be the
  • 29:40 - 29:44
    hotspots for security research. So is
    there any guidance you can give for us as
  • 29:44 - 29:50
    a community how to to retain access to the
    field and contribute?
  • 29:50 - 29:53
    Vincenzo: Yes. So. So I think
    it's an interesting situation
  • 29:53 - 29:57
    because more and more there
    are open source tools that
  • 29:57 - 30:02
    allow you to gather the data. But the
    problem with these data gathering
  • 30:02 - 30:06
    exercises is not too much how to gather
    the data. The problem is what to gather
  • 30:06 - 30:12
    and how to keep it. Because when you look
    at the cloud bill, for most
  • 30:12 - 30:17
    players, it's extraordinarily high.
    And I don't unfortunately, I don't have an
  • 30:17 - 30:23
    easy solution to that. I mean, you can use
    pretty cheap cloud providers, but
  • 30:23 - 30:29
    it's still like, the expenditure is still
    an order of magnitude higher than it used
  • 30:29 - 30:35
    to be. And I don't know, maybe academia
    can step up. I'm not sure.
  • 30:35 - 30:39
    Herald: We have one last question from the
    Internet. And you can stay at the
  • 30:39 - 30:41
    microphone if you have another question
    for Vincenzo.
  • 30:41 - 30:46
    Signal: Yes. The Internet asked that. You
    ask a lot about fuzzing at scale about
  • 30:46 - 30:51
    besides OSS-Fuzz, are you aware of any
    other scaled large fuzzing infrastructure?
  • 30:51 - 30:58
    Vincenzo: That is publicly available? No.
    But when you look at, I mean when you
  • 30:58 - 31:04
    look, for instance, of the participants
    for Cyber Grand Challenge, a lot of them
  • 31:04 - 31:13
    were effectively using a significant
    amount of CPU power for fuzzing. So I'm
  • 31:13 - 31:17
    not aware of any kind of like plug and
    play fuzzing infrastructure you can use
  • 31:17 - 31:26
    aside from OSS-Fuzz. But there is a law,
    like as far as I'm aware, everyone there
  • 31:26 - 31:33
    that does fuzzing for a living has now
    access to significant resources and tries
  • 31:33 - 31:39
    to scale fuzzing infrastructure.
    Herald: If we don't have any more
  • 31:39 - 31:43
    questions, this is your last chance to run
    to a microphone or write a question on the
  • 31:43 - 31:47
    Internet. Then I think we should give a
    big round of applause to Vincenzo.
  • 31:47 - 31:48
    Vincenzo: Thank you.
  • 31:48 - 31:53
    Applause
  • 31:53 - 32:19
    subtitles created by c3subtitles.de
    in the year 2019. Join, and help us!
Title:
36C3 - The Case for Scale in Cyber Security
Description:

more » « less
Video Language:
English
Duration:
32:19

English subtitles

Revisions