< Return to Video

The Rise of Artificial Intelligence through Deep Learning | Yoshua Bengio | TEDxMontreal

  • 0:18 - 0:21
    Our world is changing in many ways
  • 0:21 - 0:26
    and one of the things which is going
    to have a huge impact on our future
  • 0:26 - 0:29
    is artificial intelligence - AI,
  • 0:29 - 0:33
    bringing another industrial revolution.
  • 0:34 - 0:40
    Previous industrial revolutions
    expanded human's mechanical power.
  • 0:40 - 0:46
    This new revolution,
    this second machine age
  • 0:46 - 0:50
    is going to expand
    our cognitive abilities,
  • 0:50 - 0:52
    our mental power.
  • 0:53 - 0:57
    Computers are not just going
    to replace manual labor,
  • 0:58 - 1:00
    but also mental labor.
  • 1:00 - 1:03
    So, where do we stand today?
  • 1:04 - 1:08
    You may have heard
    about what happened last March
  • 1:08 - 1:12
    when a machine learning system
    called AlphaGo
  • 1:12 - 1:18
    used deep learning to beat
    the world champion at the game of Go.
  • 1:18 - 1:21
    Go is an ancient Chinese game
  • 1:21 - 1:24
    which had been much more difficult
    for computers to master
  • 1:24 - 1:26
    than the game of chess.
  • 1:27 - 1:32
    How did we succeed,
    now, after decades of AI research?
  • 1:33 - 1:37
    AlphaGo was trained to play Go.
  • 1:38 - 1:41
    First, by watching over and over
  • 1:42 - 1:47
    tens of millions of moves made
    by very strong human players.
  • 1:48 - 1:52
    Then, by playing against itself,
    millions of games.
  • 1:54 - 2:00
    Machine Learning allows computers
    to learn from examples.
  • 2:00 - 2:03
    To learn from data.
  • 2:04 - 2:07
    Machine learning
    has turned out to be a key
  • 2:07 - 2:12
    to cram knowledge into computers.
  • 2:12 - 2:14
    And this is important
  • 2:14 - 2:19
    because knowledge
    is what enables intelligence.
  • 2:20 - 2:27
    Putting knowledge into computers had been
    a challenge for previous approaches to AI.
  • 2:28 - 2:29
    Why?
  • 2:29 - 2:34
    There are many things
    which we know intuitively.
  • 2:35 - 2:38
    So we cannot communicate them verbally.
  • 2:39 - 2:43
    We do not have conscious access
    to that intuitive knowledge.
  • 2:43 - 2:47
    How can we program computers
    without knowledge?
  • 2:48 - 2:49
    What's the solution?
  • 2:49 - 2:55
    The solution is for machines to learn
    that knowledge by themselves,
  • 2:55 - 2:56
    just as we do.
  • 2:56 - 3:03
    And this is important because knowledge
    is what enables intelligence.
  • 3:03 - 3:07
    My mission has been
    to contribute to discover
  • 3:07 - 3:13
    and understand principles
    of intelligence through learning.
  • 3:13 - 3:18
    Whether animal, human or machine learning.
  • 3:19 - 3:25
    I and others believe that there are
    a few key principles,
  • 3:25 - 3:27
    just like the law of physics.
  • 3:28 - 3:33
    Simple principles which could explain
    our own intelligence
  • 3:33 - 3:37
    and help us build intelligent machines.
  • 3:38 - 3:42
    For example, think about the laws
    of aerodynamics
  • 3:42 - 3:48
    which are general enough to explain
    the flight of both, birds and planes.
  • 3:49 - 3:55
    Wouldn't it be amazing to discover
    such simple but powerful principles
  • 3:55 - 3:59
    that would explain intelligence itself?
  • 4:00 - 4:03
    Well, we've made some progress.
  • 4:04 - 4:11
    My collaborators and I have contributed
    in recent years in a revolution in AI
  • 4:12 - 4:16
    with our research on neural networks
    and deep learning,
  • 4:16 - 4:21
    an approach to machine learning
    which is inspired by the brain.
  • 4:22 - 4:25
    It started with speech recognition
  • 4:25 - 4:30
    on your phones,
    with neural networks since 2012.
  • 4:31 - 4:36
    Shortly after, came a breakthrough
    in computer vision.
  • 4:37 - 4:43
    Computers can now do a pretty good job
    of recognizing the content of images.
  • 4:44 - 4:50
    In fact, they approach human performance
    on some benchmarks over the last 5 years.
  • 4:51 - 4:55
    A computer can now get
    an intuitive understanding
  • 4:55 - 4:58
    of the visual appearance of a Go-board
  • 4:58 - 5:02
    that is comparable to that
    of the best human players.
  • 5:02 - 5:03
    More recently,
  • 5:03 - 5:07
    following some discoveries made in my lab,
  • 5:07 - 5:11
    deep learning has been used to translate
    from one language to another
  • 5:11 - 5:14
    and you are going to start seeing
    this in Google translate.
  • 5:15 - 5:18
    This is expanding the computer's ability
  • 5:18 - 5:23
    to understand and generate
    natural language.
  • 5:24 - 5:26
    But don't be fooled.
  • 5:26 - 5:30
    We are still very, very far from a machine
  • 5:30 - 5:34
    that would be as able as humans
  • 5:34 - 5:38
    to learn to master
    many aspects of our world.
  • 5:39 - 5:41
    So, let's take an example.
  • 5:42 - 5:47
    Even a two year old child
    is able to learn things
  • 5:47 - 5:51
    in a way that computers
    are not able to do right now.
  • 5:52 - 5:56
    A two year old child actually
    masters intuitive physics.
  • 5:57 - 6:02
    She knows when she drops a ball
    that it is going to fall down.
  • 6:02 - 6:06
    When she spills some liquids
    she expects the resulting mess.
  • 6:07 - 6:10
    Her parents do not need to teach her
  • 6:10 - 6:13
    about Newton's laws
    or differential equations.
  • 6:14 - 6:20
    She discovers all these things by herself
    in a unsupervised way.
  • 6:21 - 6:28
    Unsupervised learning actually remains
    one of the key challenges for AI.
  • 6:28 - 6:33
    And it may take several more decades
    of fundamental research
  • 6:33 - 6:35
    to crack that knot.
  • 6:35 - 6:41
    Unsupervised learning is actually trying
    to discover representations of the data.
  • 6:42 - 6:44
    Let me show you and example.
  • 6:44 - 6:49
    Consider a page on the screen
    that you're seeing with your eyes
  • 6:49 - 6:54
    or that the computer is seeing
    as an image, a bunch of pixels.
  • 6:55 - 7:00
    In order to answer a question
    about the content of the image
  • 7:01 - 7:05
    you need to understand
    its high-level meaning.
  • 7:06 - 7:11
    This high level meaning corresponds
    to the highest level of representation
  • 7:11 - 7:12
    in your brain.
  • 7:13 - 7:18
    Low down, you have
    the individual meaning of words
  • 7:19 - 7:24
    and even lower down, you have characters
    which make up the words.
  • 7:25 - 7:28
    Those characters could be
    rendered in different ways
  • 7:28 - 7:31
    with different strokes
    that make up the characters.
  • 7:32 - 7:35
    And those strokes are made up of edges
  • 7:35 - 7:37
    and those edges are made up of pixels.
  • 7:37 - 7:40
    So these are different
    levels of representation.
  • 7:41 - 7:44
    But the pixels are not
    sufficient by themselves
  • 7:44 - 7:47
    to make sense of the image,
  • 7:47 - 7:52
    to answer a high level question
    about the content of the page.
  • 7:53 - 7:58
    Your brain actually has
    these different levels of representation
  • 7:58 - 8:02
    starting with neurons
    in the first visual area of cortex - V1,
  • 8:02 - 8:05
    which recognizes edges.
  • 8:05 - 8:09
    And then, neurons in the second
    visual area of cortex - V2,
  • 8:09 - 8:13
    which recognizes strokes and small shapes.
  • 8:13 - 8:17
    Higher up, you have neurons
    which detect parts of objects
  • 8:17 - 8:20
    and then objects and full scenes.
  • 8:21 - 8:25
    Neural networks,
    when they're trained with images,
  • 8:25 - 8:29
    can actually discover these types
    of levels of representation
  • 8:29 - 8:33
    that match pretty well
    what we observe in the brain.
  • 8:34 - 8:39
    Both, biological neural networks,
    which are what you have in your brain
  • 8:39 - 8:43
    and the deep neural networks
    that we train on our machines
  • 8:43 - 8:48
    can learn to transform from one level
    of representation to the next,
  • 8:48 - 8:53
    with the high levels corresponding
    to more abstract notions.
  • 8:53 - 8:58
    For example the abstract notion
    of the character A
  • 8:58 - 9:01
    can be rendered in many different ways
    at the lowest levels
  • 9:01 - 9:04
    as many different configurations of pixels
  • 9:04 - 9:09
    depending on the position,
    rotation, font and so on.
  • 9:10 - 9:16
    So, how do we learn
    these high levels of representations?
  • 9:17 - 9:21
    One thing that has been
    very successful up to now
  • 9:21 - 9:23
    in the applications of deep learning,
  • 9:23 - 9:26
    is what we call supervised learning.
  • 9:26 - 9:32
    With supervised learning, the computer
    needs to be taken by the hand
  • 9:32 - 9:35
    and humans have to tell the computer
    the answer to many questions.
  • 9:35 - 9:41
    For example, on millions and millions
    of images, humans have to tell the machine
  • 9:41 - 9:44
    well... for this image, it is a cat.
  • 9:44 - 9:47
    For this image, it is a dog.
  • 9:47 - 9:50
    For this image, it is a laptop.
  • 9:50 - 9:56
    For this image, it is a keyboard,
    And so on, and so on millions of times.
  • 9:56 - 10:01
    This is very painful and we use
    crowdsourcing to manage to do that.
  • 10:01 - 10:03
    Although, this is very powerful
  • 10:03 - 10:06
    and we are able to solve
    many interesting problems,
  • 10:06 - 10:08
    humans are much stronger
  • 10:08 - 10:12
    and they can learn over many more
    different aspects of the world
  • 10:12 - 10:14
    in a much more autonomous way,
  • 10:14 - 10:18
    just as we've seen with the child
    learning about intuitive physics.
  • 10:18 - 10:24
    Unsupervised learning could also help us
    deal with self-driving cars.
  • 10:25 - 10:26
    Let me explain what I mean:
  • 10:26 - 10:32
    Unsupervised learning allows computers
    to project themselves into the future
  • 10:32 - 10:37
    to generate plausible futures
    conditioned on the current situation.
  • 10:38 - 10:43
    And that allows computers to reason
    and to plan ahead.
  • 10:43 - 10:48
    Even for circumstances
    they have not been trained on.
  • 10:49 - 10:50
    This is important
  • 10:50 - 10:54
    because if we use supervised learning
    we would have to tell the computers
  • 10:54 - 10:57
    about all the circumstances
    where the car could be
  • 10:57 - 11:01
    and how humans
    would react in that situation.
  • 11:02 - 11:06
    How did I learn to avoid
    dangerous driving behavior?
  • 11:07 - 11:11
    Did I have to die
    a thousand times in an accident?
  • 11:11 - 11:12
    (Laughter)
  • 11:12 - 11:15
    Well, that's the way we train
    machines right now.
  • 11:15 - 11:18
    So, it's not going to fly
    or at least not to drive.
  • 11:18 - 11:20
    (Laughter)
  • 11:21 - 11:26
    So, what we need is to train our models
  • 11:26 - 11:31
    to be able to generate plausible images
    or plausible futures, be creative.
  • 11:31 - 11:34
    And we are making progress with that.
  • 11:34 - 11:37
    So, we're training
    these deep neural networks
  • 11:37 - 11:41
    to go from high-level meaning to pixels
  • 11:41 - 11:43
    rather than from pixels
    to high level meaning,
  • 11:43 - 11:47
    going into the other direction
    through the levels of representation.
  • 11:47 - 11:50
    And this way, the computer
    can generate images
  • 11:51 - 11:55
    that are new images different
    from what the computer has seen
  • 11:55 - 11:56
    while it was trained,
  • 11:57 - 12:00
    but are plausible and look like natural images.
  • 12:02 - 12:06
    We can also use these models
    to dream up strange,
  • 12:06 - 12:09
    sometimes scary images,
  • 12:09 - 12:12
    just like our dreams and nightmares.
  • 12:13 - 12:17
    Here's some images
    that were synthesized by the computer
  • 12:17 - 12:20
    using these deep charted models.
  • 12:20 - 12:22
    They look like natural images
  • 12:22 - 12:25
    but if you look closely,
    you will see they are different
  • 12:25 - 12:29
    and they're still missing
    some of the important details
  • 12:29 - 12:31
    that we would recognize as natural.
  • 12:32 - 12:34
    About 10 years ago,
  • 12:34 - 12:39
    unsupervised learning has been
    a key to the breakthrough
  • 12:39 - 12:42
    that we obtained
    discovering deep learning.
  • 12:44 - 12:48
    This was happening in just few labs,
    including mine at the time
  • 12:48 - 12:51
    at a time when neural networks
    were not popular.
  • 12:51 - 12:55
    They were almost abandoned
    by the scientific community.
  • 12:56 - 12:59
    Now, things have changed a lot.
  • 12:59 - 13:01
    It has become a very hard field.
  • 13:01 - 13:07
    There are now hundreds of students
    every year applying for graduate studies
  • 13:07 - 13:10
    at my lab with my collaborators.
  • 13:11 - 13:17
    Montreal has become
    the largest academic concentration
  • 13:17 - 13:19
    of deep learning researchers in the world.
  • 13:20 - 13:26
    We just received a huge
    research grant of 94 million dollars
  • 13:26 - 13:30
    to push the boundaries
    of AI and data science
  • 13:30 - 13:36
    and also to transfer technology of deep
    learning and data science to the industry.
  • 13:37 - 13:44
    Business people stimulated by all this
    are creating start-ups, industrial labs,
  • 13:44 - 13:47
    many of which near the universities.
  • 13:49 - 13:50
    For example,
  • 13:50 - 13:55
    just a few weeks ago, we announced
    the launch of a start-up factory
  • 13:55 - 13:57
    called 'Element AI'
  • 13:57 - 14:00
    which is going to focus
    on the deep learning applications.
  • 14:02 - 14:06
    There are just not enough
    deep learning experts.
  • 14:06 - 14:11
    So, they are getting paid crazy salaries,
  • 14:11 - 14:17
    and many of my former academic colleagues
    have accepted generous deals
  • 14:17 - 14:21
    from companies to work in industrial labs.
  • 14:21 - 14:25
    I, for myself, have chosen
    to stay in university,
  • 14:25 - 14:27
    to work for the public good,
  • 14:27 - 14:29
    to work with students,
  • 14:29 - 14:31
    to remain independent.
  • 14:31 - 14:35
    To guide the next generation
    of deep learning experts.
  • 14:35 - 14:41
    One thing that we are doing
    beyond commercial value
  • 14:41 - 14:45
    is thinking about the social
    implications of AI.
  • 14:46 - 14:50
    Many of us are now starting
    to turn our eyes
  • 14:50 - 14:56
    towards social value added
    applications, like health.
  • 14:56 - 14:59
    We think that we can use deep learning
  • 14:59 - 15:03
    to improve treatment
    with personalized medicine.
  • 15:04 - 15:06
    I believe that in the future,
  • 15:06 - 15:10
    as we collect more data from millions
    and billions people around the earth,
  • 15:10 - 15:14
    we will be able to provide medical advice
  • 15:14 - 15:17
    to billions of people
    who don't have access to it right now.
  • 15:18 - 15:23
    And we can imagine many other
    applications for social value of AI.
  • 15:23 - 15:26
    For example, something
    that will come out of our research
  • 15:26 - 15:29
    on natural language understanding
  • 15:29 - 15:31
    is providing all kinds of services
  • 15:31 - 15:34
    like legal services,
    to those who can't afford them.
  • 15:35 - 15:37
    We are now turning our eyes
  • 15:37 - 15:41
    also towards the social implications
    of AI in my community.
  • 15:42 - 15:45
    But it's not just for experts
    to think about this.
  • 15:46 - 15:50
    I believe that beyond the math
    and the jargon,
  • 15:50 - 15:53
    ordinary people can get the sense
  • 15:53 - 15:56
    of what goes on under the hood
  • 15:56 - 16:01
    enough to participate
    in the important decisions
  • 16:01 - 16:07
    that will take place, in the next
    few years and decades about AI.
  • 16:08 - 16:09
    So please,
  • 16:10 - 16:16
    set aside your fees and give yourself
    some space to learn about it.
  • 16:18 - 16:23
    My collaborators and I have written
    several introductory papers
  • 16:23 - 16:25
    and a book entitled "Deep Learning"
  • 16:25 - 16:30
    to help students and engineers
    jump into this exciting field.
  • 16:31 - 16:36
    There are also many online resources:
    softwares, tutorials, videos..
  • 16:36 - 16:41
    and many undergraduate students
    are learning a lot of this
  • 16:41 - 16:45
    about research in deep learning
    by themselves,
  • 16:45 - 16:48
    to later join the ranks of labs like mine.
  • 16:49 - 16:55
    Ai is going to have a profound
    impact on our society.
  • 16:57 - 17:02
    So, it's important to ask:
    How are we going to use it?
  • 17:03 - 17:08
    Immense positives may come
    along with negatives
  • 17:08 - 17:10
    such as military use
  • 17:11 - 17:15
    or rapid disruptive changes
    in the job market.
  • 17:16 - 17:22
    To make sure the collective choices
    that will be made about AI
  • 17:22 - 17:23
    in the next few years,
  • 17:23 - 17:25
    will be for the benefit of all,
  • 17:25 - 17:29
    every citizen should take an active role
  • 17:29 - 17:33
    in defining how AI will shape our future.
  • 17:34 - 17:35
    Thank you.
  • 17:35 - 17:39
    (Applause)
Title:
The Rise of Artificial Intelligence through Deep Learning | Yoshua Bengio | TEDxMontreal
Description:

A revolution in AI is occurring thanks to progress in deep learning. How far are we towards the goal of achieving human-level AI? What are some of the main challenges ahead?

Yoshua Bengio believes that understanding the basics of AI is within every citizen’s reach. That democratizing these issues is important so that our societies can make the best collective decisions regarding the major changes AI will bring, thus making these changes beneficial and advantageous for all.

___________________________

Yoshua Bengio is one of the pioneers of Deep Learning. He is the head of the Montreal Institute for Learning Algorithms (MILA), Professor at the Université de Montréal, member of the NIPS board and co-founder of Element AI. With a PhD from McGill University (1991, Computer Science) and postdocs at MIT and AT&T Bell Labs, he holds the Canada Research Chair in Statistical Learning Algorithms, is a Senior Fellow of the Canadian Institute for Advanced Research and co-directs its program focused on deep learning. He is best known for his contributions to deep learning, recurrent nets, neural language models, neural machine translation and biologically inspired machine learning.

https://mila.umontreal.ca/en/
https://www.elementai.com/

___________________________

For more information visit http://www.tedxmontreal.com

This talk was given at a TEDx event using the TED conference format but independently organized by a local community. Learn more at https://www.ted.com/tedx

more » « less
Video Language:
English
Team:
closed TED
Project:
TEDxTalks
Duration:
17:54
  • Hi!

    I'd like to suggest a small correction at 16:09:

    set aside your fees and give yourself => set aside your fears...
    some space to learn about it.

    Thank you! :)

English subtitles

Revisions Compare revisions