< Return to Video

How do our brains process speech? - Gareth Gaskell

  • 0:07 - 0:14
    The average 20-year-old knows between
    27,000 and 52,000 different words.
  • 0:14 - 0:20
    By age 60, that number averages between
    35,000 and 56,000.
  • 0:20 - 0:24
    Spoken out loud, most of these words last
    less than a second.
  • 0:24 - 0:29
    So with every word, the brain has a quick
    decision to make:
  • 0:29 - 0:32
    which of those thousands of options
    matches the signal?
  • 0:32 - 0:36
    About 98% of the time, the brain chooses
    the correct word.
  • 0:36 - 0:41
    But how? Speech comprehension is different
    from reading comprehension,
  • 0:41 - 0:44
    but it’s similar to sign language
    comprehension—
  • 0:44 - 0:49
    though spoken word recognition has
    been studied more than sign language.
  • 0:49 - 0:51
    The key to our ability to understand
    speech
  • 0:51 - 0:55
    is the brain’s role as a
    parallel processor,
  • 0:55 - 0:59
    meaning that it can do multiple different
    things at the same time.
  • 0:59 - 1:01
    Most theories assume that each word
    we know
  • 1:01 - 1:06
    is represented by a separate processing
    unit that has just one job:
  • 1:06 - 1:11
    to assess the likelihood of incoming
    speech matching that particular word.
  • 1:11 - 1:15
    In the context of the brain, the
    processing unit that represents a word
  • 1:15 - 1:20
    is likely a pattern of firing activity
    across a group of neurons
  • 1:20 - 1:22
    in the brain’s cortex.
  • 1:22 - 1:24
    When we hear the beginning of a word,
  • 1:24 - 1:27
    several thousand such units
    may become active,
  • 1:27 - 1:32
    because with just the beginning of a
    word, there are many possible matches.
  • 1:32 - 1:36
    Then, as the word goes on, more and
    more units register
  • 1:36 - 1:41
    that some vital piece of information
    is missing and lose activity.
  • 1:41 - 1:43
    Possibly well before the end of the word,
  • 1:43 - 1:48
    just one firing pattern remains active,
    corresponding to one word.
  • 1:48 - 1:51
    This is called the ‘recognition point.’
  • 1:51 - 1:54
    In the process of honing in on one word,
  • 1:54 - 1:57
    the active units suppress
    the activity of others,
  • 1:57 - 1:59
    saving vital milliseconds.
  • 1:59 - 2:04
    Most people can comprehend up to
    about 8 syllables per second.
  • 2:04 - 2:07
    Yet, the goal is not only
    to recognize the word,
  • 2:07 - 2:10
    but also to access its stored meaning.
  • 2:10 - 2:14
    The brain accesses many possible meanings
    at the same time,
  • 2:14 - 2:17
    before the word has been fully identified.
  • 2:17 - 2:22
    We know this from studies which show
    that even upon hearing a word fragment––
  • 2:22 - 2:23
    like ‘cap’ ––
  • 2:23 - 2:27
    listeners will start to register multiple
    possible meanings,
  • 2:27 - 2:32
    like captain or capital,
    before the full word emerges.
  • 2:32 - 2:35
    This suggests that every time we hear a
    word
  • 2:35 - 2:38
    there’s a brief explosion of meanings in
    our minds,
  • 2:38 - 2:43
    and by the recognition point the brain
    has settled on one interpretation.
  • 2:43 - 2:46
    The recognition process moves more
    rapidly
  • 2:46 - 2:51
    with a sentence that gives us context
    than in a random string of words.
  • 2:51 - 2:55
    Context also helps guide us towards the
    intended meaning of words
  • 2:55 - 2:59
    with multiple interpretations, like ‘bat,’
    or ‘crane,’
  • 2:59 - 3:03
    or in cases of homophones
    like ‘no’ or ‘know.’
  • 3:03 - 3:07
    For multilingual people, the language
    they are listening to is another cue,
  • 3:07 - 3:13
    used to eliminate potential words
    that don’t match the language context.
  • 3:13 - 3:17
    So, what about adding completely new
    words to this system?
  • 3:17 - 3:21
    Even as adults, we may come across a
    new word every few days.
  • 3:21 - 3:25
    But if every word is represented as a
    fine-tuned pattern of activity
  • 3:25 - 3:27
    distributed over many neurons,
  • 3:27 - 3:32
    how do we prevent new words from
    overwriting old ones?
  • 3:32 - 3:34
    We think that to avoid this problem,
  • 3:34 - 3:39
    new words are initially stored in a part
    of the brain called the hippocampus,
  • 3:39 - 3:43
    well away from the main store of words
    in the cortex,
  • 3:43 - 3:46
    so they don’t share neurons
    with others words.
  • 3:46 - 3:49
    Then, over multiple nights of sleep,
  • 3:49 - 3:54
    the new words gradually transfer over
    and interweave with old ones.
  • 3:54 - 3:58
    Researchers think this gradual
    acquisition process
  • 3:58 - 4:01
    helps avoid disrupting existing words.
  • 4:01 - 4:03
    So in the daytime,
  • 4:03 - 4:07
    unconscious activity generates explosions
    of meaning as we chat away.
  • 4:07 - 4:12
    At night, we rest, but our brains are
    busy integrating new knowledge
  • 4:12 - 4:14
    into the word network.
  • 4:14 - 4:18
    When we wake up, this process ensures
    that we’re ready
  • 4:18 - 4:21
    for the ever-changing world of language.
Title:
How do our brains process speech? - Gareth Gaskell
Speaker:
Gareth Gaskell
Description:

more » « less
Video Language:
English
Team:
closed TED
Project:
TED-Ed
Duration:
04:21
lauren mcalpine approved English subtitles for How do our brains process speech?
lauren mcalpine accepted English subtitles for How do our brains process speech?
lauren mcalpine edited English subtitles for How do our brains process speech?
Tara Ahmadinejad edited English subtitles for How do our brains process speech?
Tara Ahmadinejad edited English subtitles for How do our brains process speech?

English subtitles

Revisions Compare revisions