Return to Video

Randomized Trials: The Ideal Weapon

  • 0:00 - 0:05
    - [Narrator] The path from cause
    to effect is dark and dangerous,
  • 0:06 - 0:09
    but the weapons
    of Econometrics are strong.
  • 0:10 - 0:15
    Behold the most powerful,
    the sword of random assignment,
  • 0:15 - 0:18
    which cuts to the heart
    of causal questions.
  • 0:23 - 0:28
    We begin with our most powerful
    and costliest weapon --
  • 0:28 - 0:30
    randomized trials.
  • 0:30 - 0:31
    - [Student] Awesome.
  • 0:31 - 0:35
    - Every metric's mission begins
    with a causal question.
  • 0:35 - 0:38
    Clear questions lead
    to clear answers.
  • 0:39 - 0:43
    The clearest answers come
    from randomized trials.
  • 0:44 - 0:48
    Let's see how and why
    randomized trials provide
  • 0:48 - 0:51
    especially convincing answers
    to causal questions.
  • 0:52 - 0:56
    - [Josh] Like a finely honed sword,
    randomized trials cut
  • 0:56 - 0:58
    to the heart of a causal problem,
  • 0:58 - 1:01
    creating convincing apples-
    to-apples comparisons.
  • 1:01 - 1:04
    Yet like any finely made weapon,
  • 1:04 - 1:07
    randomized trials are expensive
    and cannot be done quickly.
  • 1:08 - 1:12
    - Randomized trials originate
    in medical research
  • 1:12 - 1:16
    where they're called
    randomized clinical trials or RCTs.
  • 1:17 - 1:21
    The U.S. Food and Drug Administration
    requires drug manufacturers
  • 1:21 - 1:25
    to establish the safety
    and efficacy of new drugs
  • 1:25 - 1:26
    or medical treatments.
  • 1:27 - 1:30
    They do this through a series
    of RCTs.
  • 1:30 - 1:35
    That's why randomized trials are
    said to measure treatment effects.
  • 1:36 - 1:38
    You may have contributed
    to another kind
  • 1:38 - 1:40
    of randomized trial --
  • 1:40 - 1:43
    the A/B tests
    Silicon Valley companies use
  • 1:43 - 1:46
    to compare marketing strategies.
  • 1:46 - 1:50
    For example, Amazon randomizes
    search results in a constant stream
  • 1:50 - 1:52
    of hidden experiments.
  • 1:52 - 1:53
    - [Woman] Oh.
    - [Man] Interesting.
  • 1:53 - 1:57
    - Randomized trials are also
    important in education research.
  • 1:57 - 2:00
    They've been used to answer
    a causal question near and dear
  • 2:00 - 2:02
    to my beating teacher's heart --
  • 2:02 - 2:05
    should laptops
    and other electronic devices
  • 2:05 - 2:07
    be allowed in the classroom?
  • 2:07 - 2:10
    Many see classroom electronics
    as a learning aid.
  • 2:10 - 2:14
    But others, like me, think they're
    a damaging distraction.
  • 2:14 - 2:15
    Who's right?
  • 2:22 - 2:24
    Metrics masters teaching
    at West Point,
  • 2:24 - 2:27
    the military college
    that trains American Army officers,
  • 2:27 - 2:31
    designed a randomized trial
    to answer this question.
  • 2:31 - 2:33
    These masters randomly assign
    West Point cadets
  • 2:33 - 2:36
    into economics classes
    operating under different rules.
  • 2:37 - 2:39
    Unlike most American colleges,
  • 2:39 - 2:42
    the West Point default
    is no electronics.
  • 2:43 - 2:46
    For purposes of this experiment,
    some students were left
  • 2:46 - 2:49
    in such traditional
    technology-free classes --
  • 2:49 - 2:52
    no laptops, no tablets,
    and no phones!
  • 2:53 - 2:56
    This is the control group,
    or baseline case.
  • 2:56 - 3:00
    Another group was allowed
    to use electronics.
  • 3:00 - 3:03
    This is the treatment group,
    subject to a changed environment.
  • 3:03 - 3:06
    The treatment in this case
    is the unrestricted use
  • 3:06 - 3:08
    of laptops or tablets in class.
  • 3:09 - 3:12
    Every causal question has
    a clear outcome --
  • 3:12 - 3:16
    the variables we hope to influence,
    defined in advance of the study.
  • 3:16 - 3:19
    The outcomes
    in the West Point electronics study
  • 3:19 - 3:20
    are final exam scores.
  • 3:20 - 3:24
    The study seeks to answer
    the following question --
  • 3:24 - 3:28
    what is the causal effect
    of classroom electronics on learning
  • 3:28 - 3:30
    as measured by exam scores?
  • 3:30 - 3:33
    - [Narrator] West Point economics
    students were randomly assigned
  • 3:33 - 3:36
    to either the treatment
    or control groups.
  • 3:36 - 3:40
    Random assignment creates
    ceteris paribus comparisons,
  • 3:40 - 3:44
    allowing us to draw causal
    conclusions by comparing groups.
  • 3:45 - 3:49
    The causality revealing power
    of a randomized trial comes
  • 3:49 - 3:53
    from a statistical property
    called the law of large numbers.
  • 3:53 - 3:55
    When statisticians
    and mathematicians uncover
  • 3:55 - 3:57
    something important
    and reliably true
  • 3:57 - 4:01
    about the natural world,
    they call it a law.
  • 4:02 - 4:05
    The law of large numbers says
    that when the groups randomized
  • 4:05 - 4:08
    are large enough, the students
    in them are sure to be similar
  • 4:08 - 4:10
    on average, in every way.
  • 4:11 - 4:14
    This means that groups of students
    randomly divided can be expected
  • 4:14 - 4:18
    to have similar family background,
    motivation, and ability.
  • 4:19 - 4:22
    Sayonara selection bias,
    at least in theory.
  • 4:23 - 4:26
    In practice, the groups randomized
    might not be large enough
  • 4:26 - 4:28
    for the law
    of large numbers to kick in.
  • 4:29 - 4:32
    Or the researchers might have
    messed up the random assignment.
  • 4:32 - 4:35
    As in any technical endeavor,
    even experienced masters
  • 4:35 - 4:37
    are alert for a possible foul-up.
  • 4:38 - 4:40
    We therefore check for balance,
  • 4:40 - 4:43
    comparing student background
    variables across groups
  • 4:43 - 4:45
    to make sure
    they indeed look similar.
  • 4:47 - 4:49
    - [Narrator] Here's
    the balance check for West Point.
  • 4:50 - 4:52
    This table has two columns --
  • 4:52 - 4:55
    one showing data
    from the control group
  • 4:55 - 4:57
    and one showing data
    from the treatment group.
  • 4:58 - 5:02
    The rows list a few of the variables
    that we hope are balanced --
  • 5:02 - 5:06
    sex, age, race, and high school GPA,
    among others.
  • 5:07 - 5:09
    The first row indicates
    what percentage
  • 5:09 - 5:12
    of each group is female.
  • 5:12 - 5:16
    It's 17% for the control group
    and 20% for the treatment group.
  • 5:18 - 5:21
    Kamal, how does balance
    look for GPA?
  • 5:22 - 5:25
    - [Kamal] Controls have
    a GPA of 2.87,
  • 5:25 - 5:29
    while treated have
    a GPA of 2.82, pretty close.
  • 5:30 - 5:34
    - [Narrator] Happily the two groups
    look similar all around.
  • 5:34 - 5:36
    - How big does our sample have to be
  • 5:36 - 5:37
    for the law
    of large numbers to kick in?
  • 5:38 - 5:39
    - [Narrator] The West Point study,
  • 5:39 - 5:43
    which involved about 250 students
    in each group,
  • 5:43 - 5:45
    is almost certainly big enough.
  • 5:45 - 5:48
    There are no hard
    and fast rules here.
  • 5:48 - 5:51
    In another video, you'll learn
    how to confirm the hypothesis
  • 5:51 - 5:55
    of group balance
    with formal statistical tests.
  • 5:55 - 5:56
    - [Man] Exciting.
  • 6:00 - 6:02
    - The heart of the matter
    in this table
  • 6:02 - 6:04
    is the estimated treatment effect.
  • 6:05 - 6:07
    Remember the treatment
    in this case is permission
  • 6:07 - 6:09
    to use electronics in class.
  • 6:10 - 6:14
    Treatment effects compare averages
    between control and treatment groups.
  • 6:15 - 6:18
    The group Allowed Classroom
    Electronics had
  • 6:18 - 6:22
    an average final exam score
    0.28 standard deviations
  • 6:22 - 6:24
    below the score for students
    in the control group.
  • 6:26 - 6:28
    How big an effect is this?
  • 6:28 - 6:32
    Social scientists measure test scores
    in standard deviation units
  • 6:32 - 6:35
    because these units are easily
    compared across studies.
  • 6:35 - 6:39
    We know from a long history
    of research on classroom learning
  • 6:39 - 6:41
    that 0.28 is huge.
  • 6:41 - 6:46
    A decline of 0.28 is like taking
    the middle student in the class
  • 6:46 - 6:48
    and pushing him or her down
    to the bottom third.
  • 6:49 - 6:52
    How certain can we be
    that these large results
  • 6:52 - 6:53
    are meaningful?
  • 6:53 - 6:56
    After all, we're looking
    at a single randomized division
  • 6:56 - 6:58
    between treatment
    and control groups.
  • 6:58 - 7:01
    Other random splits might have
    produced something different.
  • 7:01 - 7:04
    - We therefore quantify
    the sampling variance
  • 7:04 - 7:06
    in estimates of causal effects.
  • 7:06 - 7:08
    - What is sampling variance?
  • 7:09 - 7:11
    - [Narrator] Sampling variance
    tells us how likely
  • 7:11 - 7:16
    a particular statistical finding
    is to be a chance result,
  • 7:16 - 7:19
    rather than indicative
    of an underlying relationship.
  • 7:20 - 7:23
    Sampling variance is summarized
    by a number,
  • 7:23 - 7:27
    called the standard error
    of the estimated treatment effect.
  • 7:27 - 7:29
    - [Students muttering] I don't get it.
    - What is she even talking... ?
  • 7:29 - 7:33
    - Don't worry -- we'll cover
    this important idea in depth later.
  • 7:33 - 7:34
    - [Student] Great.
  • 7:34 - 7:36
    - Just remember the smaller
    the standard error,
  • 7:36 - 7:39
    the more conclusive the findings.
  • 7:39 - 7:42
    On the other hand,
    if the standard error is large
  • 7:42 - 7:45
    relative to the effect
    we're trying to estimate,
  • 7:45 - 7:47
    there's a pretty good chance
    we'd get something different
  • 7:47 - 7:50
    if we were to rerun the experiment.
  • 7:50 - 7:54
    You can think of the standard error
    as a way to gauge how much
  • 7:54 - 7:57
    we can trust the result we're seeing.
  • 7:57 - 7:58
    - [Students] Okay.
  • 7:58 - 8:01
    - [Narrator] In this study,
    the relevant standard error is 0.1.
  • 8:02 - 8:05
    - For now it's enough to learn
    a simple rule of thumb.
  • 8:05 - 8:07
    When an estimated
    treatment effect is more
  • 8:07 - 8:09
    than double its standard error,
  • 8:09 - 8:12
    the odds this non-zero result is due
    to chance are very low,
  • 8:12 - 8:14
    roughly 1 in 20.
  • 8:15 - 8:18
    Because this is so unlikely,
    we say that estimates that are two
  • 8:18 - 8:21
    or more times larger
    than the associated standard errors
  • 8:21 - 8:23
    are statistically significant.
  • 8:24 - 8:28
    - Camilla, is the treatment effect
    in the West Point study
  • 8:28 - 8:30
    statistically significant?
  • 8:31 - 8:36
    - The standard error is 0.10
    and the treatment effect is 0.28,
  • 8:36 - 8:40
    so 0.28 is more than two times larger
    than 0.10, so yeah.
  • 8:40 - 8:42
    - Correct.
  • 8:42 - 8:46
    The lost learning caused
    by electronics use in Econ 101
  • 8:46 - 8:50
    is therefore both large
    and statistically significant.
  • 8:50 - 8:51
    - [Man] Nice.
  • 8:57 - 9:00
    - Randomized trials usually provide
    the most convincing answers
  • 9:00 - 9:01
    to causal questions.
  • 9:01 - 9:04
    When this weapon
    is in our tool kit, we use it.
  • 9:05 - 9:07
    Random assignment allows us
    to claim ceteris
  • 9:07 - 9:10
    has indeed been made paribus.
  • 9:10 - 9:13
    But randomized trials
    could be tough to organize.
  • 9:13 - 9:15
    They may be expensive,
    time-consuming,
  • 9:15 - 9:18
    and, in some cases,
    considered unethical.
  • 9:18 - 9:21
    So masters look
    for convincing alternatives.
  • 9:21 - 9:25
    The alternatives try to mimic
    the causality revealing power
  • 9:25 - 9:28
    of a randomized trial
    but without the time, trouble,
  • 9:28 - 9:31
    and expense
    of a purpose-built experiment.
  • 9:31 - 9:35
    These alternative tools are applied
    in real world scenarios
  • 9:35 - 9:37
    that mimic random assignment.
  • 9:39 - 9:43
    - [Narrator] You're on your way
    to mastering Econometrics.
  • 9:43 - 9:44
    Make sure this video sticks
  • 9:44 - 9:47
    by taking a few
    quick practice questions.
  • 9:47 - 9:51
    Or, if you're ready,
    click for the next video.
  • 9:51 - 9:54
    You can also check out MRU's website
  • 9:54 - 9:57
    for more courses,
    teacher resources, and more.
Title:
Randomized Trials: The Ideal Weapon
Description:

MIT’s Josh Angrist—aka Master Joshway—introduces us to our most powerful weapon: randomized trials!

Randomized trials originate in medical research, where they’re called “randomized clinical trials” or “RCTs.” That’s why randomized trials are said to measure “treatment effects”.

Josh covers a fascinating study from West Point that seeks to answer a common question using randomized trials: Are our devices more distracting than useful when it comes to learning?

In this video, we cover the following:
-What's the difference between control and treatment groups?
-How to "check for balance"
-The Law of Large Numbers
-Treatment effects
-Standard errors and statistical significance

***INSTRUCTOR RESOURCES***
High school teacher resources: https://bit.ly/2RcCsta
Professor resources: https://bit.ly/2XgdHQG
EconInbox: https://bit.ly/2yzTjzz

***MORE LEARNING***
Try out our practice questions: https://bit.ly/2xVuvSg
See the full course: https://bit.ly/2Xc88CQ
Receive updates when we release new videos: https://bit.ly/2VaLTdH
More from Marginal Revolution University: https://bit.ly/39QeWZo

more » « less
Video Language:
English
Team:
Marginal Revolution University
Project:
Mastering Econometrics
Duration:
10:01

English subtitles

Revisions Compare revisions