< Return to Video

Making a car for blind drivers

  • 0:00 - 0:03
    Many believe driving is an activity
  • 0:03 - 0:05
    solely reserved for those who can see.
  • 0:05 - 0:08
    A blind person driving a vehicle safely and independently
  • 0:08 - 0:11
    was thought to be an impossible task, until now.
  • 0:11 - 0:13
    Hello, my name is Dennis Hong,
  • 0:13 - 0:15
    and we're bringing freedom and independence to the blind
  • 0:15 - 0:18
    by building a vehicle for the visually impaired.
  • 0:18 - 0:21
    So before I talk about this car for the blind,
  • 0:21 - 0:23
    let me briefly tell you about another project that I worked on
  • 0:23 - 0:25
    called the DARPA Urban Challenge.
  • 0:25 - 0:27
    Now this was about building a robotic car
  • 0:27 - 0:29
    that can drive itself.
  • 0:29 - 0:31
    You press start, nobody touches anything,
  • 0:31 - 0:34
    and it can reach its destination fully autonomously.
  • 0:34 - 0:37
    So in 2007, our team won half a million dollars
  • 0:37 - 0:39
    by placing third place in this competition.
  • 0:39 - 0:41
    So about that time,
  • 0:41 - 0:43
    the National Federation of the Blind, or NFB,
  • 0:43 - 0:45
    challenged the research committee
  • 0:45 - 0:47
    about who can develop a car
  • 0:47 - 0:49
    that lets a blind person drive safely and independently.
  • 0:49 - 0:51
    We decided to give it a try,
  • 0:51 - 0:53
    because we thought, "Hey, how hard could it be?"
  • 0:53 - 0:55
    We have already an autonomous vehicle.
  • 0:55 - 0:57
    We just put a blind person in it and we're done, right?
  • 0:57 - 0:59
    (Laughter)
  • 0:59 - 1:01
    We couldn't have been more wrong.
  • 1:01 - 1:03
    What NFB wanted
  • 1:03 - 1:06
    was not a vehicle that can drive a blind person around,
  • 1:06 - 1:09
    but a vehicle where a blind person can make active decisions and drive.
  • 1:09 - 1:11
    So we had to throw everything out the window
  • 1:11 - 1:13
    and start from scratch.
  • 1:13 - 1:15
    So to test this crazy idea,
  • 1:15 - 1:17
    we developed a small dune buggy prototype vehicle
  • 1:17 - 1:19
    to test the feasibility.
  • 1:19 - 1:21
    And in the summer of 2009,
  • 1:21 - 1:24
    we invited dozens of blind youth from all over the country
  • 1:24 - 1:26
    and gave them a chance to take it for a spin.
  • 1:26 - 1:28
    It was an absolutely amazing experience.
  • 1:28 - 1:30
    But the problem with this car was
  • 1:30 - 1:33
    it was designed to only be driven in a very controlled environment,
  • 1:33 - 1:35
    in a flat, closed-off parking lot --
  • 1:35 - 1:37
    even the lanes defined by red traffic cones.
  • 1:37 - 1:39
    So with this success,
  • 1:39 - 1:41
    we decided to take the next big step,
  • 1:41 - 1:44
    to develop a real car that can be driven on real roads.
  • 1:44 - 1:46
    So how does it work?
  • 1:46 - 1:48
    Well, it's a rather complex system,
  • 1:48 - 1:51
    but let me try to explain it, maybe simplify it.
  • 1:51 - 1:53
    So we have three steps.
  • 1:53 - 1:55
    We have perception, computation
  • 1:55 - 1:57
    and non-visual interfaces.
  • 1:57 - 1:59
    Now obviously the driver cannot see,
  • 1:59 - 2:01
    so the system needs to perceive the environment
  • 2:01 - 2:03
    and gather information for the driver.
  • 2:03 - 2:06
    For that, we use an initial measurement unit.
  • 2:06 - 2:08
    So it measures acceleration, angular acceleration --
  • 2:08 - 2:10
    like a human ear, inner ear.
  • 2:10 - 2:12
    We fuse that information with a GPS unit
  • 2:12 - 2:15
    to get an estimate of the location of the car.
  • 2:15 - 2:18
    We also use two cameras to detect the lanes of the road.
  • 2:18 - 2:20
    And we also use three laser range finders.
  • 2:20 - 2:23
    The lasers scan the environment to detect obstacles --
  • 2:23 - 2:25
    a car approaching from the front, the back
  • 2:25 - 2:28
    and also any obstacles that run into the roads,
  • 2:28 - 2:30
    any obstacles around the vehicle.
  • 2:30 - 2:33
    So all this vast amount of information is then fed into the computer,
  • 2:33 - 2:35
    and the computer can do two things.
  • 2:35 - 2:38
    One is, first of all, process this information
  • 2:38 - 2:40
    to have an understanding of the environment --
  • 2:40 - 2:43
    these are the lanes of the road, there's the obstacles --
  • 2:43 - 2:45
    and convey this information to the driver.
  • 2:45 - 2:47
    The system is also smart enough
  • 2:47 - 2:49
    to figure out the safest way to operate the car.
  • 2:49 - 2:51
    So we can also generate instructions
  • 2:51 - 2:53
    on how to operate the controls of the vehicle.
  • 2:53 - 2:55
    But the problem is this: How do we convey
  • 2:55 - 2:57
    this information and instructions
  • 2:57 - 2:59
    to a person who cannot see
  • 2:59 - 3:02
    fast enough and accurate enough so he can drive?
  • 3:02 - 3:04
    So for this, we developed many different types
  • 3:04 - 3:07
    of non-visual user interface technology.
  • 3:07 - 3:09
    So starting from a three-dimensional ping sound system,
  • 3:09 - 3:11
    a vibrating vest,
  • 3:11 - 3:14
    a click wheel with voice commands, a leg strip,
  • 3:14 - 3:16
    even a shoe that applies pressure to the foot.
  • 3:16 - 3:18
    But today we're going to talk about
  • 3:18 - 3:20
    three of these non-visual user interfaces.
  • 3:20 - 3:23
    Now the first interface is called a DriveGrip.
  • 3:23 - 3:25
    So these are a pair of gloves,
  • 3:25 - 3:27
    and it has vibrating elements on the knuckle part
  • 3:27 - 3:30
    so you can convey instructions about how to steer --
  • 3:30 - 3:32
    the direction and the intensity.
  • 3:32 - 3:34
    Another device is called SpeedStrip.
  • 3:34 - 3:37
    So this is a chair -- as a matter of fact, it's actually a massage chair.
  • 3:37 - 3:41
    We gut it out, and we rearrange the vibrating elements in different patterns,
  • 3:41 - 3:44
    and we actuate them to convey information about the speed,
  • 3:44 - 3:47
    and also instructions how to use the gas and the brake pedal.
  • 3:47 - 3:49
    So over here, you can see
  • 3:49 - 3:51
    how the computer understands the environment,
  • 3:51 - 3:53
    and because you cannot see the vibration,
  • 3:53 - 3:56
    we actually put red LED's on the driver so that you can see what's happening.
  • 3:56 - 3:58
    This is the sensory data,
  • 3:58 - 4:01
    and that data is transferred to the devices through the computer.
  • 4:01 - 4:03
    So these two devices, DriveGrip and SpeedStrip,
  • 4:03 - 4:05
    are very effective.
  • 4:05 - 4:07
    But the problem is
  • 4:07 - 4:09
    these are instructional cue devices.
  • 4:09 - 4:11
    So this is not really freedom, right?
  • 4:11 - 4:13
    The computer tells you how to drive --
  • 4:13 - 4:15
    turn left, turn right, speed up, stop.
  • 4:15 - 4:17
    We call this the "backseat-driver problem."
  • 4:17 - 4:20
    So we're moving away from the instructional cue devices,
  • 4:20 - 4:22
    and we're now focusing more
  • 4:22 - 4:24
    on the informational devices.
  • 4:24 - 4:26
    A good example for this informational non-visual user interface
  • 4:26 - 4:28
    is called AirPix.
  • 4:28 - 4:30
    So think of it as a monitor for the blind.
  • 4:30 - 4:32
    So it's a small tablet, has many holes in it,
  • 4:32 - 4:34
    and compressed air comes out,
  • 4:34 - 4:36
    so it can actually draw images.
  • 4:36 - 4:38
    So even though you are blind, you can put your hand over it,
  • 4:38 - 4:40
    you can see the lanes of the road and obstacles.
  • 4:40 - 4:43
    Actually, you can also change the frequency of the air coming out
  • 4:43 - 4:45
    and possibly the temperature.
  • 4:45 - 4:48
    So it's actually a multi-dimensional user interface.
  • 4:48 - 4:51
    So here you can see the left camera, the right camera from the vehicle
  • 4:51 - 4:54
    and how the computer interprets that and sends that information to the AirPix.
  • 4:54 - 4:56
    For this, we're showing a simulator,
  • 4:56 - 4:59
    a blind person driving using the AirPix.
  • 4:59 - 5:02
    This simulator was also very useful for training the blind drivers
  • 5:02 - 5:04
    and also quickly testing different types of ideas
  • 5:04 - 5:06
    for different types of non-visual user interfaces.
  • 5:06 - 5:08
    So basically that's how it works.
  • 5:08 - 5:10
    So just a month ago,
  • 5:10 - 5:12
    on January 29th,
  • 5:12 - 5:14
    we unveiled this vehicle for the very first time to the public
  • 5:14 - 5:17
    at the world-famous Daytona International Speedway
  • 5:17 - 5:19
    during the Rolex 24 racing event.
  • 5:19 - 5:22
    We also had some surprises. Let's take a look.
  • 5:22 - 5:32
    (Music)
  • 5:32 - 5:36
    (Video) Announcer: This is an historic day in January.
  • 5:36 - 5:40
    He's coming up to the grandstand, fellow Federationists.
  • 5:40 - 5:46
    (Cheering)
  • 5:46 - 5:49
    (Honking)
  • 5:49 - 5:51
    There's the grandstand now.
  • 5:51 - 5:55
    And he's [unclear] following that van that's out in front of him.
  • 5:55 - 5:57
    Well there comes the first box.
  • 5:57 - 6:00
    Now let's see if Mark avoids it.
  • 6:00 - 6:03
    He does. He passes it on the right.
  • 6:05 - 6:08
    Third box is out. The fourth box is out.
  • 6:08 - 6:11
    And he's perfectly making his way between the two.
  • 6:11 - 6:13
    He's closing in on the van
  • 6:13 - 6:16
    to make the moving pass.
  • 6:17 - 6:19
    Well this is what it's all about,
  • 6:19 - 6:23
    this kind of dynamic display of audacity and ingenuity.
  • 6:24 - 6:27
    He's approaching the end of the run,
  • 6:27 - 6:32
    makes his way between the barrels that are set up there.
  • 6:32 - 6:35
    (Honking)
  • 6:35 - 6:38
    (Applause)
  • 6:41 - 6:43
    Dennis Hong: I'm so happy for you.
  • 6:43 - 6:45
    Mark's going to give me a ride back to the hotel.
  • 6:45 - 6:47
    Mark Riccobono: Yes.
  • 6:50 - 6:59
    (Applause)
  • 6:59 - 7:01
    DH: So since we started this project,
  • 7:01 - 7:04
    we've been getting hundreds of letters, emails, phone calls
  • 7:04 - 7:06
    from people from all around the world.
  • 7:06 - 7:09
    Letters thanking us, but sometimes you also get funny letters like this one:
  • 7:09 - 7:13
    "Now I understand why there is Braille on a drive-up ATM machine."
  • 7:13 - 7:15
    (Laughter)
  • 7:15 - 7:17
    But sometimes --
  • 7:17 - 7:19
    (Laughter)
  • 7:19 - 7:21
    But sometimes I also do get --
  • 7:21 - 7:23
    I wouldn't call it hate mail --
  • 7:23 - 7:25
    but letters of really strong concern:
  • 7:25 - 7:27
    "Dr. Hong, are you insane,
  • 7:27 - 7:29
    trying to put blind people on the road?
  • 7:29 - 7:31
    You must be out of your mind."
  • 7:31 - 7:33
    But this vehicle is a prototype vehicle,
  • 7:33 - 7:35
    and it's not going to be on the road
  • 7:35 - 7:37
    until it's proven as safe as, or safer than, today's vehicle.
  • 7:37 - 7:40
    And I truly believe that this can happen.
  • 7:40 - 7:42
    But still, will the society,
  • 7:42 - 7:44
    would they accept such a radical idea?
  • 7:44 - 7:46
    How are we going to handle insurance?
  • 7:46 - 7:48
    How are we going to issue driver's licenses?
  • 7:48 - 7:51
    There's many of these different kinds of hurdles besides technology challenges
  • 7:51 - 7:54
    that we need to address before this becomes a reality.
  • 7:54 - 7:56
    Of course, the main goal of this project
  • 7:56 - 7:58
    is to develop a car for the blind.
  • 7:58 - 8:00
    But potentially more important than this
  • 8:00 - 8:03
    is the tremendous value of the spin-off technology
  • 8:03 - 8:05
    that can come from this project.
  • 8:05 - 8:07
    The sensors that are used can see through the dark,
  • 8:07 - 8:09
    the fog and rain.
  • 8:09 - 8:11
    And together with this new type of interfaces,
  • 8:11 - 8:13
    we can use these technologies
  • 8:13 - 8:15
    and apply them to safer cars for sighted people.
  • 8:15 - 8:18
    Or for the blind, everyday home appliances --
  • 8:18 - 8:20
    in the educational setting, in the office setting.
  • 8:20 - 8:23
    Just imagine, in a classroom a teacher writes on the blackboard
  • 8:23 - 8:26
    and a blind student can see what's written and read
  • 8:26 - 8:28
    using these non-visual interfaces.
  • 8:28 - 8:31
    This is priceless.
  • 8:31 - 8:34
    So today, the things I've showed you today, is just the beginning.
  • 8:34 - 8:36
    Thank you very much.
  • 8:36 - 8:47
    (Applause)
Title:
Making a car for blind drivers
Speaker:
Dennis Hong
Description:

Using robotics, laser rangefinders, GPS and smart feedback tools, Dennis Hong is building a car for drivers who are blind. It's not a "self-driving" car, he's careful to note, but a car in which a non-sighted driver can determine speed, proximity and route -- and drive independently.

more » « less
Video Language:
English
Team:
closed TED
Project:
TEDTalks
Duration:
08:48
TED edited English subtitles for Making a car for blind drivers
TED added a translation

English subtitles

Revisions Compare revisions