< Return to Video

The birth of Wikipedia

  • 0:01 - 0:04
    Charles Van Doren, who was later
    a senior editor of Britannica,
  • 0:04 - 0:07
    said the ideal encyclopedia
    should be radical --
  • 0:07 - 0:08
    it should stop being safe.
  • 0:08 - 0:12
    But if you know anything about the history
    of Britannica since 1962,
  • 0:12 - 0:13
    it was anything but radical:
  • 0:13 - 0:18
    still a very completely safe,
    stodgy type of encyclopedia.
  • 0:18 - 0:22
    Wikipedia, on the other hand,
    begins with a very radical idea,
  • 0:22 - 0:24
    and that's for all of us
    to imagine a world
  • 0:24 - 0:26
    in which every single person on the planet
  • 0:26 - 0:29
    is given free access
    to the sum of all human knowledge.
  • 0:29 - 0:30
    And that's what we're doing.
  • 0:31 - 0:32
    So Wikipedia --
  • 0:32 - 0:35
    you just saw the little
    demonstration of it --
  • 0:35 - 0:37
    it's a freely licensed encyclopedia.
  • 0:37 - 0:39
    It's written by thousands of volunteers
  • 0:39 - 0:41
    all over the world
    in many, many languages.
  • 0:41 - 0:43
    It's written using wiki software --
  • 0:43 - 0:46
    which is the type of software
    he just demonstrated --
  • 0:46 - 0:47
    so anyone can quickly edit and save,
  • 0:47 - 0:49
    and it goes live
    on the Internet immediately.
  • 0:51 - 0:55
    And everything about Wikipedia is managed
    by virtually an all-volunteer staff.
  • 0:55 - 1:00
    So when Yochai is talking
    about new methods of organization,
  • 1:00 - 1:02
    he's exactly describing Wikipedia.
  • 1:02 - 1:04
    And what I'm going to do today
  • 1:04 - 1:07
    is tell you a little bit more
    about how it really works on the inside.
  • 1:08 - 1:11
    So Wikipedia's owned
    by the Wikimedia Foundation,
  • 1:11 - 1:12
    which I founded,
  • 1:12 - 1:14
    a nonprofit organization.
  • 1:14 - 1:18
    And our goal, the core aim
    of the Wikimedia Foundation,
  • 1:18 - 1:21
    is to get a free encyclopedia
    to every single person on the planet.
  • 1:21 - 1:23
    And so, if you think
    about what that means,
  • 1:23 - 1:26
    it means a lot more
    than just building a cool website.
  • 1:26 - 1:29
    We're really interested
    in all the issues of the digital divide,
  • 1:29 - 1:30
    poverty worldwide,
  • 1:30 - 1:34
    empowering people everywhere
    to have the information that they need
  • 1:34 - 1:35
    to make good decisions.
  • 1:35 - 1:37
    And so we're going
    to have to do a lot of work
  • 1:38 - 1:39
    that goes beyond just the Internet.
  • 1:39 - 1:43
    And so that's a big part of why
    we've chosen the free licensing model,
  • 1:43 - 1:46
    because that empowers local
    entrepreneurs or anyone who wants to --
  • 1:46 - 1:49
    they can take our content
    and do anything they like with it --
  • 1:49 - 1:50
    you can copy it, redistribute it --
  • 1:50 - 1:53
    and you can do it
    commercially or non-commercially.
  • 1:53 - 1:56
    So there's a lot of opportunities
    that are going to arise around Wikipedia
  • 1:56 - 1:58
    all over the world.
  • 1:58 - 2:00
    We're funded by donations from the public,
  • 2:01 - 2:03
    and one of the more
    interesting things about that
  • 2:03 - 2:07
    is how little money
    it actually takes to run Wikipedia.
  • 2:07 - 2:11
    So Yochai showed you the graph
    of what the cost of a printing press was.
  • 2:11 - 2:15
    And I'm going to tell you
    what the cost of Wikipedia is.
  • 2:15 - 2:17
    But first, I'll show you how big it is.
  • 2:17 - 2:21
    So we've got over 600,000
    articles in English.
  • 2:21 - 2:25
    We've got two million total articles
    across many, many different languages.
  • 2:26 - 2:29
    The biggest languages
    are German, Japanese, French --
  • 2:29 - 2:32
    all the Western-European
    languages are quite big.
  • 2:33 - 2:36
    But only around one-third
    of all of our traffic to our web
  • 2:36 - 2:38
    clusters to the English Wikipedia,
  • 2:38 - 2:39
    which is surprising to a lot of people.
  • 2:39 - 2:43
    A lot of people think in a very
    English-centric way on the Internet,
  • 2:43 - 2:45
    but for us, we're truly global.
  • 2:45 - 2:46
    We're in many, many languages.
  • 2:47 - 2:50
    How popular we've gotten to be --
    we're a top-50 website
  • 2:50 - 2:52
    and we're more popular
    than the New York Times.
  • 2:52 - 2:56
    So this is where we get
    to Yochai's discussion.
  • 2:57 - 3:00
    This shows the growth of Wikipedia --
    we're the blue line there --
  • 3:00 - 3:03
    and this is the New York Times over there.
  • 3:03 - 3:05
    And what's interesting about this
  • 3:05 - 3:09
    is the New York Times website is a huge,
    enormous corporate operation
  • 3:09 - 3:11
    with I have no idea
    how many hundreds of employees.
  • 3:12 - 3:14
    We have exactly one employee,
  • 3:14 - 3:16
    and that employee
    is our lead software developer.
  • 3:16 - 3:19
    And he's only been our employee
    since January 2005,
  • 3:19 - 3:21
    all the other growth before that ...
  • 3:21 - 3:25
    So the servers are managed
    by a ragtag band of volunteers.
  • 3:25 - 3:27
    All the editing is done by volunteers.
  • 3:27 - 3:28
    And the way that we're organized
  • 3:28 - 3:31
    is not like any traditional
    organization you can imagine.
  • 3:31 - 3:34
    People are always asking,
    "Well, who's in charge of this?"
  • 3:34 - 3:35
    or "Who does that?"
  • 3:35 - 3:37
    And the answer is:
    anybody who wants to pitch in.
  • 3:37 - 3:41
    It's a very unusual and chaotic thing.
  • 3:41 - 3:44
    We've got over 90 servers now
    in three locations.
  • 3:44 - 3:49
    These are managed by volunteer
    system administrators who are online.
  • 3:49 - 3:51
    I can go online
    any time of the day or night
  • 3:51 - 3:55
    and see eight to 10 people waiting for me
  • 3:55 - 3:59
    to ask a question or something,
    anything about the servers.
  • 3:59 - 4:01
    You could never afford
    to do this in a company.
  • 4:01 - 4:03
    You could never afford
  • 4:03 - 4:06
    to have a standby crew of people
    24 hours a day
  • 4:06 - 4:09
    and do what we're doing at Wikipedia.
  • 4:09 - 4:13
    So we're doing around
    1.4 billion page views monthly,
  • 4:13 - 4:15
    so it's really gotten to be a huge thing.
  • 4:15 - 4:18
    And everything is managed
    by the volunteers.
  • 4:18 - 4:22
    And the total monthly cost
    for our bandwidth is about 5,000 dollars.
  • 4:22 - 4:24
    And that's essentially our main cost.
  • 4:24 - 4:26
    We could actually do without the employee.
  • 4:26 - 4:31
    We hired Brian because he was
    working part-time for two years
  • 4:31 - 4:32
    and full-time at Wikipedia,
  • 4:32 - 4:33
    so we actually hired him,
  • 4:33 - 4:36
    so he could get a life
    and go to the movies sometimes.
  • 4:37 - 4:40
    So the big question when you've got
    this really chaotic organization is,
  • 4:40 - 4:42
    why isn't it all rubbish?
  • 4:42 - 4:44
    Why is the website as good as it is?
  • 4:44 - 4:45
    First of all, how good is it?
  • 4:45 - 4:47
    Well, it's pretty good.
  • 4:47 - 4:48
    It isn't perfect,
  • 4:48 - 4:50
    but it's much better
    than you would expect,
  • 4:50 - 4:52
    given our completely chaotic model.
  • 4:52 - 4:55
    So when you saw him make
    a ridiculous edit to the page about me,
  • 4:55 - 4:59
    you think, "Oh, this is obviously
    just going to degenerate into rubbish."
  • 4:59 - 5:00
    But when we've seen quality tests --
  • 5:01 - 5:03
    and there haven't been enough of these yet
  • 5:03 - 5:05
    and I'm really encouraging
    people to do more,
  • 5:05 - 5:07
    comparing Wikipedia
    to traditional things --
  • 5:07 - 5:09
    we win hands down.
  • 5:09 - 5:12
    So a German magazine
    compared German Wikipedia,
  • 5:12 - 5:15
    which is much, much smaller than English,
  • 5:15 - 5:19
    to Microsoft Encarta
    and to Brockhaus multimedial,
  • 5:19 - 5:21
    and we won across the board.
  • 5:21 - 5:24
    They hired experts to come and look
    at articles and compare the quality,
  • 5:24 - 5:27
    and we were very pleased with that result.
  • 5:27 - 5:33
    So a lot of people have heard about
    the Wikipedia Bush-Kerry controversy.
  • 5:33 - 5:36
    The media has covered this
    somewhat extensively.
  • 5:36 - 5:38
    It started out
    with an article in Red Herring.
  • 5:38 - 5:40
    The reporters called me up and they --
  • 5:40 - 5:43
    I mean, I have to say
    they spelled my name right,
  • 5:43 - 5:48
    but they really wanted to say
    the Bush-Kerry election is so contentious,
  • 5:48 - 5:51
    it's tearing apart
    the Wikipedia community.
  • 5:51 - 5:52
    And so they quote me as saying,
  • 5:52 - 5:55
    "They're the most contentious
    in the history of Wikipedia."
  • 5:55 - 5:58
    What I actually said
    is they're not contentious at all.
  • 5:58 - 6:00
    So it's a slight misquote.
  • 6:00 - 6:02
    (Laughter)
  • 6:02 - 6:04
    The articles were edited quite heavily.
  • 6:04 - 6:06
    And it is true that we did
    have to lock the articles
  • 6:06 - 6:07
    on a couple of occasions.
  • 6:07 - 6:10
    Time magazine recently reported that
  • 6:10 - 6:12
    "Extreme action sometimes has to be taken,
  • 6:12 - 6:17
    and Wales locked the entries
    on Kerry and Bush for most of 2004."
  • 6:17 - 6:21
    This came after I told the reporter
    that we had to lock it for --
  • 6:21 - 6:23
    occasionally a little bit here and there.
  • 6:23 - 6:27
    So the truth in general
    is that the kinds of controversies
  • 6:27 - 6:30
    that you would probably think
    we have within the Wikipedia community
  • 6:30 - 6:33
    are not really controversies at all.
  • 6:33 - 6:35
    Articles on controversial topics
    are edited a lot,
  • 6:35 - 6:38
    but they don't cause much controversy
    within the community.
  • 6:38 - 6:42
    And the reason for this is that most
    people understand the need for neutrality.
  • 6:43 - 6:49
    The real struggle is not
    between the right and the left --
  • 6:49 - 6:51
    that's where most people assume --
  • 6:51 - 6:53
    but it's between
    the party of the thoughtful
  • 6:53 - 6:54
    and the party of the jerks.
  • 6:54 - 6:56
    And no side of the political spectrum
  • 6:56 - 6:59
    has a monopoly
    on either of those qualities.
  • 6:59 - 7:01
    The actual truth
    about the specific Bush-Kerry incident
  • 7:01 - 7:04
    is that the Bush-Kerry articles
  • 7:04 - 7:07
    were locked less than one percent
    of the time in 2004,
  • 7:07 - 7:09
    and it wasn't because
    they were contentious;
  • 7:09 - 7:12
    it was just because
    there was routine vandalism --
  • 7:12 - 7:15
    which happens sometimes even on stage ...
  • 7:15 - 7:17
    (Laughter)
  • 7:17 - 7:20
    Sometimes even reporters have reported
    to me that they vandalized Wikipedia
  • 7:20 - 7:22
    and were amazed
    that it was fixed so quickly.
  • 7:23 - 7:25
    And I said -- you know, I always say,
    please don't do that.
  • 7:25 - 7:27
    That's not a good thing.
  • 7:27 - 7:28
    So how do we do this?
  • 7:28 - 7:31
    How do we manage the quality control?
  • 7:31 - 7:33
    How does it work?
  • 7:34 - 7:36
    So there's a few elements,
  • 7:36 - 7:39
    mostly social policies
    and some elements of the software.
  • 7:39 - 7:42
    So the biggest
    and the most important thing
  • 7:42 - 7:44
    is our neutral point of view policy.
  • 7:44 - 7:46
    This is something
    that I set down, from the very beginning,
  • 7:46 - 7:50
    as a core principle of the community
    that's completely not debatable.
  • 7:50 - 7:53
    It's a social concept of cooperation,
  • 7:53 - 7:57
    so we don't talk a lot
    about truth and objectivity.
  • 7:57 - 7:59
    The reason for this is if we say
  • 7:59 - 8:02
    we're only going to write
    the "truth" about some topic,
  • 8:02 - 8:05
    that doesn't do us a damn bit of good
    of figuring out what to write,
  • 8:05 - 8:07
    because I don't agree with you
    about what's the truth.
  • 8:07 - 8:10
    But we have this
    jargon term of neutrality,
  • 8:10 - 8:12
    which has its own long history
    within the community,
  • 8:12 - 8:16
    which basically says,
    any time there's a controversial issue,
  • 8:16 - 8:18
    Wikipedia itself should not
    take a stand on the issue.
  • 8:18 - 8:22
    We should merely report on what
    reputable parties have said about it.
  • 8:22 - 8:25
    So this neutrality policy
    is really important for us
  • 8:25 - 8:28
    because it empowers a community
    that is very diverse
  • 8:28 - 8:30
    to come together
    and actually get some work done.
  • 8:30 - 8:32
    So we have very diverse contributors
  • 8:32 - 8:36
    in terms of political,
    religious, cultural backgrounds.
  • 8:36 - 8:37
    By having this firm neutrality policy,
  • 8:37 - 8:39
    which is non-negotiable
    from the beginning,
  • 8:40 - 8:41
    we ensure that people can work together
  • 8:41 - 8:44
    and that the entries
    don't become simply a war
  • 8:44 - 8:46
    back and forth
    between the left and the right.
  • 8:46 - 8:50
    If you engage in that type of behavior,
    you'll be asked to leave the community.
  • 8:51 - 8:53
    So, real-time peer review.
  • 8:53 - 8:56
    Every single change on the site
    goes to the "Recent changes" page.
  • 8:56 - 8:59
    So as soon as he made his change,
    it went to the "Recent changes" page.
  • 8:59 - 9:03
    That recent changes page
    was also fed into an IRC channel,
  • 9:03 - 9:05
    which is an Internet chat channel
  • 9:05 - 9:08
    that people are monitoring
    with various software tools.
  • 9:09 - 9:11
    And people can get RSS feeds --
  • 9:11 - 9:14
    they can get email
    notifications of changes.
  • 9:15 - 9:17
    And then users can set up
    their own personal watch list.
  • 9:17 - 9:20
    So my page is on quite a few
    volunteers' watch lists,
  • 9:20 - 9:21
    because it is sometimes vandalized.
  • 9:21 - 9:27
    And therefore, what happens is someone
    will notice the change very quickly,
  • 9:27 - 9:31
    and then they'll just
    simply revert the change.
  • 9:31 - 9:33
    There's a "new pages feed," for example,
  • 9:33 - 9:35
    so you can go
    to a certain page of Wikipedia
  • 9:35 - 9:37
    and see every new page as it's created.
  • 9:37 - 9:38
    This is really important
  • 9:38 - 9:41
    because a lot of new pages
    are just garbage that has to be deleted,
  • 9:41 - 9:42
    you know, "ASDFASDF."
  • 9:42 - 9:45
    But also, that's some
    of the most interesting and fun things,
  • 9:45 - 9:47
    some of the new articles.
  • 9:47 - 9:49
    People will start an article
    on some interesting topic,
  • 9:49 - 9:51
    other people will find that intriguing
  • 9:51 - 9:53
    and jump in and help
    and make it much better.
  • 9:54 - 9:55
    So we do have edits by anonymous users,
  • 9:55 - 9:59
    which is one of the most controversial
    and intriguing things about Wikipedia.
  • 9:59 - 10:03
    So, Chris was able to do his change --
    he didn't have to log in or anything;
  • 10:03 - 10:06
    he just went on the website
    and made a change.
  • 10:06 - 10:08
    But it turns out
    that only about 18 percent
  • 10:08 - 10:09
    of all the edits to the website
  • 10:09 - 10:11
    are done by anonymous users.
  • 10:11 - 10:13
    And that's a really important
    thing to understand:
  • 10:13 - 10:16
    the vast majority of the edits
    that go on on the website
  • 10:16 - 10:20
    are from a very close-knit community
    of maybe 600 to 1,000 people
  • 10:20 - 10:22
    who are in constant communication.
  • 10:22 - 10:24
    And we have over 40 IRC channels,
    40 mailing lists.
  • 10:24 - 10:26
    All these people know each other.
  • 10:26 - 10:28
    They communicate.
    We have off-line meetings.
  • 10:28 - 10:31
    These are the people
    who are doing the bulk of the site,
  • 10:31 - 10:35
    and they are, in a sense,
    semi-professionals at what they're doing.
  • 10:36 - 10:38
    The standards we set for ourselves
  • 10:38 - 10:41
    are equal to or higher
    than professional standards of quality.
  • 10:41 - 10:45
    We don't always meet those standards,
    but that's what we're striving for.
  • 10:45 - 10:48
    And so that tight community
    is who really cares for the site,
  • 10:48 - 10:50
    and these are some
    of the smartest people I've ever met.
  • 10:50 - 10:53
    It's my job to say that,
    but it's actually true.
  • 10:53 - 10:56
    The type of people who were drawn
    to writing an encyclopedia for fun
  • 10:57 - 10:59
    tend to be pretty smart people.
  • 10:59 - 11:02
    The tools and the software:
    there's lots of tools that allow us --
  • 11:02 - 11:04
    allow us, meaning the community --
  • 11:04 - 11:06
    to self-monitor
    and to monitor all the work.
  • 11:06 - 11:08
    This is an example
    of a page history on "flat Earth,"
  • 11:08 - 11:10
    and you can see some changes
    that were made.
  • 11:10 - 11:14
    What's nice about this page
    is you can immediately take a look at this
  • 11:14 - 11:15
    and see, "OK, I understand now."
  • 11:15 - 11:17
    When somebody goes and looks at --
  • 11:17 - 11:19
    they see that someone,
    an anonymous IP number,
  • 11:19 - 11:21
    made an edit to my page.
  • 11:21 - 11:23
    That sounds suspicious.
    Who is this person?
  • 11:23 - 11:24
    Somebody looks at it --
  • 11:24 - 11:26
    they can immediately see
    highlighted in red
  • 11:26 - 11:28
    all of the changes that took place --
  • 11:28 - 11:32
    to see, OK, well, these words
    have changed, things like this.
  • 11:32 - 11:33
    So that's one tool that we can use
  • 11:33 - 11:36
    to very quickly monitor
    the history of a page.
  • 11:37 - 11:39
    Another thing that we do
    within the community
  • 11:39 - 11:42
    is we leave everything very open-ended.
  • 11:42 - 11:46
    Most of the social rules
    and the methods of work
  • 11:46 - 11:48
    are left completely open-ended
    in the software.
  • 11:48 - 11:50
    All of that stuff is just on Wiki pages.
  • 11:50 - 11:53
    And so there's nothing
    in the software that enforces the rules.
  • 11:53 - 11:57
    The example I've got up here
    is the Votes for Deletion page.
  • 11:57 - 12:00
    So, I mentioned earlier,
    people type "ASDFASDF" --
  • 12:00 - 12:01
    it needs to be deleted.
  • 12:01 - 12:04
    Cases like that,
    the administrators just delete it.
  • 12:04 - 12:06
    There's no reason
    to have a big argument about it.
  • 12:06 - 12:09
    But you can imagine there's a lot
    of other areas where the question is,
  • 12:09 - 12:12
    is this notable enough
    to go in an encyclopedia?
  • 12:12 - 12:15
    Is the information verifiable?
    Is it a hoax? Is it true? Is it what?
  • 12:15 - 12:18
    So we needed a social method
    for figuring out the answer to this.
  • 12:18 - 12:21
    And so the method that arose
    organically within the community
  • 12:21 - 12:23
    is the Votes For Deletion page.
  • 12:23 - 12:25
    And in the particular
    example we have here,
  • 12:25 - 12:26
    it's a film, "Twisted Issues,"
  • 12:26 - 12:28
    and the first person says,
  • 12:28 - 12:31
    "Now this is supposedly a film.
    It fails the Google test miserably."
  • 12:32 - 12:34
    The Google test is you look in Google
    and see if it's there,
  • 12:34 - 12:38
    because if something's not even in Google,
    it probably doesn't exist at all.
  • 12:38 - 12:40
    It's not a perfect rule,
  • 12:40 - 12:43
    but it's a nice starting point
    for quick research.
  • 12:43 - 12:46
    So somebody says, "Delete it, please.
    Delete it -- it's not notable."
  • 12:47 - 12:50
    And then somebody says,
    "Wait, I found it. I found it in a book,
  • 12:50 - 12:53
    'Film Threat Video Guide:
    the 20 Underground Films You Must See.'"
  • 12:53 - 12:55
    So the next persons says, "Clean it up."
  • 12:55 - 12:58
    Somebody says, "I've found it on IMDB.
    Keep, keep, keep."
  • 12:58 - 13:01
    And what's interesting about this
    is that the software is --
  • 13:01 - 13:05
    these votes are just text
    typed into a page.
  • 13:05 - 13:09
    This is not really a vote
    so much as it is a dialogue.
  • 13:09 - 13:11
    Now it is true that at the end of the day,
  • 13:11 - 13:15
    an administrator can go through here
    and take a look at this and say,
  • 13:15 - 13:17
    "OK, 18 deletes, two keeps:
    we'll delete it."
  • 13:17 - 13:21
    But in other cases,
    this could be 18 deletes and two keeps,
  • 13:21 - 13:24
    and we would keep it,
    because if those last two keeps say,
  • 13:24 - 13:27
    "Wait a minute. Nobody else saw this
    but I found it in a book,
  • 13:27 - 13:29
    and I found a link
    to a page that describes it,
  • 13:29 - 13:32
    and I'm going to clean it up tomorrow,
    so please don't delete it,"
  • 13:32 - 13:33
    then it would survive.
  • 13:33 - 13:36
    And it also matters
    who the people are who are voting.
  • 13:36 - 13:38
    Like I say, it's a tight-knit community.
  • 13:38 - 13:40
    Down here at the bottom,
    "Keep, real movie," RickK.
  • 13:40 - 13:44
    RickK is a very famous Wikipedian
    who does an enormous amount of work
  • 13:44 - 13:47
    with vandalism, hoaxes
    and votes for deletion.
  • 13:47 - 13:49
    His voice carries a lot of weight
    within the community
  • 13:49 - 13:51
    because he knows what he's doing.
  • 13:52 - 13:54
    So how is all this governed?
  • 13:54 - 13:57
    People really want to know
    about administrators, things like that.
  • 13:58 - 14:01
    So the Wikipedia governance model,
    the governance of the community,
  • 14:01 - 14:06
    is a very confusing,
    but workable mix of consensus --
  • 14:06 - 14:08
    meaning we try not to vote
    on the content of articles,
  • 14:08 - 14:12
    because the majority view
    is not necessarily neutral --
  • 14:12 - 14:14
    some amount of democracy --
  • 14:14 - 14:15
    all of the administrators --
  • 14:15 - 14:18
    these are the people
    who have the ability to delete pages.
  • 14:18 - 14:21
    That doesn't mean that they have
    the right to delete pages.
  • 14:21 - 14:25
    They still have to follow all the rules --
    but they're elected by the community.
  • 14:25 - 14:27
    Sometimes people --
    random trolls on the Internet --
  • 14:27 - 14:29
    like to accuse me
    of handpicking the administrators
  • 14:29 - 14:31
    to bias the content of the encyclopedia.
  • 14:31 - 14:35
    I always laugh at this, because I have
    no idea how they're elected, actually.
  • 14:35 - 14:37
    There's a certain amount of aristocracy.
  • 14:37 - 14:39
    You got a hint of that
    when I mentioned, like,
  • 14:39 - 14:41
    RickK's voice would carry
    a lot more weight
  • 14:41 - 14:43
    than someone we don't know.
  • 14:43 - 14:45
    I give this talk sometimes with Angela,
  • 14:46 - 14:48
    who was just re-elected to the board
    from the community --
  • 14:48 - 14:50
    to the Board of the Foundation,
  • 14:50 - 14:53
    with more than twice the votes
    of the person who didn't make it.
  • 14:53 - 14:56
    And I always embarrass her because I say,
  • 14:56 - 14:57
    "Well, Angela, for example,
  • 14:57 - 15:01
    could get away with doing
    absolutely anything within Wikipedia,
  • 15:01 - 15:03
    because she's so admired and so powerful."
  • 15:03 - 15:07
    But the irony is, of course, that Angela
    can do this because she's the one person
  • 15:07 - 15:10
    who you know would never, ever
    break any rules of Wikipedia.
  • 15:10 - 15:12
    And I also like to say
    she's the only person
  • 15:12 - 15:15
    who actually knows
    all the rules of Wikipedia, so ...
  • 15:15 - 15:16
    And then there's monarchy,
  • 15:16 - 15:18
    and that's my role
    on the community, so ...
  • 15:18 - 15:21
    (Laughter)
  • 15:21 - 15:23
    I was describing this in Berlin once,
  • 15:23 - 15:26
    and the next day
    in the newspaper the headline said,
  • 15:26 - 15:28
    "I am the Queen of England."
  • 15:28 - 15:29
    (Laughter)
  • 15:29 - 15:31
    And that's not exactly what I said, but --
  • 15:31 - 15:33
    (Laughter)
  • 15:33 - 15:35
    the point is my role in the community --
  • 15:35 - 15:40
    Within the free software world,
    there's been a long-standing tradition
  • 15:40 - 15:43
    of the "benevolent dictator" model.
  • 15:43 - 15:46
    So if you look at most
    of the major free software projects,
  • 15:46 - 15:48
    they have one single person in charge
  • 15:48 - 15:51
    who everyone agrees
    is the benevolent dictator.
  • 15:51 - 15:54
    Well, I don't like the term
    "benevolent dictator,"
  • 15:54 - 15:58
    and I don't think that it's my job
    or my role in the world of ideas
  • 15:58 - 16:01
    to be the dictator of the future
    of all human knowledge
  • 16:01 - 16:02
    compiled by the world.
  • 16:02 - 16:04
    It just isn't appropriate.
  • 16:04 - 16:07
    But there is a need still
    for a certain amount of monarchy,
  • 16:07 - 16:09
    a certain amount of --
  • 16:09 - 16:10
    sometimes we have to make a decision
  • 16:10 - 16:13
    and we don't want
    to get bogged down too heavily
  • 16:13 - 16:15
    in formal decision-making processes.
  • 16:15 - 16:20
    So as an example
    of how this can be important:
  • 16:20 - 16:24
    we recently had a situation where
    a neo-Nazi website discovered Wikipedia,
  • 16:24 - 16:27
    and they said,
    "Oh, well, this is horrible,
  • 16:27 - 16:29
    this Jewish conspiracy of a website,
  • 16:29 - 16:32
    and we're going to get certain articles
    deleted that we don't like.
  • 16:32 - 16:35
    And we see they have a voting process,
    so we're going to send --
  • 16:35 - 16:38
    we have 40,000 members
    and we're going to send them over
  • 16:38 - 16:41
    and they're all going to vote
    and get these pages deleted."
  • 16:41 - 16:43
    Well, they managed
    to get 18 people to show up.
  • 16:43 - 16:46
    That's neo-Nazi math for you.
  • 16:46 - 16:49
    They always think they've got
    40,000 members when they've got 18.
  • 16:49 - 16:53
    But they managed to get 18 people
    to come and vote in a fairly absurd way
  • 16:53 - 16:55
    to delete a perfectly valid article.
  • 16:55 - 16:58
    Of course, the vote ended up
    being about 85 to 18,
  • 16:58 - 17:01
    so there was no real danger
    to our democratic processes.
  • 17:01 - 17:04
    On the other hand, people said,
    "But what are we going to do?
  • 17:04 - 17:06
    I mean, this could happen.
  • 17:06 - 17:08
    What if some group
    gets really seriously organized
  • 17:08 - 17:11
    and comes in and wants to vote?"
  • 17:11 - 17:13
    Then I said, "Well, fuck it,
    we'll just change the rules."
  • 17:13 - 17:15
    That's my job in the community:
  • 17:15 - 17:19
    to say we won't allow
    our openness and freedom
  • 17:19 - 17:21
    to undermine the quality of the content.
  • 17:21 - 17:25
    And so, as long as people
    trust me in my role,
  • 17:25 - 17:26
    then that's a valid place for me.
  • 17:26 - 17:29
    Of course, because of the free licensing,
  • 17:29 - 17:32
    if I do a bad job, the volunteers
    are more than happy to take and leave --
  • 17:32 - 17:34
    I can't tell anyone what to do.
  • 17:34 - 17:39
    So the final point here
    is that to understand how Wikipedia works,
  • 17:39 - 17:42
    it's important to understand
    that our wiki model is the way we work,
  • 17:42 - 17:46
    but we are not fanatical web anarchists.
  • 17:46 - 17:49
    In fact, we're very flexible
    about the social methodology,
  • 17:49 - 17:51
    because ultimately,
    the passion of the community
  • 17:51 - 17:53
    is for the quality of the work,
  • 17:53 - 17:56
    not necessarily for the process
    that we use to generate it.
  • 17:56 - 17:58
    Thank you.
  • 17:58 - 18:02
    (Applause)
  • 18:02 - 18:04
    Ben Saunders: Yeah, hi, Ben Saunders.
  • 18:04 - 18:08
    Jimmy, you mentioned impartiality
    being a key to Wikipedia's success.
  • 18:08 - 18:13
    It strikes me that much of the textbooks
    that are used to educate our children
  • 18:13 - 18:15
    are inherently biased.
  • 18:15 - 18:19
    Have you found Wikipedia
    being used by teachers
  • 18:19 - 18:21
    and how do you see Wikipedia
    changing education?
  • 18:21 - 18:25
    Jimmy Wales: Yeah, so, a lot of teachers
    are beginning to use Wikipedia.
  • 18:25 - 18:30
    There's a media storyline about Wikipedia,
    which I think is false.
  • 18:30 - 18:34
    It builds on the storyline
    of bloggers versus newspapers.
  • 18:34 - 18:37
    And the storyline is,
    there's this crazy thing, Wikipedia,
  • 18:37 - 18:40
    but academics hate it
    and teachers hate it.
  • 18:40 - 18:43
    And that turns out to not be true.
  • 18:43 - 18:45
    The last time I got an email
    from a journalist saying,
  • 18:45 - 18:47
    "Why do academics hate Wikipedia?"
  • 18:47 - 18:49
    I sent it from my Harvard email address
  • 18:49 - 18:51
    because I was recently
    appointed a fellow there.
  • 18:51 - 18:53
    And I said, "Well,
    they don't all hate it."
  • 18:53 - 18:54
    (Laughter)
  • 18:54 - 18:57
    But I think there's going
    to be huge impacts.
  • 18:57 - 19:01
    And we actually have a project
    that I'm personally really excited about,
  • 19:01 - 19:02
    which is the Wikibooks project,
  • 19:02 - 19:05
    which is an effort to create
    textbooks in all the languages.
  • 19:05 - 19:07
    And that's a much bigger project.
  • 19:07 - 19:12
    It's going to take 20 years or so
    to come to fruition.
  • 19:12 - 19:14
    But part of that is to fulfill our mission
  • 19:14 - 19:17
    of giving an encyclopedia
    to every single person on the planet.
  • 19:17 - 19:20
    We don't mean we're going
    to Spam them with AOL-style CDs.
  • 19:20 - 19:22
    We mean we're going to give them
    a tool that they can use.
  • 19:22 - 19:24
    And for a lot of people in the world,
  • 19:24 - 19:27
    if I give you an encyclopedia
    that's written at a university level,
  • 19:27 - 19:31
    it doesn't do you any good
    without a whole host of literacy materials
  • 19:31 - 19:34
    to build you up to the point
    where you can actually use it.
  • 19:34 - 19:36
    The Wikibooks project
    is an effort to do that.
  • 19:36 - 19:39
    And I think that we're going to see --
    it may not even come from us;
  • 19:39 - 19:41
    there's all kinds of innovation going on.
  • 19:41 - 19:44
    But freely licensed textbooks
    are the next big thing in education.
Title:
The birth of Wikipedia
Speaker:
Jimmy Wales
Description:

Jimmy Wales recalls how he assembled "a ragtag band of volunteers," gave them tools for collaborating and created Wikipedia, the self-organizing, self-correcting, never-finished online encyclopedia.

more » « less
Video Language:
English
Team:
closed TED
Project:
TEDTalks
Duration:
19:45
Krystian Aparta commented on English subtitles for The birth of Wikipedia
Krystian Aparta edited English subtitles for The birth of Wikipedia
Krystian Aparta edited English subtitles for The birth of Wikipedia
Joanna Pietrulewicz edited English subtitles for The birth of Wikipedia
Joanna Pietrulewicz edited English subtitles for The birth of Wikipedia
Joanna Pietrulewicz edited English subtitles for The birth of Wikipedia
TED edited English subtitles for The birth of Wikipedia
TED edited English subtitles for The birth of Wikipedia
Show all
  • The English transcript was updated on 10/6/2015. Throughout the transcript, "Rick Kay" was changed to "RickK."

English subtitles

Revisions Compare revisions