Return to Video

Sam Harris: Can Science Determine Human Values?

  • 0:04 - 0:08
    Fora TV. The world is thinking.
  • 0:11 - 0:13
    In partnership with Berkeley Arts & Letters
  • 0:15 - 0:19
    I'm always tempted when I see an audience like this to say
  • 0:19 - 0:22
    "Sam Harris needs no introduction", and then sit down.
  • 0:24 - 0:28
    But... you're such a weird group, I know that
  • 0:29 - 0:33
    But maybe I'll say something to try to find out why you are here
  • 0:34 - 0:37
    Of the "three horsemen"
  • 0:38 - 0:42
    or the Great Trinity of those people who've been called "The new atheists"
  • 0:42 - 0:46
    we have had or tried to have all three of them here:
  • 0:46 - 0:48
    Cristopher Hitchens got sick and had to cancel,
  • 0:49 - 0:52
    Richard Dawson was here, filled the place,
  • 0:52 - 0:56
    and lectured for an hour about evolution.
  • 0:56 - 1:00
    In Berkeley; we had to go out and try to find somebody in Berkeley
  • 1:00 - 1:02
    who didn't believe in evolution.
  • 1:03 - 1:07
    So, I'm expecting more from Sam Harris.
  • 1:07 - 1:13
    Sam Harris' credentials are so numerous:
  • 1:13 - 1:17
    he, after an experience with ecstasy,
  • 1:17 - 1:22
    he dropped out of college and decided to try to find
  • 1:22 - 1:27
    the way to wisdom and enlightenment in a more direct way.
  • 1:27 - 1:30
    He went to India, meditated, learned a great deal about
  • 1:30 - 1:34
    this tradition, came back, went to Stanford and graduated.
  • 1:34 - 1:39
    After that, he also took a degree in Neuroscience,
  • 1:39 - 1:41
    a PhD in Neuroscience.
  • 1:41 - 1:45
    He's written... I believe this is four books:
  • 1:46 - 1:48
    End of Faith, Letter to a Christian Nation,
  • 1:48 - 1:51
    and now the Moral Landscape,
  • 1:51 - 1:56
    which is very appropriate for Berkeley because it's trying to see how science
  • 1:56 - 2:01
    can determine and enlighten values, and our senses of "right" and "wrong",
  • 2:01 - 2:05
    good and evil. I think it's a...
  • 2:05 - 2:11
    even gentle to say that Sam is an iconoclast,
  • 2:11 - 2:14
    especially, of religion. He's been very hard on
  • 2:14 - 2:18
    all forms of fundamentalism, especially Islam,
  • 2:18 - 2:22
    and that might give some of us
  • 2:22 - 2:26
    in the progressive religious movement
  • 2:26 - 2:27
    ...make us feel more comfortable.
  • 2:27 - 2:29
    But the fact is, he is very critical
  • 2:29 - 2:32
    of progressive religion also. Seeing it
  • 2:32 - 2:36
    as very often enabling of those with more radical beliefs.
  • 2:36 - 2:40
    I always think of when I read of him
  • 2:40 - 2:46
    how Paul Tillich said, you know, we always need atheists around, believers need atheists.
  • 2:46 - 2:49
    Sam is not an atheist, but he
  • 2:49 - 2:57
    and Hitchens, and Richard can be said
    "anti-theists".
  • 2:58 - 3:03
    They are warriors in the battle against ideas of god
  • 3:03 - 3:08
    which are demeaning and which produce violence and ignorance.
  • 3:09 - 3:11
    So, it's nice to have him here.
  • 3:11 - 3:15
    I don't think "nice" is the word - it would be illuminating
    to have him here!
  • 3:15 - 3:16
    Sam!
  • 3:16 - 3:26
    [wild applause]
  • 3:25 - 3:27
    Thank you.
  • 3:31 - 3:32
    Thank you so much.
  • 3:32 - 3:35
    Well Sam, that is the first introduction that has detailed
  • 3:35 - 3:41
    my history of drug use. I think that is probably appropriate here in Berkeley.
  • 3:41 - 3:47
    Well, it's an honor to be here - I've never spoken in Berkeley before.
  • 3:47 - 3:54
    And, as many of you know I've been speaking for the last
  • 3:54 - 3:59
    six years or so, very critically about religion
  • 3:59 - 4:03
    and when you criticize religion in public,
  • 4:03 - 4:06
    what you immediately get are all the reasons
  • 4:06 - 4:07
    why people think it's a terrible idea.
  • 4:07 - 4:14
    The first reason: is never that there's so much evidence that god exists,
  • 4:14 - 4:19
    that you're denying the obvious, that you haven't read the Bible closely enough.
  • 4:19 - 4:24
    The first reason, even for fundamentalists, the truth of religion
  • 4:24 - 4:27
    is not the first line of defense.
  • 4:27 - 4:34
    The first reason is that religion, it is imagined, offers the only possible
  • 4:34 - 4:40
    framework to think about morality in truly global terms, in universal terms.
  • 4:42 - 4:45
    And, there are several things wrong with this claim.
  • 4:45 - 4:47
    Obviously, it's not an argument.
  • 4:47 - 4:52
    Even if it were true, it wouldn't prove that god exists:
  • 4:52 - 4:55
    religion could function as a placebo.
  • 4:55 - 5:00
    It also doesn't reconcile the obvious contradictions among the world's religions.
  • 5:00 - 5:07
    It doesn't reconcile the contradictory truth claims of Christianity and Judaism,
  • 5:07 - 5:13
    or Christianity and Islam. So, it is kind of a non sequitur as a defensive god.
  • 5:13 - 5:16
    But, I used to think it was a totally empty claim,
  • 5:16 - 5:24
    but now I've come to believe that actually religious people, and even fundamentalists,
  • 5:24 - 5:30
    perhaps, especially fundamentalists, are worried about something that is worth worrying about.
  • 5:30 - 5:38
    They are worried that secular people, for the most part, have become convinced
  • 5:38 - 5:42
    that something has happened in the last two hundred years of intellectual discourse,
  • 5:42 - 5:45
    that has made it impossible to speak about moral truth.
  • 5:47 - 5:53
    And I continually meet people who seem to have had their convictions of moral truth eroded by something
  • 5:53 - 5:55
    that has happened in science and philosophy.
  • 5:58 - 6:02
    And this, I think troubles them. I think we are in danger of waking up in a world
  • 6:02 - 6:06
    where the only people who are sure that moral truths exist,
  • 6:06 - 6:10
    are religious demagogues that think the Universe is 6,000 years old.
  • 6:10 - 6:13
    That's not the world we should be eager to live in,
  • 6:13 - 6:18
    so I'm going to push your intuitions around on this front.
  • 6:19 - 6:28
    But, just to give you a sense of the problem, and to tell you how this issue was really seared onto my brain.
  • 6:28 - 6:32
    I'll tell a little story. I was at a conference few years ago
  • 6:32 - 6:35
    talking about the link between morality and human well-being
  • 6:35 - 6:42
    as I'm going to tonight; and I said something that I thought would be quite uncontroversial in this context.
  • 6:42 - 6:47
    I said, "Listen, we know that morality relates to human well-being.
  • 6:47 - 6:57
    We know that human well-being relates to the facts that allow mind to emerge in the brain,
  • 6:57 - 7:03
    so it's constrained by truth claims, in some sense.
  • 7:03 - 7:09
    And therefore, we know that certain cultures are wrong about how to maximize human well-being -
  • 7:09 - 7:13
    and therefore, they are wrong in terms of what they value."
  • 7:13 - 7:19
    And I cited as an example life for women, especially
  • 7:19 - 7:26
    under the Taliban. It seemed to me their violent misogyny and religious lunacy
  • 7:26 - 7:32
    was a pretty obvious context in which people, especially women, were not thriving.
  • 7:32 - 7:39
    Now, it turns out to denigrate the Taliban at a scientific conference is to court controversy.
  • 7:42 - 7:45
    So, after I spoke another speaker came up to me and said:
  • 7:45 - 7:53
    "How could you ever say that the compulsory veiling of women is wrong, from the point of view of science?"
  • 7:53 - 7:58
    I said, "Whoa, ok, the moment you link the questions of right and wrong
  • 7:58 - 8:05
    to questions of human well-being, then it seems pretty clear that forcing half of the population to live in cloth bags
  • 8:05 - 8:12
    and beating them or killing them, when they try to get out, is not a way of maximizing well-being,
  • 8:12 - 8:16
    and therefore, not a good practice. And she said, "Well, that's just your opinion."
  • 8:19 - 8:23
    I said, "Ok, well, let's just make it easier - let's imagine we found a culture
  • 8:23 - 8:30
    that was removing the eyeballs of every third child. Would you then agree that we've found a culture
  • 8:30 - 8:34
    that was not perfectly maximizing human well-being?"
  • 8:36 - 8:38
    And she said, "It would depend on why they were doing it."
  • 8:38 - 8:43
    [laughter]
  • 8:43 - 8:47
    So, after I picked my jaw back up off the floor, I said:
  • 8:47 - 8:50
    "Ok, let's say they're doing it for religious reasons.
  • 8:50 - 8:56
    Let's say, they have a scripture which says "every third should walk in darkness"
  • 8:56 - 8:57
    or some such nonsense.
  • 8:57 - 9:00
    [laughter]
  • 9:03 - 9:08
    And, you'll be pleased or horrified to know that she just bit the bullet here and said:
  • 9:08 - 9:10
    "Then you could never say that they were wrong!"
  • 9:10 - 9:16
    OK, this was a woman who has background in science and philosophy;
  • 9:16 - 9:19
    she's now on the President's Council for bio-ethics;
  • 9:19 - 9:24
    she's one of 13 people, advising our President on all of the issues
  • 9:24 - 9:27
    related to progress in medicine and life-science generally.
  • 9:27 - 9:35
    She had just delivered a totally lucid lecture on the moral implications,
  • 9:35 - 9:39
    as she saw them, for the use of Neuroscience in our courts.
  • 9:39 - 9:43
    She was very worried that we have been developing lie-detection technology,
  • 9:43 - 9:47
    and that we are using this on captured terrorists,
  • 9:47 - 9:51
    and she viewed this as an invasion of cognitive liberty.
  • 9:51 - 9:56
    On the one hand, her moral scruples were really finely calibrated
  • 9:56 - 10:01
    to our own possible missteps, in this case in our war on terror.
  • 10:01 - 10:08
    But she was rather sanguine about the ritual enucleation of children
  • 10:10 - 10:16
    and seemed to me terrifyingly detached from the very real suffering of millions of women
  • 10:16 - 10:21
    in Afghanistan at this moment. So, this kind of impossible juxtaposition of views
  • 10:21 - 10:29
    is something I'm encountering a lot now, among disproportionately well-educated and liberal people.
  • 10:29 - 10:32
    So, it's something we have to grapple with.
  • 10:32 - 10:41
    Now, the issue for most people is that it has been said, over and over again,
  • 10:41 - 10:44
    that there is a distinction between facts and values,
  • 10:45 - 10:51
    and that science and rationality generally, can only really make truth claims
  • 10:51 - 10:55
    about the former. So, science obviously can deal with facts;
  • 10:55 - 10:59
    we have a universe of facts that we can understand to greater or lesser degree.
  • 10:59 - 11:03
    Facts transcend culture in some basic sense.
  • 11:03 - 11:07
    But it's not that values are another thing entirely;
  • 11:07 - 11:11
    values are the domain of questions of right and wrong, and good and evil,
  • 11:11 - 11:18
    and inconveniently for us, this is the area where the most important questions of human life arise.
  • 11:18 - 11:22
    These are questions like: how should you raise your children,
  • 11:22 - 11:27
    what goals should you strive for in life, what constitutes a good life.
  • 11:27 - 11:33
    And it's not that science will never be able to tell us the right answers to these questions,
  • 11:33 - 11:37
    just as science is never going to tell you whether you should like chocolate over vanilla,
  • 11:37 - 11:41
    it's not going to tell you how you should raise your children or treat your neighbour.
  • 11:41 - 11:45
    I think this is an illusion, this is quite confused.
  • 11:47 - 11:53
    It's a dangerous illusion, for as I said, it erodes the conviction of very smart people
  • 11:53 - 12:04
    in the face of really barbaric practices, which occasion needless human misery.
  • 12:06 - 12:11
    Now, it's long been obvious that we have needed a universal framework
  • 12:11 - 12:15
    to think about values and morality.
  • 12:15 - 12:23
    In the immediate aftermath of World War II, the UN put forward its Universal Declaration of Human Rights,
  • 12:23 - 12:32
    and yet, the American Anthropological Association, in all its wisdom, came forward
  • 12:32 - 12:37
    and said: "This is a fool's errand. You can't put forward a Universal Declaration of Human Rights
  • 12:37 - 12:44
    because you're merely foisting one provincial notion of value onto the rest of humanity.
  • 12:44 - 12:48
    It is an intellectually illegitimate thing to do."
  • 12:48 - 12:52
    Notice, this is the best our social scientists could do,
  • 12:52 - 12:57
    really with the crematoria of Auschwitz still smoking: this is 1947.
  • 13:00 - 13:07
    Now, I think the connection between facts and values is actually straight forward and philosophically uninteresting.
  • 13:07 - 13:14
    And i'm going to alight many of the common categories and distinctions one finds in philosophy
  • 13:14 - 13:20
    in what's called 'meta-ethics'. I just want to say there's going to be a Q & A afterwards,
  • 13:20 - 13:25
    and if you think I've missed some crucial bit of philosophy in my treatment here,
  • 13:25 - 13:30
    please come to the mic and deliver the devastating argument.
  • 13:30 - 13:34
    I want to hear from you. I don't want you to leave with your doubts intact,
  • 13:34 - 13:41
    but I'm convinced that many of the categories we have in philosophy,
  • 13:41 - 13:48
    to talk about the truth value of morality and the relationship between values and facts,
  • 13:48 - 13:56
    are needlessly confusing, and there's no reason to keep rehashing these ancient,
  • 13:56 - 14:00
    and I think, truly moribund debates.
  • 14:03 - 14:07
    Values reduce to facts about the well-being of conscious creatures.
  • 14:07 - 14:12
    The well-being of conscious creatures is what can be valued in this universe.
  • 14:14 - 14:19
    If you doubt this, just imagine a universe populated entirely by rocks.
  • 14:19 - 14:24
    OK, rocks, presumably, have no inner dimension:
  • 14:24 - 14:26
    there's nothing that is like to be a rock.
  • 14:27 - 14:31
    We are right to be unconcerned about the experience of rocks,
  • 14:31 - 14:34
    because we think there's no possibility of experience there.
  • 14:34 - 14:40
    And I think, we are right to be more concerned about our fellow primates than we are about insects,
  • 14:40 - 14:49
    because we think there's an inner dimension there that can be modulated to a much greater degree
  • 14:49 - 14:55
    by changes in the universe. A universe... for changes in the universe to matter,
  • 14:55 - 14:58
    they have to matter to some conscious system.
  • 15:00 - 15:06
    Now, if you doubt this, here's the one bit of philosophy I'm going to anchor this to.
  • 15:07 - 15:09
    It seems to me the only assumption you have to buy.
  • 15:11 - 15:16
    Imagine a universe in which every conscious creature suffers as much as it possibly can
  • 15:16 - 15:21
    for as long as it can - I call this "the worst possible misery for everyone".
  • 15:21 - 15:26
    The worst possible misery for everyone is bad.
  • 15:27 - 15:33
    If the word "bad" is to mean anything, surely it applies to the worst possible misery for everyone.
  • 15:33 - 15:35
    [laughter]
  • 15:35 - 15:41
    Now, if you don't think the worst possible misery for everyone is bad,
  • 15:41 - 15:44
    or you think there might be something worse, or you think it might have a silver lining, -
  • 15:44 - 15:48
    I don't know what you're talking about.
  • 15:48 - 15:51
    [laughter]
  • 15:51 - 15:53
    And I'm pretty sure, you don't know what you're talking about either.
  • 15:56 - 16:01
    The moment you grant me that the worst possible misery for everyone is bad,
  • 16:01 - 16:07
    therefore worth avoiding, and if we should do anything in this universe
  • 16:07 - 16:10
    we should avoid the worst possible misery for everyone.
  • 16:10 - 16:14
    Well, then you've the worst possible misery for everyone over here
  • 16:14 - 16:18
    and you have every other possible constellation of conscious experience
  • 16:18 - 16:22
    which, by definition, is better.
  • 16:22 - 16:27
    You see, we have this continuum here of possible states of consciousness,
  • 16:27 - 16:33
    and given that consciousness is related to the way the universe is.
  • 16:33 - 16:36
    It's constrained by the laws of nature in some way
  • 16:36 - 16:41
    there are going to be right and wrong ways to move along this continuum.
  • 16:41 - 16:46
    There going to be ways to think you're avoiding the worst possible misery and to fail.
  • 16:46 - 16:51
    It's going to be possible to have erroneous beliefs about how best to move
  • 16:51 - 16:55
    from your current space in this continuum to something better.
  • 17:00 - 17:07
    Now, this is, in philosophy, a somewhat controversial statement, again, I do not see how.
  • 17:10 - 17:14
    But it meets with objections of the following sort:
  • 17:14 - 17:21
    You hear people say, "What if someone wanted to torture all conscious beings to the point of madness?
  • 17:21 - 17:25
    How could you ever prove that he's not as good as you are?"
  • 17:27 - 17:28
    This is kind of email I get.
  • 17:28 - 17:33
    [laughter]
  • 17:34 - 17:41
    Or, more relevantly, if a member of the Taliban wants to throw battery acid
  • 17:41 - 17:44
    in the face of a little girl for the crime of learning to read.
  • 17:44 - 17:48
    How could you ever convince him that he is not as moral as you are?
  • 17:48 - 17:54
    Again, this is really the kind of email I get.
  • 17:54 - 17:56
    Now, this is a pseudo-problem.
  • 17:59 - 18:04
    Notice, we don't confront this (I failed to anticipate a slide.)
  • 18:06 - 18:10
    A religious conception of morality, again,
  • 18:10 - 18:15
    falls into the same concept of well-being and its possible changes,
  • 18:15 - 18:17
    the conscious well-being of creatures.
  • 18:17 - 18:21
    So if you are concerned about a lifetime of happiness with god after death in heaven
  • 18:21 - 18:24
    and avoiding a lifetime of misery in hell, again,
  • 18:24 - 18:27
    you are concerned about consciousness and its changes.
  • 18:27 - 18:30
    You just happen to think consciousness and its changes
  • 18:30 - 18:34
    are most importantly experienced after death for eternity.
  • 18:34 - 18:37
    Now again, this is a claim about which science has some rather obvious doubts,
  • 18:37 - 18:39
    but this is still the same framework.
  • 18:39 - 18:41
    We're talking about consciousness and its changes.
  • 18:41 - 18:54
    Now, this challenge, this fundamental skeptical challenge, about well-being not being worth valuing,
  • 18:54 - 18:58
    and how you could convince someone who doesn't value it.
  • 18:58 - 19:02
    Notice, how this maps onto our notion of physical health.
  • 19:02 - 19:08
    Physical health is loosely defined; we don't have a clear definition of health.
  • 19:08 - 19:12
    We sort of know when we see it: has something to do with not dying too early,
  • 19:12 - 19:15
    has something to do with not always vomiting,
  • 19:15 - 19:18
    not being in continuous pain.
  • 19:18 - 19:24
    Now, if we re-engineer our genome so that we can live to be a thousand,
  • 19:24 - 19:27
    and regrow missing limbs like a salamander,
  • 19:27 - 19:32
    that would become part of our expectation of basic physical health.
  • 19:32 - 19:40
    So, this notion of health is truly elastic, and yet, the fact that it can't be clearly typed out
  • 19:42 - 19:44
    is not a problem for the science of medicine.
  • 19:44 - 19:50
    You don't hear a philosophical challenge to medicine of the following sort:
  • 19:50 - 19:55
    "Who are you to say that not always vomiting is healthy?"
  • 19:55 - 19:57
    [laughter]
  • 19:57 - 20:01
    "What if you meet someone who wants to always vomit and wants to be dead tomorrow?"
  • 20:02 - 20:05
    You don't hear someone say:
  • 20:05 - 20:12
    "How do you convince a person with terminal small pox that he is not as healthy as you are?"
  • 20:12 - 20:17
    This kind of attack upon medicine would make no sense,
  • 20:17 - 20:23
    and, yet, this is precisely the attack one hears from moral relativists and multiculturalists,
  • 20:23 - 20:30
    when you talk about the very obvious and needless, and horrific misery
  • 20:30 - 20:39
    of millions of people in situations that are anchored to really pathological notions of good and evil.
  • 20:43 - 20:55
    Now, even the most basic, apparently value-free descriptions of fact in science
  • 20:55 - 20:59
    are also anchored to values in a way that would never survive
  • 20:59 - 21:03
    this kind of skeptical challenge you meet when you talk about morality.
  • 21:03 - 21:07
    Consider water. Water is a substance we now dimly understand;
  • 21:07 - 21:11
    for about 150 years we've known it's 2 parts hydrogen and 1 part oxygen.
  • 21:11 - 21:16
    What do we do when someone doubts that proposition?
  • 21:16 - 21:19
    What do we do if someone comes into the room and says:
  • 21:19 - 21:21
    "Well, that's not how I choose to think about water"?
  • 21:24 - 21:27
    What do we do when someone says, "Well, I'm actually, I'm a biblical chemist.
  • 21:27 - 21:32
    [laughter]
  • 21:32 - 21:36
    You can get your chemistry from science, but I get my chemistry from the book of Genesis,
  • 21:36 - 21:40
    and whatever squares with Genesis, is chemistry for me."
  • 21:42 - 21:45
    Again, see the analogy to moral truth.
  • 21:48 - 21:54
    The only thing we can do in that case - is appeal to scientific values.
  • 21:54 - 21:58
    I mean, you have to value understanding the Universe,
  • 21:58 - 22:02
    you have to value evidence, you have to value logical consistency.
  • 22:02 - 22:06
    If you don't value these things, the conversation stops.
  • 22:06 - 22:12
    There is no convincing someone who doesn't value evidence that they should value.
  • 22:12 - 22:17
    What evidence are you going to provide to convince someone they should value evidence?
  • 22:17 - 22:19
    [laughter]
  • 22:19 - 22:23
    What logical argument are you going to offer to convince them
  • 22:23 - 22:26
    of the necessity of obeying the rules of logic?
  • 22:28 - 22:33
    So, even the most basic scientific statements:
  • 22:33 - 22:36
    You don't get more basic than the chemistry of water.
  • 22:36 - 22:39
    You don't get more value free than the chemistry of water.
  • 22:39 - 22:42
    ...are anchored to values at every point.
  • 22:44 - 22:49
    This covers Hume's famous notion of "you can't get an 'ought' from an 'is'."
  • 22:49 - 22:53
    You can't get a statement of how we should behave or how the world should be,
  • 22:53 - 22:58
    based on a description of the way it is. You can't get an 'is' without an 'ought';
  • 22:58 - 23:02
    you can't make the most basic scientific statement
  • 23:02 - 23:07
    without conforming to the norms of scientific rationality.
  • 23:07 - 23:10
    So, science is very much in the values business.
  • 23:10 - 23:15
    It is a myth that there is this division between facts and values in science.
  • 23:17 - 23:21
    Now, there's another way to bridge this supposed gap between facts and values,
  • 23:21 - 23:25
    and it's this. When you look more closely at what beliefs are,
  • 23:25 - 23:29
    we form beliefs about the world in many domains:
  • 23:29 - 23:32
    We form beliefs about facts, obviously, and this constitutes science,
  • 23:32 - 23:36
    but it constitutes every truth claim we make about the world;
  • 23:36 - 23:40
    journalism, and common sense, and your personal memories of your past.
  • 23:40 - 23:42
    But we also form beliefs about values,
  • 23:42 - 23:48
    and this includes religion, and ethics, and questions of right and wrong, and good and evil.
  • 23:48 - 23:52
    We decide to look at these operations to level the brain;
  • 23:52 - 23:56
    we put people in fMRI scanners and gave them statements to read,
  • 23:56 - 23:57
    statements from many different categories,
  • 23:57 - 24:03
    and just looking to see if the difference between judging something true,
  • 24:03 - 24:05
    and rejecting another statement is false,
  • 24:05 - 24:08
    was sensitive to content. So,
  • 24:08 - 24:12
    statements from science, statements from religion, statements from ethics...
  • 24:12 - 24:18
    and we found that the brain is essentially doing the same thing, independent of content.
  • 24:18 - 24:22
    So, on the left, we have all of our categories in together,
  • 24:22 - 24:27
    and you get this region of signal, in what is called 'ventromedial prefrontal cortex',
  • 24:27 - 24:32
    and in the middle we have mathematics, and on the right we have ethics.
  • 24:32 - 24:36
    So, mathematics and ethics were probably our most different category areas.
  • 24:36 - 24:39
    Mathematics was just equations; you just had 'true' and 'false' equations.
  • 24:39 - 24:44
    Ethics was statements like "It's good to be kind to children"
  • 24:44 - 24:47
    vs. "It's good to torture children".
  • 24:47 - 24:50
    So, obviously very value-laden statements,
  • 24:50 - 24:52
    and the difference between accepting a proposition
  • 24:52 - 24:57
    and rejecting it seemed to be importantly similar,
  • 24:57 - 25:00
    within the tolerance of an fMRI experiment
  • 25:00 - 25:02
    those are essentially the same maps.
  • 25:02 - 25:07
    So, I'm not placing too much emphasis on this,
  • 25:07 - 25:10
    but I would just suggest to you that, if the brain thinks it's doing the same thing,
  • 25:10 - 25:16
    when it's accepting a proposition about ethics vs. rejecting another proposition about ethics,
  • 25:16 - 25:21
    and accepting mathematical statements vs. rejecting mathematical statements,
  • 25:21 - 25:26
    then I think we should be very slow to break our world-view
  • 25:26 - 25:29
    into separate fundamental categories of values and facts.
  • 25:29 - 25:39
    So, I would suggest you that belief is really our best effort to map reality in our thoughts.
  • 25:39 - 25:46
    And when we seem to succeed in doing this, when our beliefs about the world survive the tests
  • 25:46 - 25:50
    that the world throws up at them, then we call it knowledge,
  • 25:50 - 25:55
    then we don't talk about believing things, we talk about knowing things.
  • 25:55 - 25:59
    But still we're simply talking about linguistic representations of the world.
  • 26:04 - 26:10
    I think it's clear there's a continuum of facts about which we can form true or false beliefs
  • 26:10 - 26:15
    that relate very directly to human happiness,
  • 26:15 - 26:21
    to the question of how human beings, both individually and collectively, can flourish.
  • 26:21 - 26:28
    There's clearly continuum where on a one hand, you can live in a failed state,
  • 26:28 - 26:31
    where everything that can go wrong, does go wrong.
  • 26:31 - 26:34
    Think of place like Congo at the moment.
  • 26:34 - 26:37
    This photo is from Somalia in the '80s,
  • 26:37 - 26:41
    but Congo at this moment, and Somalia also a failed state,
  • 26:41 - 26:46
    but the worst example, probably, of any place to be at this moment, is Congo,
  • 26:46 - 26:53
    where everyone's daily concern, essentially, is to avoid being raped and killed by drug-addled soldiers.
  • 26:53 - 27:01
    There's absolutely no basis for cooperation, to think about the education of children.
  • 27:01 - 27:06
    Things can go disastrously wrong in human communities,
  • 27:06 - 27:09
    but they go wrong for a reason,
  • 27:09 - 27:17
    and how we move from a state of absolute and needless misery
  • 27:17 - 27:22
    to something much more idyllic, that movement is still constrained, clearly,
  • 27:22 - 27:31
    by the dynamics of human psychology, and social systems, and economic incentives,
  • 27:31 - 27:35
    and the rule of law and all these... We can't just make everything up.
  • 27:35 - 27:39
    This is not merely culture that explains these movements,
  • 27:39 - 27:42
    and clearly there are many levels at which we could understand
  • 27:42 - 27:46
    how human life can be improved,
  • 27:46 - 27:47
    how human well-being can be maximized.
  • 27:47 - 27:53
    There's the level of the genome, and clearly there are genes for positive social emotions
  • 27:53 - 27:55
    that people can have to a greater or lesser degree,
  • 27:58 - 28:05
    and there's at the other extreme, the dynamics of economies and political arrangements.
  • 28:05 - 28:08
    But every place in between, we're talking about anything
  • 28:08 - 28:12
    that can possibly influence states of the human brain.
  • 28:12 - 28:17
    We're talking about genetics and neurobiology, and psychology, and sociology and economics.
  • 28:18 - 28:24
    So, I'm not talking... when I talk about science, giving us an understanding of human values,
  • 28:24 - 28:28
    I'm not narrowly talking about white-lab-coated experimentalists scanning brains,
  • 28:28 - 28:34
    I'm talking about any area of human life where we make truth claims honestly,
  • 28:34 - 28:38
    based on honest observation and clear reasoning about the nature of reality.
  • 28:38 - 28:44
    There are clearly right and wrong ways for human beings to seek to thrive.
  • 28:49 - 28:52
    This I think, is easy to see. If you just imagine two people living on Earth,
  • 28:52 - 28:54
    we can call them Adam and Eve.
  • 28:54 - 29:04
    If you think of just two people here, it's pretty clear that they have better and worse options.
  • 29:04 - 29:11
    That there're better and worse responses to the predicament of simply appearing in this place.
  • 29:11 - 29:18
    Bad option number 1: they could smash each other in the face with a large rock. OK.
  • 29:18 - 29:26
    But clearly that is the failure to discover the most promising sources of collaboration
  • 29:26 - 29:29
    that the human condition offers.
  • 29:29 - 29:37
    Now, how does this change when you add 6.7 billion people to the experiment?
  • 29:39 - 29:41
    I don't think it does, it just gets more complicated.
  • 29:46 - 29:50
    So, what I ask you to consider is what I call 'a moral landscape'
  • 29:50 - 29:54
    where the peaks correspond to the heights of well-being,
  • 29:54 - 29:57
    and the valleys correspond to the lowest depths of suffering,
  • 29:57 - 30:06
    and one thing to realize about this analogy that it's pretty clear there can be multiple peaks on this landscape.
  • 30:06 - 30:11
    It's not obviously so, it doesn't have to be so but I think, it's quite plausable,
  • 30:11 - 30:16
    that there are many different ways for human beings and human communities to thrive
  • 30:16 - 30:21
    that are dissimilar, and therefore importantly different,
  • 30:21 - 30:23
    and if you were living one way, you can't live another way,
  • 30:23 - 30:28
    but clearly there are going to be many more ways to not be on the peak.
  • 30:28 - 30:33
    There gonna be many more ways to fail to be as happy and as creative,
  • 30:33 - 30:37
    and as intelligent as human communities could be.
  • 30:40 - 30:43
    Now, why wouldn't multiple right answers be a problem?
  • 30:43 - 30:46
    Well, consider the analogy of food.
  • 30:46 - 30:50
    I would never be tempted to argue that there must be one right food to eat.
  • 30:50 - 30:53
    There are clearly many right answers to the question, 'What is food?',
  • 30:53 - 30:56
    but there are obvious wrong answers.
  • 30:56 - 31:03
    The distinction between food and poison is still scientifically true,
  • 31:03 - 31:07
    and it's true even with all of the caveats.
  • 31:07 - 31:10
    Some people are allergic to peanuts, they'll die if they eat them;
  • 31:10 - 31:12
    peanuts are poison for them, but a food for us.
  • 31:12 - 31:16
    We can understand all of this in the context of chemistry and biology
  • 31:16 - 31:20
    and every science related to human health.
  • 31:20 - 31:27
    People also worry that, if there's going to be an ethical principle
  • 31:27 - 31:30
    that's true, it has to always be true,
  • 31:30 - 31:32
    and if you find a single exception,
  • 31:32 - 31:35
    well then there's no such a thing as 'moral truth'.
  • 31:35 - 31:36
    Well, consider by analogy the game of chess.
  • 31:36 - 31:43
    The principle of not loosing your queen in chess is absolutely
  • 31:43 - 31:47
    worth following, almost all the time it is one of the best things
  • 31:47 - 31:49
    you can seek to do,
  • 31:49 - 31:53
    and yet it admits of nearly countless exceptions.
  • 31:53 - 31:57
    There are moments where loosing your queen is a brilliant thing to do.
  • 31:57 - 32:04
    Now, this is, think by analogy the principle of not lying.
  • 32:04 - 32:10
    Not lying, 'truth telling', is obviously, in so many circumstances, a good thing to do.
  • 32:10 - 32:16
    Telling lies is almost the easiest way to screw up your life.
  • 32:16 - 32:21
    Yet, when the Nazis come knocking on the door, and say,
  • 32:21 - 32:23
    asking whether you have Jews in the basement,
  • 32:23 - 32:29
    that might be a time to forget about Kant and tell your first lie.
  • 32:29 - 32:32
    [some laughter]
  • 32:32 - 32:36
    Now, the fact that there's a situational exception to the principle of not lying
  • 32:36 - 32:39
    does not mean that there's no such thing as 'moral truth'.
  • 32:45 - 32:49
    This model of a landscape also admits of the possibility
  • 32:49 - 32:52
    of what we call 'spiritual' or 'mystical' experience.
  • 32:52 - 32:56
    I think that there's no question that human mind
  • 32:56 - 33:01
    is capable of having remarkable self-transcending experiences,
  • 33:01 - 33:07
    many of which can be very hard-won, many of which you have to have a talent, perhaps, to access,
  • 33:07 - 33:11
    and certainly, training to access.
  • 33:11 - 33:14
    And many positive social emotions that we all experience
  • 33:14 - 33:21
    can be brought to a much higher register, than we bring them.
  • 33:21 - 33:23
    It's something like compassion or empathy.
  • 33:23 - 33:29
    I used to be in the habit of saying that, 'undoubtedly, there's a Tiger Woods of compassion out there'.
  • 33:29 - 33:34
    For obvious reasons that analogy doesn't work so well at the moment.
  • 33:34 - 33:35
    [laughter]
  • 33:37 - 33:39
    But I think, compassion is best thought of as a skill.
  • 33:39 - 33:44
    It's clearly trainable, it's clearly something that people have
  • 33:44 - 33:48
    to a greater or lesser degree, and we're beginning to understand
  • 33:48 - 33:54
    the neurology of both of its appearance and its encouragement.
  • 33:54 - 33:58
    Our minds are, to a significant degree, plastic.
  • 33:58 - 34:01
    I mean, you're sort of become what you pay attention to.
  • 34:01 - 34:06
    And, this is something a maturing science of the human mind
  • 34:06 - 34:09
    really can put us in a position to understand.
  • 34:15 - 34:17
    How many of you recognize this photo?
  • 34:17 - 34:20
    This is a photo of, apparently, very happy Nazis.
  • 34:20 - 34:21
    (You can't quite make it out.)
  • 34:21 - 34:24
    This is from a book that's now called "The Auschwitz Album";
  • 34:24 - 34:28
    there was a photo album that was discovered in an attic somewhere,
  • 34:28 - 34:32
    of just Nazis having a gay old time.
  • 34:32 - 34:36
    And it took some research to figure out who they were
  • 34:36 - 34:37
    and what they were doing.
  • 34:37 - 34:43
    These were people working at Auschwitz during the peak of its productivity
  • 34:43 - 34:49
    as a factory of death, and this was a... they were at a kind of a chalet
  • 34:49 - 34:52
    that was a few kilometers from the death camp.
  • 34:53 - 34:58
    So, this is, more less their mood as they were listening to accordion music
  • 34:58 - 35:00
    and lying in the sun, and eating blueberries
  • 35:00 - 35:05
    under the plume of human ash coming out of the crematoria of Auschwitz.
  • 35:05 - 35:10
    Now, there's nothing about my view of the moral landscape,
  • 35:10 - 35:13
    or the link I'm drawing between morality and human well-being,
  • 35:13 - 35:21
    that insists that I deny the obvious happiness of these people.
  • 35:21 - 35:24
    I don't think they are fake smiles that we see there,
  • 35:24 - 35:27
    and I don't think these people are psychopaths.
  • 35:27 - 35:31
    I think, for the most part, these are normally normal people who would go home
  • 35:31 - 35:38
    and pet their dogs and cats, and they love their children, and they would listen to Wagner and shed a tear.
  • 35:41 - 35:46
    The problem with these people was not that they had a radically different conception of morality,
  • 35:46 - 35:51
    they have a 'moral circle' that they had radically delimited.
  • 35:51 - 35:56
    They had put most of humanity outside the sphere of their moral concern
  • 35:56 - 35:58
    and that's what we continually run into in the world,
  • 35:58 - 36:04
    based on some divisive dogmatism, we run into groups of people,
  • 36:04 - 36:10
    who just managed to put the better part of humanity, or significant groups of humanity,
  • 36:10 - 36:14
    outside the theater of their moral concern.
  • 36:17 - 36:20
    I want to talk about Islam for a moment,
  • 36:20 - 36:26
    because I think we are wise to be concerned about it.
  • 36:26 - 36:28
    As you know, I'm concerned about religion in general,
  • 36:28 - 36:34
    but I think, we are wise to differentiate specific religious beliefs.
  • 36:37 - 36:43
    And we are, I think, quite encumbered by political correctness and just frank confusion on this front.
  • 36:43 - 36:49
    One problem is that we have this one word, "religion", which names
  • 36:49 - 36:59
    truly diverse spectrum of fascinations and ideological commitments.
  • 36:59 - 37:05
    And religion is a nearly useless term; it's a term like "sports".
  • 37:07 - 37:11
    There are sports like Badminton and there are sports like Thai Boxing,
  • 37:11 - 37:15
    and they have almost nothing in common, apart from breathing.
  • 37:15 - 37:17
    [some laughter]
  • 37:17 - 37:23
    There are sports that are just synonymous with the risk of physical injury, or even death.
  • 37:23 - 37:25
    I mean, there are sports that are just synonymous with violence.
  • 37:25 - 37:29
    If you get injured playing badminton, you're just embarrassed.
  • 37:29 - 37:31
    [laughter]
  • 37:37 - 37:38
    We're facing a problem at this moment:
  • 37:38 - 37:44
    there is, I'm happy to say, a religion of peace in this world,
  • 37:44 - 37:46
    but it's not Islam.
  • 37:48 - 37:52
    To call Islam "a religion of peace", as we hear, ceaselessly reiterated,
  • 37:52 - 37:54
    is completely delusional.
  • 37:54 - 37:58
    Now, Jainism actually is a religion of peace.
  • 37:58 - 38:04
    In Jainism, the core principle of Jainism is non-violence.
  • 38:04 - 38:06
    Gandhi got his non-violence from the Jains.
  • 38:08 - 38:12
    The crazier you get as a Jain, the less we have to worry about you.
  • 38:12 - 38:14
    [laughter]
  • 38:20 - 38:29
    Jain extremists are actually, they are paralysed by their pacifism.
  • 38:31 - 38:35
    Jane extremists just can't take their eyes off the ground when they walk,
  • 38:35 - 38:36
    lest they step on an ant.
  • 38:36 - 38:43
    They filter every sip of water through cheese cloth, lest they swallow and thereby kill a bug,
  • 38:43 - 38:46
    and, needless to say, they are vegetarian.
  • 38:49 - 38:52
    So the problem, notice, is not religious extremism,
  • 38:52 - 38:59
    because extremism is not a problem, if your core beliefs are truly non-violent.
  • 38:59 - 39:06
    The problem isn't fundamentalism, which we often hear this said, these are euphemisms.
  • 39:06 - 39:11
    The only problem with Islamic fundamentalism are the fundamentals of Islam.
  • 39:11 - 39:15
    [some laughter]
  • 39:15 - 39:19
    Now, we have Melo Omar, Osama bin Laden, and Khalid Sheikh Mohammed.
  • 39:19 - 39:24
    These guys agree about the nature of reality and how to live within it,
  • 39:24 - 39:30
    and the problem is they are giving a very plausible version of the faith.
  • 39:32 - 39:36
    Osama bin Laden is not the Reverend Jim Jones of the Muslim world.
  • 39:36 - 39:49
    It would be wonderful if he were, but the problem is he's giving a truly straightforward version of Islam,
  • 39:49 - 39:57
    and you really have to be an acrobat to figure out how he is distorting the faith.
  • 39:57 - 40:06
    Now, if these guys were Jains, or Buddhists, or Amish, or Quakers, it would be patently obvious
  • 40:06 - 40:08
    how they were distorting their religion.
  • 40:08 - 40:11
    In fact, their behaviour would be unintelligible.
  • 40:11 - 40:17
    Ok. It is not obvious by the light of Islam; this is just a fact we have to speak honestly about.
  • 40:17 - 40:23
    And no-one should be speaking more honestly about this, and more volubly about this,
  • 40:23 - 40:29
    than moderate Muslims. Moderate Muslims have to find some way to grapple with this fact.
  • 40:29 - 40:38
    But to say that Osama bin Laden is David Koresh, is just a lie,
  • 40:38 - 40:41
    and it's a dangerous lie at this point.
  • 40:41 - 40:48
    Now, I just want to rehearse for you what these core beliefs are and what they entail.
  • 40:48 - 40:56
    The belief is, that Muhammad got the Koran directly from the archangel Gabriel in his cave, in the 7th century,
  • 40:56 - 41:00
    and it is the perfect word, therefore, of the Creator of the Universe.
  • 41:00 - 41:05
    I apologize for the cartoonish nature of this image of Muhammad,
  • 41:05 - 41:07
    but quality images are difficult to come by at the moment.
  • 41:07 - 41:11
    [laughter]
  • 41:13 - 41:18
    The consequences: we have this single book, which is imagined to be the best book
  • 41:18 - 41:23
    on any subject ever written, never to be superseded by any human effort
  • 41:23 - 41:27
    at any point in the future. Now, this is a problem
  • 41:27 - 41:30
    because this is a profoundly mediocre book.
  • 41:30 - 41:32
    [some laughter]
  • 41:32 - 41:38
    It is dangerous to say this. It is suicidal to say this as a Muslim.
  • 41:38 - 41:42
    It is true, and we have to grapple with this fact.
  • 41:42 - 41:48
    And, the idea that this is the best book ever written on any subject
  • 41:50 - 41:55
    can only be maintained in a kind of fantastical, intellectual isolation.
  • 41:57 - 42:00
    And this isolation has actually been achieved in the Arab world
  • 42:00 - 42:04
    to an astonishing degree, some of you probably have heard this fact,
  • 42:04 - 42:13
    but the country of Spain translates more of the world's literature and learning into Spanish every year,
  • 42:13 - 42:19
    than the entire Arab world has translated into Arabic since the 9th century.
  • 42:19 - 42:29
    That's scary! It's scary given that the contents of this book really offers
  • 42:29 - 42:38
    precious little rationale for living in a sane and pluralistic global civilization.
  • 42:42 - 42:47
    What it does give you a rationale for, is ceaselessly worshiping the perfect being
  • 42:47 - 42:49
    who has given you this mediocre book.
  • 42:49 - 42:55
    This is a photo of Muslims in Kashmir worshiping at a shrine,
  • 42:55 - 42:59
    believed to contain a single beard hair from the prophet Muhammad.
  • 43:00 - 43:07
    In showing you this image, I don't actually mean to denigrate the positive emotions
  • 43:07 - 43:09
    that can be associated with this kind of practice.
  • 43:09 - 43:13
    I think, devotion is a positive emotion that we want in our lives.
  • 43:13 - 43:18
    And I certainly don't mean to make a light of how difficult life, undoubtedly,
  • 43:18 - 43:21
    is for Muslims in Kashmir.
  • 43:21 - 43:27
    But it seems to me patently obvious, given the challenges that they face and that we all face
  • 43:27 - 43:32
    in creating world worth living in, these people have something more important to do
  • 43:32 - 43:38
    than worship the beard hair of a man, who may well have been a schizophrenic.
  • 43:42 - 43:48
    And again, when I talk about Islam, one thing I should make perfectly clear, and should've made it clear at the top,
  • 43:48 - 43:52
    I'm talking about the logical and behavioral consequences of ideas.
  • 43:52 - 43:55
    I'm not talking about people, i'm not talking about all Muslims.
  • 43:55 - 44:00
    Not all Muslims are terrorists, obviously, not all Muslims support terrorism,
  • 44:00 - 44:02
    not all Muslims take Islam all that seriously.
  • 44:02 - 44:09
    I'm certainly not talking about a race of people, I'm not talking about Arabs,
  • 44:09 - 44:11
    I'm not talking about nationalities or ethnicities.
  • 44:11 - 44:16
    I'm talking about ideas, and what people can plausibly do on their basis.
  • 44:20 - 44:27
    And what people do on the basis of these ideas, turns out to be very bad for women
  • 44:27 - 44:31
    to a remarkable degree, and in general.
  • 44:31 - 44:36
    And there really is no basis, in Islam, to argue that there should be true equality
  • 44:36 - 44:43
    between the sexes. You can finesse this issue, but what you cannot get is
  • 44:43 - 44:49
    a clear statement of "men and women are equal and have the same moral stature".
  • 44:49 - 44:54
    It is just not true. You have to edit the faith to get that.
  • 44:57 - 45:03
    And one thing you also get, by the logic of the faith at this moment,
  • 45:03 - 45:07
    is this death-cult behaviour we're now all too familiar with;
  • 45:07 - 45:15
    suicide bombing of the most extraordinary and ceaseless kind
  • 45:15 - 45:22
    (you can reliable turn to page 8 of New York Times every day and discover that someone
  • 45:22 - 45:26
    has blown up a mosque somewhere); first thing to point out is that
  • 45:26 - 45:29
    no-one suffers by this more than Muslims.
  • 45:29 - 45:33
    I mean, there are Muslims are getting killed by this; it is not Americans, for the most part.
  • 45:33 - 45:41
    This is a bombing in Pakistan, where Sunni bombed Shia procession.
  • 45:41 - 45:46
    The Sunnis view the Shia as apostates, and bombing them makes sense.
  • 45:50 - 45:53
    The core issue here is a notion of martyrdom -
  • 45:53 - 45:59
    the notion that death, in defense of the faith, is the surest way to paradise.
  • 45:59 - 46:03
    This belief is what makes sense of this behaviour.
  • 46:03 - 46:06
    I wanna linger over this image for a minute, actually.
  • 46:06 - 46:12
    This is a little girl (you can't quite see her in this lighting) crouched over the corpse of her mother.
  • 46:12 - 46:16
    And this is actually not the initial bombing.
  • 46:16 - 46:20
    What they did is, they bombed a procession of Shia pilgrims
  • 46:20 - 46:27
    and then they sent another suicide bomber to the hospital, to wait for the ambulances to come in,
  • 46:27 - 46:31
    to blow up the casualties, and the doctors, and the nurses, and the ambulance drivers.
  • 46:31 - 46:37
    So, this is a shot from outside the hospital. Just imagine this...
  • 46:37 - 46:42
    Just imagine what was like for someone to come up with that idea;
  • 46:42 - 46:47
    it's kind of creative in a truly diabolical way.
  • 46:47 - 46:51
    Someone had to volunteer to be the suicide bomber;
  • 46:51 - 46:55
    someone had to get up in the morning, thinking this is the best use of his life,
  • 46:55 - 46:58
    to blow up this little girl's mother.
  • 46:58 - 47:04
    This is, again, realize, this is not a collateral damage, this is the point of the exercise.
  • 47:04 - 47:10
    In fact, by the logic of this belief, it is impossible to kill the wrong people.
  • 47:10 - 47:15
    Because all the good Muslims, you blow up, are going to go to paradise, and they are going to thank you.
  • 47:15 - 47:22
    All the bad Muslims, all the apostates and the infidels, are going to go to hell where they belong.
  • 47:22 - 47:25
    It is impossible to screw this up.
  • 47:32 - 47:35
    And just, so you don't get the sense I'm narrowly focused on Islam.
  • 47:35 - 47:38
    [some laughter]
  • 47:40 - 47:45
    I would point out that the moment you see the link between morality and human well-being,
  • 47:45 - 47:49
    you can see that notions of "right" and "wrong", and "good" and "evil",
  • 47:49 - 47:56
    that come to us from religion, often break this connection,
  • 47:56 - 47:59
    and that's what so dangerous about religion.
  • 47:59 - 48:05
    In the best case, religion gives people bad reasons to be good, where good reasons are actually available.
  • 48:05 - 48:09
    In the worst case, it just disregards human well-being entirely.
  • 48:09 - 48:15
    For instance, the Catholic Church is simply more concerned about stopping contraception
  • 48:15 - 48:22
    than stopping the rape of children. That is a fact about both of the beliefs of this
  • 48:22 - 48:30
    particular person, the use of his energy over the last several decades, and just the energy of the Church.
  • 48:30 - 48:33
    If you know anything about the child-rape scandal in the Catholic Church,
  • 48:33 - 48:39
    it is mind-boggling, the effort that was not spent to protect children,
  • 48:39 - 48:44
    and the effort that was spent on sheltering the rapists from secular justice.
  • 48:44 - 48:52
    The Catholic Church is also more concerned is about stopping gay marriage than stopping genocide.
  • 48:52 - 48:54
    This is what its attention is on.
  • 48:54 - 48:58
    When you realize that questions of "right" and "wrong"
  • 48:58 - 49:02
    actually relate to questions of human well-being,
  • 49:02 - 49:08
    this is not morality, this is not an altered moral framework that we have to take seriously.
  • 49:08 - 49:12
    The Catholic Church could talk about physics. It could say:
  • 49:12 - 49:16
    "Well, we're actually interested in the physics of the transubstantiation,
  • 49:16 - 49:21
    or the physics that allows the holy ghost to be here and there, and everywhere, all at once."
  • 49:21 - 49:26
    But there's not a physicist alive who would have to take those utterances seriously.
  • 49:26 - 49:34
    I'm saying that, if you talk that way, you don't get invited back to the physics conference.
  • 49:34 - 49:36
    [laughter]
  • 49:39 - 49:43
    The moment we admit that morality relates to human and animal well-being,
  • 49:43 - 49:46
    you don't have to get invited back to the morality conference either.
  • 49:51 - 49:57
    Now, I'm gonna briefly remind you of some of the reasons
  • 49:57 - 50:02
    why religion can't be the repository of our moral wisdom generally.
  • 50:02 - 50:08
    One is that when we go to scripture, we are the guarantors of the wisdom we find there.
  • 50:08 - 50:13
    We find 'the golden rule' and we say, a-ha, that's why you should read the Bible,
  • 50:13 - 50:17
    and that's just about the wisest thing ever said.
  • 50:17 - 50:23
    But then we, and "we" includes Fundamentalists and Orthodox Jews, we ignore the rest of the book;
  • 50:23 - 50:29
    we ignore the places in Deuteronomy, and Leviticus, and Second Samuel
  • 50:29 - 50:34
    where the most mind-numbing, theocratic barbarism is recommended.
  • 50:35 - 50:37
    We bowdlerize the book.
  • 50:37 - 50:42
    So, clearly, our moral tools are coming from outside the text.
  • 50:47 - 50:53
    There's also the fact that there are obvious contradictions between the world's faiths.
  • 50:53 - 50:55
    This is a map of world religion.
  • 50:57 - 51:00
    (This is an example I've taken from my colleague, Richard Dawkins.)
  • 51:00 - 51:03
    This is not how knowledge spreads over the face of the Earth.
  • 51:03 - 51:07
    There's no reason, if you are in the business of understanding truths of any kind,
  • 51:07 - 51:11
    truths about human well-being, truths about the Universe,
  • 51:11 - 51:15
    you wouldn't expect your belief system to hug national boundaries in this way.
  • 51:15 - 51:22
    And no one, i think... Just imagine India. Is it possible that they believe,
  • 51:22 - 51:29
    the billion Hindus in India at the moment, that they alone, among all the worlds' people
  • 51:33 - 51:36
    (they could probably build me a better remote, there's no doubt about that),
  • 51:36 - 51:39
    [Laughter]
  • 51:39 - 51:47
    do they even think that they know that Ganesh, the elephant-headed god, exists and must be worshiped?
  • 51:47 - 51:54
    This is not the way we're discovering things about the nature of the world,
  • 51:54 - 51:59
    that people are accidentally born into the right belief system by dint of geography.
  • 52:05 - 52:10
    Apart from the contradictions between faiths, there are impressive patterns of contradiction within every faith.
  • 52:10 - 52:15
    Each one of these red arcs is linking two verses in the Bible,
  • 52:15 - 52:18
    that are just deal-breakers for omniscience.
  • 52:18 - 52:27
    These are statements that the Bible is just self-refuting in hundreds of places.
  • 52:27 - 52:33
    Jesus was crucified the day before the Passover meal;
  • 52:33 - 52:35
    Jesus was crucified the day after the Passover meal.
  • 52:35 - 52:40
    You just cannot make sense of these two claims.
  • 52:43 - 52:52
    There's also the inconvenient fact that the most important and easily resolved moral conundrums,
  • 52:52 - 52:57
    are conundrums that the Creator of the Universe apparently gets wrong.
  • 52:57 - 53:02
    So, slavery is, perhaps, the most consequential and easiest moral problem
  • 53:02 - 53:06
    we've ever had to confront. Slavery is supported in the Bible,
  • 53:06 - 53:09
    both in the old Testament and the new.
  • 53:09 - 53:15
    The god of Abraham never envisioned a time, where human beings ceased to keep slaves.
  • 53:15 - 53:19
    That is a fact, and then Jefferson Davis, the President of the Confederacy,
  • 53:19 - 53:24
    was right to point out that theology was on the side of the slaveholders.
  • 53:24 - 53:28
    It was despite theology that we got rid of slavery.
  • 53:31 - 53:36
    Now, I want to talk to you for a minute, before I wrap up, about moral intuition
  • 53:36 - 53:38
    and how it can seem to be confounded.
  • 53:38 - 53:44
    Because people take the difficulties we have in answering moral problems,
  • 53:44 - 53:47
    as a sign that there can be no such a thing as 'moral truth',
  • 53:47 - 53:51
    and I want to argue that that is, in fact, a sign of no such a thing.
  • 53:51 - 53:55
    This is the 'trolley problem', which some of you have probably seen.
  • 53:55 - 54:01
    A trolley is coming down the tracks. If you do nothing, it will hit and kill 5 people.
  • 54:01 - 54:03
    But you stand at the switch, and you can throw the switch
  • 54:03 - 54:07
    diverting the trolley onto another track, so that it kills only one person.
  • 54:07 - 54:12
    So, you'd save a net 4 lives.
  • 54:14 - 54:18
    Presented this way, something like 95% of people think, "you absolutely have to throw that switch."
  • 54:18 - 54:21
    You'd be a monster not to throw the switch.
  • 54:21 - 54:24
    But the problem is, you can present it another way:
  • 54:24 - 54:26
    the trolley is coming down the track now,
  • 54:26 - 54:31
    but you stand on a foot bridge, and beside you is a suitably large fellow
  • 54:31 - 54:33
    who you can push into the path of the oncoming trolley,
  • 54:33 - 54:39
    killing him, unfortunately, but saving 5 workmen and a net 4 lives.
  • 54:39 - 54:44
    When presented this way, everyone's intuition seems to flip,
  • 54:44 - 54:49
    and something like 95% of people will think, "you'd be a monster to push that man."
  • 54:50 - 54:56
    I think, this is actually somewhat ill-posed because, I think, we have an intuitive physics
  • 54:56 - 55:00
    and most of us burn a lot of fuel, worrying about whether the guy's really gonna stop the trolley.
  • 55:00 - 55:02
    [laughter]
  • 55:05 - 55:07
    But even if you stipulate that he will stop the trolley,
  • 55:07 - 55:10
    and these are truly identical outcomes,
  • 55:10 - 55:14
    it's still a problem for most people, and our intuitions get pushed around.
  • 55:14 - 55:19
    That is not a sign that there is no right answer to the 'trolley problem'.
  • 55:19 - 55:24
    I think, in fact, one of the details is that they are not equivalent,
  • 55:24 - 55:30
    and it may just be different to push someone up close and personal, than to throw a switch,
  • 55:30 - 55:33
    in terms of the consequences of everyone involved.
  • 55:33 - 55:37
    In one scenario, you could wake up with nightmares for the rest of your life;
  • 55:37 - 55:40
    and the other, you could think you're a hero,
  • 55:40 - 55:44
    and if that's just a difference of human psychology, we have to take that into account
  • 55:44 - 55:48
    in our evaluation of consequences.
  • 55:48 - 55:50
    But consider what we do with our logical intuitions.
  • 55:50 - 55:53
    This is the Monty Hall problem.
  • 55:53 - 55:55
    How many of you have seen the Monty Hall problem?
  • 55:55 - 55:59
    OK, this is Berkeley, I'm just taking coals to Newcastle here...
  • 55:59 - 56:00
    [some laughter]
  • 56:00 - 56:06
    So you're on a game show, and you are given a choice of three doors:
  • 56:06 - 56:09
    behind one door is a new car, behind the other two are goats,
  • 56:09 - 56:13
    and you pick door number one (obviously, if you pick the car, you get to keep it).
  • 56:13 - 56:19
    You pick door number one, and Monty Hall then opens door number two,
  • 56:19 - 56:22
    revealing a goat and he gives you a choice
  • 56:22 - 56:25
    to switch your bet to door number 3.
  • 56:25 - 56:27
    How many of you think you should switch?
  • 56:27 - 56:30
    How many of you think there's no reason to switch?
  • 56:30 - 56:40
    OK, there's a powerful intuition that many people, certainly many naive people share,
  • 56:40 - 56:46
    that there's no reason to switch. There are two doors: we've got a car behind one,
  • 56:46 - 56:49
    a goat behind the other, why would you switch? It's a coin toss.
  • 56:49 - 56:53
    It turns out, you should switch - and your chances double if you switch -
  • 56:53 - 56:56
    because you had a one-third chance, when you picked door number one,
  • 56:56 - 57:00
    and now the [remaining] two-thirds chance has fully collapsed onto door number 3.
  • 57:00 - 57:06
    But even very smart people, even mathematicians, people who understand probability theory,
  • 57:06 - 57:11
    can just get led back into thinking, "Wait a minute, there are two doors; there's a car,
  • 57:11 - 57:13
    there's a goat, why do you switch?"
  • 57:13 - 57:19
    And now, incidentally, it is easier to see, if you imagine being confronted
  • 57:19 - 57:24
    with a thousand doors. And you pick door number one, and then Monty Hall nullifies
  • 57:24 - 57:31
    998 doors, leaving door 576. Here, it's pretty obvious you've been given a tonne of information,
  • 57:31 - 57:39
    and switching makes sense. The point, however, is that the fact that our intuitions get pushed around,
  • 57:39 - 57:43
    never leads anyone to say, "maybe there's no right answer to the Monty Hall problem",
  • 57:43 - 57:45
    "maybe there's no such a thing as logical truth."
  • 57:45 - 57:48
    There's not... we're never tempted to do that.
  • 57:48 - 57:53
    There are perceptual illusions that are truly reliable.
  • 57:53 - 57:56
    Anyone who's got a clear shot of this screen,
  • 57:56 - 58:00
    sees the tower on the right leaning further to the right than the tower on the left.
  • 58:00 - 58:03
    And yet, these are the same photograph.
  • 58:03 - 58:09
    We can get behind our failures of intuition in science.
  • 58:09 - 58:15
    Our failures of intuition tell us a lot about how we are wired - in this case,
  • 58:15 - 58:17
    about how the visual system is organised.
  • 58:17 - 58:22
    But there's no question that just as there are perceptual illusions,
  • 58:22 - 58:27
    there are moral illusions, and these are illusions that we have to find a way to get behind.
  • 58:27 - 58:30
    This is based on research by Paul Slovic.
  • 58:30 - 58:36
    He asked groups of people: How much would you give to help a little girl in need?
  • 58:36 - 58:41
    When you ask people that question, you get a maximum rating of empathy
  • 58:41 - 58:48
    and a maximum donation, and if you ask a group: "How much would you give to help a little boy in need?"
  • 58:48 - 58:51
    you get the same response; maximum empathy, maximum donation.
  • 58:51 - 58:59
    The problem, however, is that when you ask them how much they would give to a little girl and to a little boy in need,
  • 58:59 - 59:06
    you get a 25% diminishment in both self-reported empathy and material donation.
  • 59:06 - 59:12
    Our concern goes down by a quarter by adding another child.
  • 59:12 - 59:19
    This is clearly not a normative result, this is a bug, not a feature.
  • 59:19 - 59:21
    [laughter]
  • 59:21 - 59:25
    I mean, if you care about a little girl and you care about a little boy,
  • 59:25 - 59:28
    you should care, at least, as much about their combined fate.
  • 59:28 - 59:36
    And it gets worse; the more kids you add, the more altruism and empathy diminishes.
  • 59:36 - 59:47
    This explains what Slovak has called 'genocide neglect', this fact about ourselves,
  • 59:47 - 59:53
    that many of us have noticed, that we find genocides boring.
  • 59:53 - 60:03
    We do not have the attentional and emotional resources to pay attention to the greatest occasions of human misery,
  • 60:03 - 60:08
    and yet we have endless resources to pay attention to the story of one little girl
  • 60:08 - 60:13
    trapped in a well. This is an image that perhaps many of you are too young to
  • 60:13 - 60:17
    recognize, but this is baby Jessica pulled from a well.
  • 60:17 - 60:22
    For something like 120 hours she was in this well and there was just
  • 60:22 - 60:25
    wall-to-wall television coverage of it.
  • 60:25 - 60:28
    Everyone with a television was desperate to see how this was going to turn out.
  • 60:28 - 60:32
    And yet 800 thousand people can be hacked to death in Rwanda
  • 60:32 - 60:39
    and even if it makes the news we can barely pay attention to it.
  • 60:39 - 60:43
    This is something we have to... We have to engineer our better selves
  • 60:43 - 60:49
    into our laws and social institutions, so that we can protect ourselves
  • 60:49 - 60:51
    from our moment-to-moment failures of moral intuition.
  • 60:54 - 61:00
    So, to conclude, I would point out that the moment we admit that
  • 61:00 - 61:03
    there are right answers to questions of human well-being and that
  • 61:03 - 61:07
    morality relates to this domain of facts,
  • 61:07 - 61:11
    then we have to admit that certain people care about the wrong things.
  • 61:11 - 61:16
    Which is to say that certain people, certain individuals, subcultures,
  • 61:16 - 61:22
    and even whole cultures perhaps, care about things that reliably produce
  • 61:22 - 61:25
    needless human misery.
  • 61:25 - 61:31
    And it seems to me that the only way we're going to converge
  • 61:33 - 61:38
    on a truly common project and build a global civilization, in which we
  • 61:38 - 61:47
    can live toward the same shared values, is to admit that this intellectual terrain exists,
  • 61:47 - 61:53
    to admit that morality relates to questions of well-being
  • 61:53 - 61:56
    and well-being relates to how we are, at some basic level,
  • 61:56 - 62:00
    and how the universe is. Thank you very much.
  • 62:00 - 62:09
    [Loud Applause]
  • 62:10 - 62:11
    Thank you.
  • 62:26 - 62:28
    Thank you very much.
  • 62:33 - 62:34
    [Questioner 1] I guess I get to be first here.
  • 62:34 - 62:40
    [Q1] Sam, in many of the speeches you get sort of the same thing, you know,
  • 62:40 - 62:45
    there's some quest for some type of universal truth based on science
  • 62:45 - 62:49
    and then there's also putting down all the other moral systems that
  • 62:49 - 62:53
    religion has that obviously don't work, you know, completely,
  • 62:53 - 62:56
    some work to some extent, some work to the other.
  • 62:56 - 63:02
    But did you know that there's already a system written of
  • 63:02 - 63:07
    morality based on science? If you look it up, you know you can put yourself like
  • 63:07 - 63:11
    way ahead of the game because it's already been done.
  • 63:11 - 63:12
    [Sam] What is that system?
  • 63:12 - 63:14
    [Q1] Church of reality!
  • 63:14 - 63:15
    [Sam] Church of reality, okay...
  • 63:15 - 63:21
    [Q1] Church of reality. Look at the website. It's something I started about 12 years ago.
  • 63:21 - 63:23
    [Sam] I've heard of the Church of reality but...
  • 63:23 - 63:26
    [Q1] Okay, you gotta read the website because it's more than just the name.
  • 63:26 - 63:31
    It started out as just the name but then it occurred to me like it occurred to you that...
  • 63:31 - 63:35
    [Sam] Well I'm inclined to shorten it a bit, to just "reality."
  • 63:35 - 63:39
    [Applause]
  • 63:42 - 63:44
    Then I think we can talk about it.
  • 63:44 - 63:47
    [Q1] I know people turn off when they hear the word "church",
  • 63:47 - 63:50
    but I put the word "church" in there because I wanted to get the attention of
  • 63:50 - 63:54
    the religious world, so that I can win souls for Darwin.
  • 63:54 - 63:58
    [Sam] Well, I mean we're clearly on the same team and I wish you the best of luck with it.
  • 63:58 - 64:01
    Okay, but here's the thing about it is that...
  • 64:01 - 64:06
    [Sam] Apologies, but there's a very long line forming behind you.
  • 64:06 - 64:08
    But everyone can go to the website.
  • 64:08 - 64:13
    [Q1] If you read the site, I have built a complete moral system based on science.
  • 64:13 - 64:16
    Read the site.
  • 64:16 - 64:18
    [Sam] Thank you.
  • 64:18 - 64:22
    [Q2] Bertrand Russell said philosophy is basically gorging upon the stew
  • 64:22 - 64:26
    of every conceivable idea and recently Peter Hacker had an article that
  • 64:26 - 64:31
    followed that up, dealing with how language kind of confounds a lot of our
  • 64:31 - 64:33
    problems by creating nonsense...
  • 64:33 - 64:34
    [Sam] Right.
  • 64:35 - 64:40
    [Q2] How much of that do you think is sort of inherent in the whole problem with
  • 64:40 - 64:43
    creating a science out of religion, and that a lot of people who have objections
  • 64:44 - 64:49
    seem to be basing their objections on things that really are more
  • 64:49 - 64:51
    tricks of language than actual?...
  • 64:51 - 64:53
    [Sam] Yeah, yeah, it's a huge problem, It's a...
  • 64:53 - 65:00
    People have associations with words that are very difficult to correct for
  • 65:00 - 65:06
    in an elegant way and... For instance... and it's true in science as well.
  • 65:06 - 65:12
    I run into scientists who think that the difference between subjective and objective
  • 65:12 - 65:18
    is hugely important and that subjectivity in some sense can never be understood scientifically.
  • 65:18 - 65:22
    It happens more among physicists than...
  • 65:24 - 65:28
    It's just that how would we ever understand what it's like to be another person?
  • 65:28 - 65:35
    And, there are some very philosophically unsophisticated views even among
  • 65:35 - 65:37
    very smart scientists.
  • 65:41 - 65:45
    This is a crucial distinction that I'll just briefly go into:
  • 65:45 - 65:50
    We use words like "subjective" and "objective" in two very different ways.
  • 65:50 - 65:58
    Actually, a local giant here John Surrel is the first to make this point, at least in my hearing.
  • 65:58 - 66:05
    We talk about subjectivity and objectivity ontologically, which is to say,
  • 66:05 - 66:09
    in terms of what exists, but we also talk about it epistemologically in terms of
  • 66:09 - 66:13
    how we know things. Ontologically there are just objective facts,
  • 66:13 - 66:17
    third person facts about the physical world, and there are subjective facts,
  • 66:17 - 66:21
    facts about what it's like to be a certain conscious creature,
  • 66:21 - 66:27
    and we can talk about both of those honestly and searchingly and in the context of science.
  • 66:27 - 66:34
    The epistemological subjective-and-objective difference is where bias
  • 66:34 - 66:38
    and personal views, and "I like chocolate and you like vanilla" and...
  • 66:38 - 66:45
    "how should I cut my hair", and all of this, those kinds of questions, merely personal,
  • 66:45 - 66:53
    merely subjective, the notion of subjectivity that diminishes truth claims,
  • 66:53 - 66:59
    that has nothing to do with studying the nature of any conscious system.
  • 66:59 - 67:03
    We're using the difference between subjectivity and objectivity in
  • 67:03 - 67:07
    very confusing ways, and so people think: "Well, it's just merely subjective,
  • 67:07 - 67:10
    that's just your opinion. Who can say what a good life is?"
  • 67:11 - 67:17
    It doesn't make any sense when you talk about physical health, and it
  • 67:17 - 67:21
    shouldn't make any sense when you talk about psychological health or the health of societies.
  • 67:23 - 67:27
    [Q3] I have two hopefully brief questions, I hope you'll answer at least one.
  • 67:27 - 67:31
    I think the point evangelist christians might make if they ever read your book,
  • 67:32 - 67:37
    which is wishful thinking, I agree, is this: even if there are theories of morality that
  • 67:37 - 67:41
    do not depend on God, as you clearly demonstrate, they'll say they're not relevant.
  • 67:41 - 67:45
    Why? Because without God to punish them, people will do whatever they like.
  • 67:45 - 67:49
    Dostoyevsky, I'm sure people know, is often quoted as writing
  • 67:49 - 67:54
    "If God is not then all is permitted." So you still need religion, they would say,
  • 67:54 - 67:58
    or at least a history of religion, even now in godless Denmark, for human beings
  • 67:58 - 68:01
    to behave themselves. So how would you respond to that? That's question one...
  • 68:01 - 68:06
    [Sam] Well, actually I just gotta limit you to one, cause it's just the line behind you is daunting.
  • 68:10 - 68:15
    One, it's not true, because clearly atheists can be motivated to be good in the same way,
  • 68:15 - 68:20
    and in fact the most secular societies on Earth, in fact the most atheistic societies on Earth
  • 68:20 - 68:25
    right now, are societies in Western Europe that are characterized by highly moral behavior
  • 68:25 - 68:28
    by any index you would use.
  • 68:28 - 68:29
    [Q3] But they have a history of religion.
  • 68:29 - 68:32
    [Sam] Okay, but they've shed their history and they seem to be better for it.
  • 68:32 - 68:35
    So that's inconvenient for the thesis, but...
  • 68:35 - 68:39
    [Applause]
  • 68:40 - 68:46
    Clearly, there is a challenge here for us to want the right... I mean there
  • 68:46 - 68:51
    are people who are not capable of wanting what they should want, and there are psychopaths.
  • 68:51 - 68:55
    They are people who have brain damage which we're beginning to understand.
  • 68:55 - 69:02
    And they don't feel empathy for other people. Okay, so there's... certain people are not
  • 69:02 - 69:10
    up to the challenge of living a wise and ethical life. And we have to understand that.
  • 69:10 - 69:17
    There are cultures that make their ways of raising children,
  • 69:17 - 69:22
    ways of talking to one another in the public sphere, and institutional mechanisms to put in place,
  • 69:22 - 69:28
    that can encourage the greatest number of people to want to collaborate freely,
  • 69:28 - 69:33
    and not oppressively, and creatively, with everyone else, and thats the challenge for us.
  • 69:33 - 69:38
    We have to build a global civilization that allows most of the people who want to do that, to do that,
  • 69:38 - 69:45
    and there are some obvious principles that are so obvious that they basically should be non-negotiable.
  • 69:45 - 69:50
    Things like free speech, and the rights of women, you know. Those are
  • 69:50 - 69:57
    probably not on the table to be doubted, and yet, so many societies don't even have that.
  • 69:57 - 70:02
    The most basic rudiments of building a sane sphere for public discourse.
  • 70:02 - 70:12
    So we have to... kind of the first pass of what we should do, is what any sane person would want to do,
  • 70:12 - 70:19
    given all the facts, and the trade-offs between individuals and the collective,
  • 70:19 - 70:25
    or between free speech VS privacy, all of these things, there are difficult ethical dilemmas that
  • 70:25 - 70:32
    we can run into at the margins, or in our lives personally. But the biggest moves for us to make,
  • 70:32 - 70:39
    as whole cultures, are so obvious, and they're moves that would lift every boat with the same tide.
  • 70:39 - 70:44
    I mean, stopping nuclear proliferation, stopping our contribution to climate change,
  • 70:44 - 70:50
    stopping the causes of war, stopping pandemics, all these things are good for everybody.
  • 70:53 - 71:02
    So you don't need divisive religious dogmatism to help that project along.
  • 71:02 - 71:05
    In fact, it's one of the most obvious things standing in its way, I think.
  • 71:05 - 71:07
    [Q3] Thank you.
  • 71:09 - 71:13
    [Q4] The proof of concept visual you showed about the moral landscape...
  • 71:15 - 71:20
    How do you think we can take... how do you think scientists, your peers or yourself,
  • 71:20 - 71:26
    could get inspiration from that and actually come up with, say, a mathematical model
  • 71:26 - 71:31
    or framework that will be ever-evolving for sure? Something that we can start with,
  • 71:31 - 71:36
    so we can get something actionable out of the scientific research. I mean, I can't wait for the day
  • 71:36 - 71:40
    where I could I go the BBC news website and be able to see:
  • 71:40 - 71:47
    "The moral finding of the week" with the landscape visual there.
  • 71:47 - 71:56
    [Sam] I think it's, you know, in detail it would certainly take a long time, but I think the most important moves,
  • 71:56 - 72:01
    again, don't even require more data. We know that throwing battery acid
  • 72:01 - 72:05
    in the face of little girls who want to learn to read, is not a good thing to do.
  • 72:05 - 72:09
    We don't have to scan anybody's brain to figure out that that's not really compassion,
  • 72:09 - 72:11
    and that's not really good for society.
  • 72:14 - 72:17
    The core move is to admit that there are right answers
  • 72:17 - 72:21
    to the question of how human beings can live lives worth living.
  • 72:21 - 72:26
    [Q4] I'm just appealing to the visual market ability of...
  • 72:26 - 72:33
    [Sam] I think it does... There are many things that fall out of it. One is just the many peaks,
  • 72:33 - 72:37
    as I've said, but there are other dynamics of it, there is the fact that
  • 72:37 - 72:41
    we might often have to move downward in order to move upward to a higher place,
  • 72:41 - 72:47
    and I think there's an evolutionary argument, for instance, that altruism could only have formed
  • 72:47 - 72:51
    in communities that were warring with other communities.
  • 72:51 - 72:56
    And so our core moral tool at the moment,
  • 72:56 - 72:59
    of caring to collaborate with one another within a group,
  • 72:59 - 73:05
    is something that could've only been born, if Samuel Bowles is right,
  • 73:05 - 73:10
    by first kind of descending into this valley of
    internecine struggle.
  • 73:10 - 73:13
    Now, if that's true then so be it. But clearly that's not our circumstance now.
  • 73:13 - 73:19
    I understand your question. I just think we have to, one, admit that
  • 73:19 - 73:24
    human well-being is an intelligible subject to be studied scientifically.
  • 73:24 - 73:30
    That has started to happen in psychology and neuroscience, but we then have to admit
  • 73:30 - 73:36
    that whatever we find there has truly trans-cultural consequences,
  • 73:36 - 73:44
    in the same way that facts about human health do, and if culture does make a contribution,
  • 73:44 - 73:49
    which I think it certainly does, if culture changes us in ways that are relevant to human well-being,
  • 73:49 - 73:54
    it does that by changing our brains. I mean that still it's realized at the level of the brain
  • 73:54 - 73:57
    and we can understand that in science, but it's gonna be a long time before we have
  • 73:57 - 74:00
    a lot of detail. I'm just trying to get the project motivated.
  • 74:00 - 74:01
    [Q4] Thanks.
  • 74:03 - 74:07
    [Q5] I apologize for the vagueness of the question but... I don't stay up at night wondering
  • 74:07 - 74:13
    about whether 2 plus 2 really equals 4 but I do stay up at night wondering about
  • 74:13 - 74:17
    how much I should donate to charity and, you know, what it does mean to live a good life.
  • 74:17 - 74:22
    I'm wondering if you can envision a point at which the certainty that we have about
  • 74:22 - 74:26
    mathematical truth... Is it just a matter of the complexity of the issue, do you think,
  • 74:26 - 74:33
    that begets this kind of, you know, difficulty in understanding these things or is it?...
  • 74:33 - 74:35
    Are there other factors?
  • 74:35 - 74:40
    [Sam] Yeah, a good question, that there are... There are a few factors, one is that there's...
  • 74:40 - 74:48
    Understanding it in terms of science, clearly it's a much more complex issue and
  • 74:48 - 74:51
    one analogy is economics, you know. When is economics gonna be a science?
  • 74:51 - 74:56
    Who knows? I mean, clearly, we're in a position to be just surprised by the dynamics
  • 74:56 - 75:00
    of economic systems, and maybe that's always gonna be true,
  • 75:00 - 75:05
    but no one would be tempted to say "Well, there's just no truth there",
  • 75:05 - 75:11
    or "It's a sign of bigotry to criticize somebody's response to a banking crisis".
  • 75:11 - 75:15
    I mean, clearly, we're operating in a situation where there are truths and we just don't know them,
  • 75:15 - 75:17
    and we are worried.
  • 75:17 - 75:23
    There's an analogous issue there with morality. But, speaking personally,
  • 75:23 - 75:30
    it's nowhere written that it's easy to be good, you know, and we have other motives.
  • 75:30 - 75:35
    We're selfish and we want certain things that we know we probably shouldn't want,
  • 75:35 - 75:41
    which is to say we would be happier and we wouldn't regret it if we could overcome these wants,
  • 75:41 - 75:45
    you know, but we still want. We want to lose weight but we also want a hot fudge sundae.
  • 75:45 - 75:54
    So we're multiform in our motives. And I think the most important move for all of us to
  • 75:54 - 76:03
    make is in our most reflective moments to come up with an honest opinion about what should happen,
  • 76:03 - 76:08
    and then to engineer that at the level of society. So, if our tax code
  • 76:08 - 76:14
    dealt with the problem of how much we should be giving to, homelessness, say,
  • 76:14 - 76:19
    we wouldn't be having to recalculate every minute about whether we give or we don't give.
  • 76:19 - 76:25
    We have to solve the problem of homelessness, to take one problem among a myriad.
  • 76:27 - 76:31
    There are more and less intelligent ways to address it, we clearly haven't discovered the way
  • 76:31 - 76:38
    that is actionable, that is really gonna work. But the idea that there's no way to address it,
  • 76:38 - 76:44
    seems frankly crazy, and the idea that we're all left with just deciding whether to take a dollar
  • 76:44 - 76:50
    out of our pocket, that's clearly not the remedy. Yet that's the kind of problem that each of us,
  • 76:50 - 76:55
    in the privacy of our own lives, worries about in terms of our own ethical responsibility.
  • 76:55 - 76:59
    I think the bigger swings are going to happen at the level of what we engineer
  • 76:59 - 77:00
    at the level of society.
  • 77:00 - 77:05
    But it's a hard problem, it's just the cold one we have to solve moment to moment.
  • 77:08 - 77:13
    [Q6] Mark Lewis, optimalhumanvalues.com. Thank you for your work, Sam, in this area.
  • 77:15 - 77:19
    I see two central challenges to your basic premise or your basic conclusion,
  • 77:19 - 77:21
    that science can determine human values.
  • 77:21 - 77:26
    One is the philosophical case, which I believe you address relatively effectively in your book.
  • 77:26 - 77:33
    The second is, we could say, a fear that if we cede the realm of values to science,
  • 77:33 - 77:39
    that science will make a mistake and tell us to value the wrong thing,
  • 77:39 - 77:43
    and in the process, destroy things that are truly valuable and important.
  • 77:45 - 77:50
    The challenge that that poses, seems to me, from the people I've spoken with about your book,
  • 77:50 - 77:55
    has them not want to face the arguments in your book, because they're afraid that
  • 77:55 - 77:59
    if they listen to your arguments then they'll have to see that...
  • 77:59 - 78:03
    [Sam] People have creepy associations with the word "science",
  • 78:03 - 78:07
    and they think what I'm advocating is a kind of brave new world scenario,
  • 78:07 - 78:11
    where everyone's just going to be medicated with the right drug.
  • 78:14 - 78:21
    Science is our truly open conversation in which we are most constrained by
  • 78:21 - 78:25
    honest observation and clear reasoning.
  • 78:25 - 78:29
    It's when we make our best effort to get our biases out of the way,
  • 78:29 - 78:33
    and our wishful thinking out of the way, and just talk honestly about
  • 78:33 - 78:34
    what we know and what we don't know.
  • 78:37 - 78:42
    The suspicion that that might... it's a very strange intuition,
  • 78:42 - 78:48
    that the most important questions in human life must fall outside of science,
  • 78:48 - 78:52
    because what we're saying is that when you become most intellectually honest,
  • 78:52 - 78:55
    when you get your wishful thinking out of the way, when you get your biases out of the way,
  • 78:55 - 78:58
    when you rely upon clear reasoning and honest observation,
  • 78:58 - 79:03
    that's precisely the mood you can't be in to address the most important questions in human life.
  • 79:03 - 79:09
    Okay, that's weird and we should point that out.
  • 79:09 - 79:13
    [Applause]
  • 79:13 - 79:17
    There is no other mood to be in to address the most important questions.
  • 79:21 - 79:26
    Again, I don't define science narrowly. It's evidence-based rational discussion,
  • 79:26 - 79:33
    where people's convictions are going to scale with the quality of the arguments
  • 79:33 - 79:35
    and the quality of the evidence,
  • 79:35 - 79:38
    and that is really the antithesis of what happens in religion,
  • 79:38 - 79:44
    and it's what's brought to an exquisite refinement in science per se,
  • 79:44 - 79:47
    but it's also true of all intellectual discourse.
  • 79:47 - 79:55
    So it's not about getting your science... your fear is... or the fear you're expressing
  • 79:55 - 80:01
    could be applied to medicine, you know, we're all afraid that science is gonna get human health wrong,
  • 80:01 - 80:07
    and disease wrong and cancer wrong and... Well, it's possible to get these things wrong,
  • 80:07 - 80:12
    but the remedy for getting them wrong is always just better science,
  • 80:12 - 80:16
    it's understanding the facts more clearly,
  • 80:16 - 80:26
    and the antidote is never some other process of irrational faith-based claims
  • 80:26 - 80:28
    about the nature of reality. So...
  • 80:28 - 80:32
    [Q6] I hope you'll lead a forum to that end as well.
  • 80:32 - 80:34
    [Sam] Well, thank you for that vote of confidence.
  • 80:37 - 80:44
    [Q7] Hi, I did want to preface this by saying that I am not a muslim, I am not any religion at all,
  • 80:44 - 80:51
    but it bothered me when you were talking about Islam being not a religion of peace.
  • 80:54 - 80:56
    Mohammed was a very peaceful man...
  • 80:56 - 80:58
    [Sam] No, he wasn't.
  • 80:58 - 81:03
    [Q7] Yes, well if you read... yes... and he had a great respect for women.
  • 81:03 - 81:09
    His wives were given a great deal of power over him...
  • 81:09 - 81:13
    [Sam] Have you been reading Karen Armstrong? Is that where this is coming from?
  • 81:13 - 81:15
    [Q7] No...no, I've read Karen Armstrong, but no.
  • 81:15 - 81:21
    You know, what's happening now with Islam, is it's bastardized,
  • 81:21 - 81:27
    just like all the other religions, but it bothers me that you kind of single...
  • 81:27 - 81:29
    [Sam] So you're saying that the true Islam, if we could only return to it,
  • 81:29 - 81:33
    would just nullify all of my concerns about Islam.
  • 81:34 - 81:42
    [Q7] I don't think any religion is a good religion, but it bothers me that you singled Islam out
  • 81:42 - 81:52
    and that you said that... actually that Mohammed was a peaceful man, and that...
  • 81:52 - 81:58
    [Sam] Mohammed, whoever he may be... who knows who Mohammed was...
  • 81:58 - 82:01
    I mean, Mohammed was actually closer to history than Jesus and many of the other patriarchs,
  • 82:01 - 82:02
    and we know more about him,
  • 82:02 - 82:07
    but obviously there's a lot of uncertainty about what's factual,
  • 82:07 - 82:16
    but the example of Mohammed as held in Islam universally is not of a pacifist.
  • 82:16 - 82:22
    He was a conquering warlord who spread the faith with the sword quite successfully,
  • 82:22 - 82:31
    and the expectation is, this is a way of being in the world that is by example totally justified.
  • 82:31 - 82:37
    Now, this is different from Jesus. Jesus did not spread the faith with a sword,
  • 82:37 - 82:42
    he was essentially a hippy who got crucified. That's a different example,
  • 82:42 - 82:49
    and it's a difference that is a benefit to christianity at the moment,
  • 82:49 - 82:53
    because in christianity you can come up with a rationale for saying:
  • 82:53 - 82:59
    "Listen, it's not about conquering the world, it's not about winning in this life.
  • 82:59 - 83:02
    We just gotta wait for Jesus to come back and get raptured."
  • 83:05 - 83:07
    You can't really do that in Islam.
  • 83:07 - 83:12
    Christianity has a line like "render unto God that which is God's and unto Caesar that which is Caesar's"
  • 83:12 - 83:15
    and that line has done, from Matthew, has done huge work,
  • 83:15 - 83:21
    to separate christianity from claim upon terrestrial power.
  • 83:21 - 83:25
    Now, it's imperfect work and there are a lot of crazy christians who want terrestrial power,
  • 83:25 - 83:29
    but at least there's some rationale within christianity for doing that.
  • 83:29 - 83:33
    That rationale doesn't exist in Islam. There's no line in the Koran which says:
  • 83:33 - 83:38
    "Listen, guys, this is not about politics, it's not about controlling people's lives,
  • 83:38 - 83:42
    you can be privately religious and let everyone else flourish."
  • 83:42 - 83:46
    [Q7] But the subjugation of women was not something that Mohammed was...
  • 83:46 - 83:49
    [Sam] You're just not... please, just go read the Koran...
  • 83:49 - 83:51
    [Q7] I've read the Koran.
  • 83:51 - 83:56
    [Sam] Okay, then surely you read the part in the Koran which talked about husbands scorching
  • 83:56 - 84:00
    or whipping (depending on the translation) their wives who were disobedient.
  • 84:00 - 84:02
    [Q7] That's in judaism also.
  • 84:02 - 84:09
    [Sam] What was that?... Okay, so it's in judaism but... I will grant you that the worst books ever written
  • 84:09 - 84:12
    were in the Old Testament. Okay?
  • 84:12 - 84:19
    But... there's a reason why Jews are not stoning their wives for adultery,
  • 84:19 - 84:22
    at this moment in history, and we can talk about those reasons.
  • 84:22 - 84:30
    What I said about Islam was intended to counter exactly the presuppositions you are now bringing to me,
  • 84:30 - 84:36
    I don't have time to do it here, but I can only invite you to read more on the subject,
  • 84:36 - 84:41
    because we are deluding ourselves with a lot of wishful thinking.
  • 84:41 - 84:47
    We are desperate to believe that all of the problems in the world are of our own making,
  • 84:47 - 84:52
    if we could just spread more money around and behave better on the global stage,
  • 84:52 - 84:58
    people will treat us well, and that everyone wants the same thing
  • 84:58 - 85:02
    and it's all a matter of just more education and this religion is intrinsically benign.
  • 85:02 - 85:05
    It's just not true. And... Osama Bin Laden...
  • 85:05 - 85:10
    I've said this but I think I should say it again, the real problem is
  • 85:10 - 85:13
    that Osama Bin Laden is giving a very plausible version of the faith.
  • 85:13 - 85:19
    Now you can split a few hairs and say: "Well, listen, apostates shouldn't be killed
  • 85:19 - 85:28
    unless they speak kind of endlessly against the faith." But the penalty for apostasy is death under Islam,
  • 85:28 - 85:32
    and there's no school of Islam which says: "Oh no no, we don't mind apostates."
  • 85:32 - 85:37
    That just doesn't exist. There is no 'reformed Judaism' version of Islam.
  • 85:37 - 85:39
    [Laughter]
  • 85:39 - 85:44
    That's a huge problem. It's a problem that has to be solved by muslims.
  • 85:44 - 85:49
    We can't help them solve it by just lying to ourselves about the nature of their religion.
  • 85:50 - 85:52
    [Q7] I also want to say one more thing...
  • 85:52 - 85:54
    [Host] I'm sorry, thank you...
  • 85:55 - 85:57
    [Host] Are you good to keep going?
  • 85:57 - 86:03
    [Sam] Yeah... Maybe... There's no way we'll get to the back of the line,
  • 86:03 - 86:08
    so I'm going to let you decide. I'm long-winded and I'm sorry.
  • 86:08 - 86:13
    [Q8] I'll try to be short-winded. Thank you for your books, they made a huge impact on me,
  • 86:13 - 86:18
    and your breaking down of taboos has been phenomenal.
  • 86:18 - 86:24
    In this book which I enjoyed very much, one thought kept nagging at me and I just want some clarification.
  • 86:24 - 86:25
    [Sam] Sure.
  • 86:25 - 86:31
    [Q8] When I think of this moral landscape and your visual in my mind, it seems to assume...
  • 86:31 - 86:38
    I think it doesn't assume, but intuitively it assumes that the elevation starts the same, right?
  • 86:38 - 86:42
    And the question you just brought up of resources, as some people think
  • 86:42 - 86:45
    you can just spread money around, kept coming to mind. And I thought of...
  • 86:45 - 86:48
    I'm sure you encountered Robert Sapolsky at Stanford, and his work,
  • 86:48 - 86:51
    when looking at the same species of monkey...
  • 86:51 - 86:56
    You know, we are ourselves biped primates, right? I thought of this example that he has
  • 86:56 - 87:00
    of these monkeys that in an arid barren environment take risks more,
  • 87:01 - 87:03
    are more aggressive, have a different social network,
  • 87:03 - 87:12
    and then in a more resource-rich environment they are more forgiving and equitable and so on.
  • 87:12 - 87:20
    I kept thinking, I'm with you so much on this project, but when I get to the prescriptive part,
  • 87:20 - 87:26
    we, I think, can say we can break down the fact-value distinction and say very clearly:
  • 87:26 - 87:36
    "We don't like people doing X, Y and Z either in a subculture domestically or internationally,
  • 87:36 - 87:43
    Taliban or anywhere else." However, are there situations in that moral landscape
  • 87:43 - 87:46
    where some community will have organized its...
  • 87:46 - 87:47
    [Inaudible shout]
  • 87:47 - 87:49
    [Q8] I'm sorry?
  • 87:49 - 87:50
    [Sam] Just continue.
  • 87:50 - 87:56
    [Q8] ... organized itself in some kind of optimal or suboptimal,
  • 87:56 - 88:02
    but somewhat utility-maximizing way, but it's abhorrent to us and how do we deal with that?
  • 88:02 - 88:03
    Resource question.
  • 88:03 - 88:09
    [Sam] Yeah, a good question. I think there are... Well, there are clearly islands
  • 88:09 - 88:14
    of sort of pathological happiness, the example from Auschwitz was one.
  • 88:14 - 88:19
    You get the guards of Auschwitz all together agreeing that they're loving life.
  • 88:19 - 88:25
    Clearly that can't be a peak on the moral landscape because all of that well-being,
  • 88:25 - 88:32
    such as it is, is predicated on just an immensity of suffering occasioned outside that circle,
  • 88:32 - 88:39
    and I think there are some obvious kinds of happiness that people like that are not
  • 88:39 - 88:45
    experiencing. In so far as compassion and the connection to other people
  • 88:45 - 88:51
    is a source of well-being, and I think there are probably frontiers for us all to discover there
  • 88:51 - 88:59
    just how happy and connected it's possible to be in the midst of others.
  • 88:59 - 89:07
    It's clearly not a peak, so we can grant that there are sort of weird areas where people,
  • 89:07 - 89:12
    based on isolation and based on what they've established culturally...
  • 89:12 - 89:18
    You can get an island of perfectly matched sadists and masochists, say,
  • 89:18 - 89:23
    where it's just... some people just like to be beaten on and some people just like to mistreat them
  • 89:23 - 89:27
    and they're just perfectly matched and pretty happy, you know.
  • 89:27 - 89:34
    But clearly it's not a peak, and clearly that kind of valuation of experience
  • 89:34 - 89:42
    isn't really well-packaged for export. [Laughter] And I think it's not an accident that...
  • 89:42 - 89:51
    I think we can converge, just as we converge on logical understanding of basic facts
  • 89:51 - 89:58
    and in a scientific understanding of the world. We don't converge perfectly, and obviously
  • 89:58 - 90:01
    in certain situations it's not even a majority of people who converge.
  • 90:01 - 90:06
    25% of Americans think evolution is a fact, the rest apparently don't,
  • 90:06 - 90:12
    but biology can still thrive in that context, and I think it's possible that there are moral truths
  • 90:12 - 90:17
    that even a majority may not be up to realizing, but they may still be true,
  • 90:17 - 90:22
    just as there are medical truths that the majority may not be up to understanding, but they're true,
  • 90:22 - 90:27
    and we have to find some way of empowering... just as we want to empower the biologist
  • 90:27 - 90:33
    to talk about real biology, even though they would get voted out of office if their neighbors could.
  • 90:33 - 90:39
    We have to empower the people who really understand the danger of nuclear proliferation, say,
  • 90:39 - 90:49
    and know that it's worthy of our attention and gay marriage isn't, to prioritize those things
  • 90:49 - 90:52
    at the level of public policy.
  • 90:55 - 91:01
    But yeah, I think there are probably weird places and weird things to pass through,
  • 91:01 - 91:11
    in order to advance but what is the alternative? All we have is human conversation
  • 91:11 - 91:18
    where we're trying to influence one another to share a common project
  • 91:18 - 91:21
    of peaceful cooperation.
  • 91:21 - 91:24
    [Q8] I guess it would refrain from some of the harsh judgement,
  • 91:24 - 91:27
    recognising that in situations of lack of resources...
  • 91:27 - 91:36
    [Sam] Oh, yeah, I'm sorry, I didn't deal with that piece. Clearly, material resources are a huge variable
  • 91:36 - 91:42
    in... it's Maslow's hierarchy of needs. If you're starving, you really don't have the free attention
  • 91:42 - 91:49
    to worry about whether stealing is right or wrong, whether you should value other people's children
  • 91:49 - 91:55
    as much as your children or how much should you discount their hunger over your children's.
  • 91:55 - 92:00
    You can't think about human happiness when you're starving and so we need...
  • 92:00 - 92:04
    and you can't think about human happiness when you're being chased by somebody wielding a machete;
  • 92:04 - 92:12
    you're just reacting. So we need to create the most basic conditions of stability
  • 92:12 - 92:18
    in any society, where people then have the free attention to worry about things like education
  • 92:18 - 92:26
    and creativity, and how they should act so as to maximize their well-being in society.
  • 92:26 - 92:27
    [Q8] Thank you.
  • 92:27 - 92:30
    [Host] Let's do two more quick questions because we want you to stay and sign books after this.
  • 92:30 - 92:31
    Sam: Okay.
  • 92:31 - 92:35
    [Q9] This is the 'brave new world' question.
  • 92:35 - 92:36
    [Sam] Okay.
  • 92:36 - 92:43
    [Q9] If I understand your premise that morality is based on the experience of human happiness
  • 92:43 - 92:48
    or the experience of the reduction of suffering, then how do you distinguish
  • 92:48 - 92:54
    a drug-induced experience of happiness from a reality-based one?
  • 92:54 - 93:01
    [Sam] Well, I think we want our subjective states of well-being to be coupled to reality
  • 93:01 - 93:06
    for very obvious reasons, because if they're not coupled to reality then they're vulnerable to
  • 93:06 - 93:12
    the next insult from reality. So why not just take heroin all day long? It's a very pleasant state to be in,
  • 93:12 - 93:17
    you can just keep doing it. Well, a problem is you can't just keep doing it and you can't value...
  • 93:17 - 93:22
    you can't attend to all the other things you value in life that are your sources of happiness,
  • 93:22 - 93:32
    like your relationships, like having a career, like not dying. It's not a stable source of well-being,
  • 93:32 - 93:41
    and most of what we care about in life, like loving other people and experiencing love in return,
  • 93:41 - 93:48
    is predicated on our states of consciousness actually tracking the reality of our lives,
  • 93:48 - 93:53
    not perfectly perhaps, but to a significant degree. You can't have real relationships if you're delusional,
  • 93:53 - 94:02
    or if you're stoned all the time. There's a... And this is something we actually will face.
  • 94:02 - 94:08
    We will get drugs that can really, for instance, take away grief. I think at some point,
  • 94:08 - 94:11
    in the lifetime of someone in this room, there will be an antidote to sadness.
  • 94:11 - 94:16
    So that when your wife dies and you're inconsolable, there will be a pill that you can take
  • 94:16 - 94:20
    that will take that feeling away. Now the question is when do you want to take it?
  • 94:20 - 94:25
    Do you want to take it 15 minutes after she dies? You want to take it a month after?
  • 94:25 - 94:29
    You want to take it a year after? You want to take it before she dies so that you'll be indifferent to her dying?
  • 94:29 - 94:31
    [Laughter]
  • 94:31 - 94:36
    These are real ethical problems.
  • 94:36 - 94:44
    [Q9] That movie, the Matrix, kind of created this scenario where the people
  • 94:44 - 94:53
    who were living in this drug-induced world - it was complete. They did not need reality to fulfill...
  • 94:53 - 94:56
    [Sam] OK, but that's not our situation in any foreseeable future.
  • 94:56 - 95:01
    If Ray Kurzweil is right... if the singularity is true and Ray Kurzweil is right,
  • 95:01 - 95:05
    and we're just going to upload ourselves onto the internet in 30 years or whatever it is,
  • 95:05 - 95:10
    well then that's a real challenge. The question is:
  • 95:10 - 95:18
    what connection to reality do we want if we can sort of disappear into this dreamscape
  • 95:18 - 95:24
    where our happiness can be perfectly maintained? My moral intuitions get a little shaky there,
  • 95:24 - 95:28
    but in some sense we are in a bit of a dreamscape already,
  • 95:28 - 95:35
    I mean, all of this is being run on our brains and we are... We have to creatively respond
  • 95:35 - 95:42
    to the opportunity to experience happiness and avoid suffering in this space.
  • 95:42 - 95:45
    [Q9] So there's a second logical premise, which is experience...
  • 95:45 - 95:51
    [Sam] I'm saying... The answer to your question is, in our context, we want to be truth-tracking
  • 95:51 - 95:57
    to a significant degree, for obvious reasons, because the moment we're not, we begin to suffer mightly.
  • 95:57 - 96:06
    If that ever ceases to be true then we can have a conversation about what reality we want to live in.
  • 96:06 - 96:15
    But at the moment we're living in this one and we have to understand how we're entangled with it
  • 96:15 - 96:16
    or we will suffer.
  • 96:16 - 96:18
    [Q9] Thank you.
  • 96:18 - 96:20
    [Host] Please stay seated in your seats for one last question.
  • 96:20 - 96:23
    [Q10] Kind of feel like the chosen one here.
  • 96:23 - 96:25
    [Laughter]
  • 96:25 - 96:29
    I want to ask a question because I'm trying to understand if you're claiming
  • 96:29 - 96:35
    that there is one moral truth for a given action or if there can be multiple opposing ones,
  • 96:35 - 96:41
    because I can imagine some sort of moral dilemma that can have a
  • 96:41 - 96:47
    positive well-being effect on one person while at the same time having a negative well-being
  • 96:47 - 96:47
    effect on the other.
  • 96:47 - 96:48
    [Sam] That's true.
  • 96:48 - 96:53
    [Q10] I'm trying to understand if you're claiming that there is only one. Can there be multiple?
  • 96:53 - 96:56
    Can they be opposing? Is it not absolute? Cause...
  • 96:56 - 97:00
    I'm sort of getting the impression that you're saying it's absolute.
  • 97:00 - 97:05
    [Sam] Well no, what I'm saying is that there... for that case there are clearly zero-sum moments
  • 97:05 - 97:09
    when there's, you know, one slice of pie left and only one person is gonna get it,
  • 97:09 - 97:14
    and if person A gets it, person B doesn't and vice versa.
  • 97:14 - 97:22
    I think there are probably right solutions, if we could understand human well-being
  • 97:22 - 97:26
    in a truly fine-grained way, there are probably right solutions to most of those...
  • 97:26 - 97:30
    I mean, either it doesn't matter who gets it in the scheme of things,
  • 97:30 - 97:35
    or it would be a little bit better if one person got it or it would be much better if one person got it.
  • 97:35 - 97:41
    So when it really matters, I think the trade-offs begin to get obvious.
  • 97:41 - 97:49
    But even there there are situations in which some people really want something,
  • 97:49 - 97:56
    and some people really want its antithesis, and there's gonna be some obvious suffering either way you play it.
  • 97:56 - 98:00
    [Q10] I'm thinking like a euthanization sort of situation,
  • 98:00 - 98:05
    where you have a lifetime of pain versus just ending someone's life.
  • 98:05 - 98:08
    I mean, never mind even the effects on other family members and things,
  • 98:08 - 98:15
    but you've got those two things; how does science tell us, based on your claim,
  • 98:15 - 98:20
    which one of those we should value more, or less bad I suppose I should say?
  • 98:22 - 98:30
    [Sam] Well, again it would be science narrowly described as a mature science of the human mind,
  • 98:30 - 98:34
    well then it would have a lot to say. We would be able to just know how much people suffer,
  • 98:34 - 98:42
    the person suffering will be able to say: "I'm suffering this much" and we'll have our suffering detector,
  • 98:42 - 98:46
    which says: "Yeah, they are suffering that much and boy does that suck!"
  • 98:46 - 98:48
    [Short laughter]
  • 98:48 - 98:53
    But short of that all we can do is talk honestly about the trade-offs here.
  • 98:53 - 98:58
    But what I'm really trying to fight for is the trade-offs are in terms of human well-being.
  • 98:58 - 99:04
    They're not in terms of something else. So, if you're gonna advance a moral argument,
  • 99:04 - 99:08
    you have to at least talk that talk, otherwise it's not a moral argument.
  • 99:08 - 99:15
    So, if you're going to oppose gay marriage, and say: "there are all these trade-offs, who knows what is true,"
  • 99:15 - 99:22
    at least your side of the argument has to be: "here is all the suffering that gay marriage is going to cause,
  • 99:22 - 99:27
    here is what's going to happen to children if they get adopted by gay people,
  • 99:27 - 99:28
    this is what's going to happen to..."
  • 99:28 - 99:31
    There's no burden on anyone to make that argument at the moment,
  • 99:31 - 99:35
    because cause we're living in a world where the president of the USA can say:
  • 99:35 - 99:40
    "My faith tells me that marriage is between a man and a woman." End of argument.
  • 99:40 - 99:47
    That's the move that should no longer be open to smart people, and certainly people with responsibility.
  • 99:47 - 99:51
    [Host] I'm sorry, we have to stop now. Thank you!
  • 99:54 - 99:57
    [Sam] Thank you very much.
Title:
Sam Harris: Can Science Determine Human Values?
Description:

In this highly anticipated, explosive new book, the author of The End of Faith and Letter to a Christian Nation calls for an end to religion's monopoly on morality and human values. In The Moral Landscape: How Science Can Determine Human Values, Sam Harris tears down the wall between scientific facts and human values to dismantle the most common justification for religious faith -- that a moral system cannot be based on science.

The End of Faith ignited a worldwide debate about the validity of religion. In its aftermath, Harris discovered that most people, from secular scientists to religious fundamentalists, agree on one point: Science has nothing to say on the subject of human values. Even among religious fundamentalists, the defense one most often hears for belief in God is not that there is compelling evidence that God exists, but that faith in Him provides the only guidance for living a good life. Controversies about human values are controversies about which science has officially had no opinion. Until now.

Morality, Harris argues, is actually an undeveloped branch of neuroscience, and answers to questions of human value can be visualized on a "moral landscape" -- a space of real and potential outcomes whose peaks and valleys correspond to human states of greater or lesser wellbeing. Different ways of thinking and behaving -- different cultural practices, ethical codes, modes of government, etc. -- translate into movements across this landscape. Such changes can be analyzed objectively on many levels, ranging from biochemistry to economics, but they have their crucial realization as experiences in the human brain.

Bringing a fresh, secular perspective to age-old questions of right and wrong, and good and evil, Harris shows that we know enough about the human brain and its relationship to events in the world to say that there are right and wrong answers to the most pressing questions of human life. Because such answers exist, cultural relativism is simply false -- and comes at increasing cost to humanity. And just as there is no such thing as Christian physics or Muslim algebra, there can be no Christian or Muslim morality. Using his expertise in philosophy and neuroscience, along with his experience on the front lines of our "culture wars," Sam Harris delivers a game-changing argument about the future of science and about the real basis of human cooperation.

11.10.10
Berkeley Arts and Letters
First Congregational Church of Berkeley (2345 Channing Way at Dana, Berkeley)
Berkeley, CA

Sam Harris is the author of the New York Times bestsellers, The End of Faith and Letter to a Christian Nation. The End of Faith won the 2005 PEN Award for Nonfiction.

Mr. Harris' writing has been published in over fifteen languages. He and his work have been discussed in Newsweek, TIME, The New York Times, Scientific American, Nature, Rolling Stone, and many other journals. His writing has appeared in Newsweek, The New York Times, The Los Angeles Times, The Times (London), The Boston Globe, The Atlantic, The Annals of Neurology, and elsewhere.

Mr. Harris is a Co-Founder and CEO of Project Reason, a nonprofit foundation devoted to spreading scientific knowledge and secular values in society. He received a degree in philosophy from Stanford University and a Ph.D. in neuroscience from UCLA.

http://fora.tv/2010/11/10/Sam_Harris_Can_Science_Determine_Human_Values#fullprogram

http://www.samharris.org/site/media_video/

more » « less
Video Language:
English
Duration:
01:40:02

English subtitles

Revisions