text
stringlengths 277
5.29k
| textbook_name
stringclasses 16
values |
---|---|
The Blackwell Companion
to Philosophy
Blackwell Companions to Philosophy
This outstanding student reference series offers a comprehensive and authoritative
survey of philosophy as a whole. Written by today’s leading philosophers, each volume
provides lucid and engaging coverage of the key figures, terms, topics, and problems of
the field. Taken together, the volumes provide the ideal basis for course use, represent-
ing an unparalleled work of reference for students and specialists alike.
Already published in the series:
1 The Blackwell Companion to
Philosophy, Second Edition
Edited by Nicholas Bunnin and
E. P. Tsui-James
2 A Companion to Ethics
Edited by Peter Singer
3 A Companion to Aesthetics
Edited by David Cooper
4 A Companion to Epistemology
Edited by Jonathan Dancy and Ernest Sosa
5 A Companion to Contemporary Political
Philosophy
Edited by Robert E. Goodin and
Philip Pettit
6 A Companion to Philosophy of Mind
Edited by Samuel Guttenplan
7 A Companion to Metaphysics
Edited by Jaegwon Kim and Ernest Sosa
8 A Companion to Philosophy of Law and
Legal Theory
Edited by Dennis Patterson
9 A Companion to Philosophy of Religion
Edited by Philip L. Quinn and
Charles Taliaferro
10 A Companion to the Philosophy of
Language
Edited by Bob Hale and Crispin Wright
11 A Companion to World Philosophies
Edited by Eliot Deutsch and Ron
Bontekoe
12 A Companion to Continental Philosophy
Edited by Simon Critchley and
William Schroeder
14 A Companion to Cognitive Science
Edited by William Bechtel and
George Graham
15 A Companion to Bioethics
Edited by Helga Kuhse and Peter Singer
16 A Companion to the Philosophers
Edited by Robert L. Arrington
17 A Companion to Business Ethics
Edited by Robert E. Frederick
18 A Companion to the Philosophy of
Science
Edited by W. H. Newton-Smith
19 A Companion to Environmental
Philosophy
Edited by Dale Jamieson
20 A Companion to Analytic Philosophy
Edited by A. P. Martinich and David Sosa
21 A Companion to Genethics
Edited by Justine Burley and John Harris
22 A Companion to Philosophical Logic
Edited by Dale Jacquette
23 A Companion to Early Modern
Philosophy
Edited by Steven Nadler
Forthcoming
A Companion to African American
Philosophy
Edited by Tommy Lott and John Pittman
A Companion to African Philosophy
Edited by Kwasi Wiredu
A Companion to Ancient Philosophy
Edited by Mary Louise Gill
13 A Companion to Feminist Philosophy
Edited by Alison M. Jaggar and
Iris Marion Young
A Companion to Medieval Philosophy
Edited by Jorge J. E. Gracia, Greg Reichberg,
and Timothy Noone
The Blackwell
Companion to
Philosophy
SECOND EDITION
Edited by
NICHOLAS BUNNIN
and
E. P. TSUI-JAMES
Copyright © 1996, 2003 Blackwell Publishers Ltd, a Blackwell Publishing company
Editorial matter, selection and arrangement copyright © Nicholas Bunnin and Eric Tsui-James
1996, 2003
First edition published 1996
Reprinted 1996 (twice), 1998, 1999, 2002
Second edition published 2003
350 Main Street, Malden, MA 02148-5018, USA
108 Cowley Road, Oxford OX4 1JF, UK
550 Swanston Street, Carlton South, Victoria 3053, Australia
Kurfürstendamm 57, 10707 Berlin, Germany
The right of Nicholas Bunnin and Eric Tsui-James to be identified as the Authors of the Editorial
Material in this work has been asserted in accordance with the UK Copyright, Designs and Patents
Act 1988.
All rights reserved. No part of this publication may be reproduced, stored in a retrieval system,
or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording
or otherwise, except as permitted by the UK Copyright, Designs and Patents Act 1988, without
the prior permission of the publisher.
Library of Congress Cataloging-in-Publication Data
The Blackwell companion to philosophy / edited by Nicholas Bunnin and
E. P. Tsui-James. –– 2nd ed.
p.
cm. –– (Blackwell companions to philosophy)
Includes bibliographical references and index.
ISBN 0–631–21907–2 –– ISBN 0–631–21908–0 (pbk.)
I. Bunnin, Nicholas.
1. Philosophy.
II. Tsui-James, E. P.
III. Series.
B21 .B56 2003
100––dc21
2002023053
A catalogue record for this title is available from the British Library.
| Blackwell |
Set in 10 on 121/2 pt Photina
by SNP Best-set Typesetter Ltd, Hong Kong
Printed and bound in the United Kingdom by T. J. International, Padstow, Cornwall
For further information on
Blackwell Publishing, visit our website:
http://www.blackwellpublishing.com
For Antonia and Oliver Bunnin and Jamie Perry
Contents
Preface to the Second Edition
Preface to the First Edition
Notes on Contributors
Contemporary Philosophy in the United States – John R. Searle
Contemporary Philosophy: A Second Look – Bernard Williams
Part I Areas of Philosophy
1 Epistemology – A. C. Grayling
2 Metaphysics – Simon Blackburn, with a section on Time by Robin Le Poidevin
3 Philosophy of Language – Martin Davies
4 Philosophy of Logic – A. W. Moore
5 Philosophy of Mind – William G. Lycan
6 Ethics – John Skorupski
7 Aesthetics – Sebastian Gardner
8 Political and Social Philosophy – David Archard
9 Philosophy of Science – David Papineau
10 Philosophy of Biology – Elliott Sober
11 Philosophy of Mathematics – Mary Tiles
12 Philosophy of Social Science – Martin Hollis
13 Philosophy of Law – N. E. Simmonds
14 Philosophy of History – Leon Pompa
15 Philosophy of Religion – Charles Taliaferro
16 Applied Ethics – John Haldane
17 Bioethics, Genethics and Medical Ethics – Rebecca Bennett,
Charles A. Erin, John Harris and Søren Holm
18 Environmental Ethics – Holmes Rolston, III
19 Business Ethics – Georges Enderle
20 Philosophy and Feminism – Jean Grimshaw and Miranda Fricker
21 Ethnicity, Culture and Philosophy – Robert Bernasconi
ix
x
xii
1
23
35
37
61
90
147
173
202
231
257
286
317
345
375
403
428
453
490
499
517
531
552
567
CONTENTS
Part II History of Philosophy
22 Ancient Greek Philosophy – Robert Wardy
23 Plato and Aristotle – Lesley Brown
24 Medieval Philosophy – Jorge J. E. Gracia
25 Bacon – Stephen Gaukroger
26 Descartes and Malebranche – Richard Francks and George Macdonald Ross
27 Spinoza and Leibniz – Richard Francks and George Macdonald Ross
28 Hobbes – Tom Sorell
29 Locke – R. S. Woolhouse
30 Berkeley – Howard Robinson
31 Hume – Peter Jones
32 Kant – David Bell
33 Hegel – Michael Inwood
34 Marx – Richard Norman
35 Bentham, Mill and Sidgwick – Ross Harrison
36 Pragmatism – Susan Haack
37 Frege and Russell – R. M. Sainsbury
38 Moore – Thomas Baldwin
39 Wittgenstein – David Pears
40 Nietzsche – David E. Cooper
41 Husserl and Heidegger – Taylor Carman
42 Sartre, Foucault and Derrida – Gary Gutting
Glossary
Appendix
Index
583
585
601
619
634
644
658
671
682
694
709
725
741
750
759
774
790
805
811
827
842
860
875
893
905
Preface to the Second Edition
We thank readers for their gratifying response to the first edition of the Companion.
The second edition provides new chapters on Philosophy of Biology; Bioethics,
Genethics and Medical Ethics; Environmental Ethics; Business Ethics; Ethnicity, Culture
and Philosophy; Plato and Aristotle; Francis Bacon; Nietzsche; Husserl and Heidegger;
and Sartre, Foucault and Derrida. There are significant revisions or extensions to
chapters on Metaphysics, Philosophy of Language, Philosophy of Mind, Political and
Social Philosophy, Philosophy of Religion, Philosophy and Feminism, and Hobbes. The
discussion of Descartes, Spinoza and Leibniz is now divided between two chapters, and
in a new section Malebranche is considered along with Descartes in the first of these.
A longer chapter on Medieval Philosophy replaces the chapter by C. F. J. Martin, who
was unavailable to extend his work. We welcome our new contributors and hope that
readers will continue to be challenged and delighted by the Companion as a whole.
Nicholas Bunnin
E. P. Tsui-James
Preface to the First Edition
This Companion complements the Blackwell Companions to Philosophy series by
presenting a new overview of philosophy prepared by thirty-five leading British and
American philosophers. Introductory essays by John Searle and Bernard Williams,
which assess the changes that have shaped the subject in recent decades, are followed
by chapters exploring central problems and debates in the principal subdisciplines of
philosophy and in specialized fields, chapters concerning the work of great historical | Blackwell |
figures and chapters discussing newly developing fields within philosophy. Throughout
the course of
its chapters, the Companion examines the views of many of the
most widely influential figures of contemporary philosophy.
Although wide-ranging, the Companion is not exhaustive, and emphasis is placed
on developments in Anglo-American philosophy in the latter part of the twentieth
century. A premise underlying the Companion is that major participants in philosoph-
ical debate can provide accounts of their own fields that are stimulating, accessible,
stylish and authoritative.
In its primary use, the Companion is an innovative textbook for introductory courses
in philosophy. Teachers can use the broad coverage to select chapters in a flexible way
to support a variety of courses based on contemporary problems or the historical devel-
opment of the subject. Specialist chapters can be used selectively to augment standard
introductory topics or to prepare students individually for term papers or essays. Chap-
ters include initial summaries, boxed features, cross-references, suggestions for further
reading, references and discussion questions. In addition, terms are marked for a
common glossary. These features and the problem-setting nature of the discussions
encourage students to see the subject as a whole and to gain confidence that explo-
rations within philosophy can lead to unexpected and rewarding insights. In this
aspect, the Companion reflects the contributors’ experience of small group teaching,
in which arguments and perspectives are rigorously tested and in which no solution is
imposed.
In its secondary use, the Companion will accompany students throughout their
undergraduate careers and will also serve the general reader wishing to understand
the central concepts and debates within philosophy or its constituent disciplines.
Students are unlikely to read the whole volume in their first year of study, but those
continuing with philosophy will find their appreciation of the work deepening over time
PREFACE TO THE FIRST EDITION
as they gain insight into the topics of the more advanced chapters. The Companion
will help them to formulate questions and to see connections between what they
have already studied and new terrain.
In its final use, the Companion bears a special relationship to the Blackwell Com-
panions to Philosophy series. Many readers will wish to read the integrated discussions
of the chapters of the present Companion for orientation before turning to the detailed,
alphabetically arranged articles of the volumes in the Companion series. Although con-
ceived as a separate volume, the Companion to Philosophy will serve as a useful guide
to the other excellent Companions in what amounts to a comprehensive encyclopedia
of philosophy.
The general reader might begin with the introductory essays and turn to chapters
on Epistemology, Metaphysics, Ethics and Political and Social Philosophy, or to histori-
cal chapters from Ancient Greek Philosophy to Hume. Cross-references and special
interests will lead readers to other chapters.
Cross-references in the text are marked in small capitals followed by a chapter
number or page numbers in parentheses: Ethics (chapter 6) or Probability (pp.
308–11). We have used our judgement in marking terms appearing many times in the
text for cross-references, and hope that we have supplied guidance without distracting
readers. The Companion also provides a glossary of 210 terms and a comprehensive
index. Both appear at the end of the volume, and readers are advised to use them reg-
ularly for help in reading the chapters. When an author does not refer to a book by its
first edition, a recent publication is cited in the text, and the original date of publica-
tion (or in some cases of composition) will appear in square brackets in the references.
As editors, we are fully aware of our good fortune in attracting superb contributors. | Blackwell |
The complexity of their insights and the clarity of their presentations are the chief
attractions of the Companion. We appreciate their care in making the difficult not only
accessible but delightful as well. We also wish to thank the Departments of Philosophy
at the University of Essex and the University of Hong Kong for their support through-
out the preparation of this volume. We are especially grateful to Laurence Goldstein,
Tim Moore and Frank Cioffi for their comments and advice. A version of the Com-
panion is published in Chinese by the Shandong Academy of Social Sciences, and we
appreciate the friendly co-operation of our Chinese co-editors.
Our cover illustration, R. B. Kitaj’s philosophically resonant If Not, Not, is a work by
an American artist working in London during the period that provides the main focus
of our volume.
Nicholas Bunnin
E. P. Tsui-James
xi
Notes on Contributors
David Archard is Reader in Moral Philosophy and Director of the Centre for Ethics,
Philosophy and Public Affairs at the University of St Andrews. He is the author of
Sexual Consent (1998) and co-editor of The Moral and Political Status of Children: New
Essays (2002).
Thomas Baldwin is Professor of Philosophy at the University of York. He previously
taught at the University of Cambridge (where he was Fellow of Clare College) and at
Makerere University. He has published G. E. Moore (1990) and Contemporary Philoso-
phy: Philosophy in English since 1945 (2001) in addition to many articles on issues in
metaphysics and the philosophy of language.
David Bell is Professor of Philosophy at the University of Sheffield. He is the author of
works on Frege, Husserl and Kant. His interests include the foundations of arithmetic,
solipsism and the nature and origins of the analytic tradition.
Rebecca Bennett is Lecturer in Bioethics at the Centre for Social Ethics and Policy,
School of Law, University of Manchester. She edited (with Charles Erin) HIV and AIDS:
Testing, Screening and Confidentiality (1999).
Robert Bernasconi is Moss Professor of Philosophy at the University of Memphis.
He is the author of The Question of Language in Heidegger’s History of Being (1985) and
Heidegger in Question (1993) as well as numerous articles on Hegel and on twentieth-
century European philosophy. He has edited collections of essays on Derrida and on
Levinas and most recently Race (2001).
Simon Blackburn is Professor of Philosophy at the University of Cambridge. A former
editor of the journal Mind, he has written Ruling Passions (1998), Spreading the Word
(1984), Essays in Quasi Realism (1993) and The Oxford Dictionary of Philosophy (1994).
His current work concerns problems of realism and its alternatives as they have
emerged in historical and contemporary work.
Lesley Brown is Tutorial Fellow in Philosophy, Somerville College, University of
Oxford. She has written on Plato, especially his metaphysics and epistemology, and on
ancient philosophy of language.
CONTRIBUTORS
Nicholas Bunnin is Director of the Philosophy Project at the Institute for Chinese
Studies, University of Oxford and previously taught at the University of Glasgow and
the University of Essex. He compiled (with Jiyuan Yu) the Dictionary of Western Philos-
ophy: English–Chinese (2001) and edited (with Chung-ying Cheng) Contemporary
Chinese Philosophy (2002). His main interests are in metaphysics, the philosophy of
mind and political philosophy.
Taylor Carman is Assistant Professor of Philosophy at Barnard College, Columbia
University. He is co-editor of The Cambridge Companion to Merleau-Ponty (forthcoming)
and the author of Heidegger’s Analytic: Interpretation, Discourse, and Authenticity in
‘Being and Time’ (forthcoming), and of other articles on Husserl, Heidegger, and
Merleau-Ponty.
David E. Cooper is Professor of Philosophy at the University of Durham and Director
of the Durham Institute of Comparative Ethics. His books include Metaphor (1986), | Blackwell |
Existentialism: A Reconstruction (2nd revd edn 2000), World Philosophies: An Historical
Introduction (2nd revd edn 2002) and The Measure of Things: Humanism, Humility and
Mystery (2002).
Martin Davies is Professor of Philosophy in the Research School of Social Sciences,
Australian National University. He was formerly Wilde Reader in Mental Philosophy at
the University of Oxford. He has published widely in the areas of philosophy of
language, mind and psychology.
Georges Enderle is Arthur and Mary O’Neil Professor of International Business Ethics
at the University of Notre Dame. His books include Business Students Focus on Ethics
(1993), translated into Portuguese (1997) and Chinese (2001).
Charles A. Erin is Senior Lecturer in Applied Philosophy and Fellow of the Institute
of Medicine, Law and Bioethics at the University of Manchester. He has written widely
on topics in bioethics and edited (with Rebecca Bennett) HIV and AIDS: Testing,
Screening and Confidentiality (1999).
Richard Francks is Director of Undergraduate Studies in Philosophy at the
University of Leeds. His main interests are in epistemology, the history of philosophy
and the philosophy of history.
Miranda Fricker is Lecturer in Philosophy at Birkbeck College, University of London
and was previously Lecturer in Philosophy and British Academy Postdoctoral Fellow at
Heythrop College, University of London. She has published articles in epistemology,
ethics and social philosophy, and edited (with Jennifer Hornsby) The Cambridge
Companion to Feminism in Philosophy (2000). Her current work focuses on the idea of
an ethics of epistemic practice.
Sebastian Gardner is Reader in Philosophy at University College, London. He is the
author of Irrationality and the Philosophy of Psychoanalysis (1993) and Kant and the
Critique of Pure Reason (1999). His interests lie in aesthetics, psychoanalysis and the
history of philosophy.
xiii
CONTRIBUTORS
Stephen Gaukroger is Professor of History of Philosophy and History of Science at
the University of Sydney. He is the author of Explanatory Structures (1978), Cartesian
Logic (1989), Descartes: An Intellectual Biography (1995), Francis Bacon and the
Transformation of Early Modern Philosophy (2000) and Descartes’ System of Natural
Philosophy (2001). He has also edited four collections of essays and published
translations of Descartes and Arnaud.
Jorge J. E. Gracia is a State University of New York Distinguished Professor and holds
the Samuel F. Capon Chair in the Department of Philosophy, State University of New
York, University at Buffalo. He has written widely on medieval philosophy, metaphysics,
philosophical historiography, philosophy of language and philosophy in Latin America.
His books include Introduction to the Problem of Individuation in the Early Middle Ages
(2nd revd edn 1988), Individuality: An Essay on the Foundations of Metaphysics (1988),
Philosophy and Its History: Issues in Philosophical Historiography (1992), A Theory of
Textuality: The Logic and Epistemology (1995), Texts: Ontological Status, Identity, Author,
Audience (1996) and Metaphysics and Its Task: The Search for the Categorical Foundation
of Knowledge (1999).
A. C. Grayling is Reader in Philosophy at Birkbeck College, London, and Supernu-
merary Fellow at St Anne’s College, Oxford. Among his books are An Introduction
to Philosophical Logic (3rd edn 1992), The Refutation of Scepticism (1985), Berkeley:
The Central Arguments (1986), Wittgenstein (1988), Russell (1993), Moral Values (1998),
The Quarrel of the Age (2000) and The Meaning of Things (2001). He has edited
Philosophy: A Guide Through the Subject (1995) and Philosophy: Further Through the
Subject (1998).
Jean Grimshaw taught Philosophy and Women’s Studies at the University of the West
of England, Bristol. She is the author of Feminist Philosophers: Women’s Perspectives on
Philosophical Traditions (1986) and a number of articles, mainly on feminism and phi- | Blackwell |
losophy. She has edited (with Jane Arthurs) Women’s Bodies: Discipline and Transgression
(1999).
Gary Gutting is Professor of Philosophy at the University of Notre Dame. He is the
author of Religious Belief and Religious Skepticism (1982), Michel Foucault’s Archaeology
of Knowledge (1989) and French Philosophy in the Twentieth Century (2001).
Susan Haack formerly Professor of Philosophy at the University of Warwick,
currently Professor of Philosophy at the University of Miami, is the author of Deviant
Logic (1974), Philosophy of Logic (1978), Evidence and Inquiry: Towards Reconstruction
in Epistemology (1993) and Manifesto of a Passionate Moderate: Unfashionable Essays
(1998). Her main areas of interest are the philosophy of logic and language, episte-
mology and metaphysics and pragmatism. She is a past President of the Charles
Peirce Society.
John Haldane is Professor of Philosophy and formerly Director of the Centre for
Philosophy and Public Affairs at the University of St Andrews. He has published
widely in the philosophy of mind, the philosophy of value and the history of philoso-
phy. He is co-author with J. J. C. Smart of Atheism and Theism (1996) in the Blackwell
Great Debates in Philosophy series.
xiv
CONTRIBUTORS
John Harris is Sir David Alliance Professor of Bioethics, Institute of Medicine, Law and
Bioethics, University of Manchester. He is a member of the United Kingdom Human
Genetics Commission and of the Ethics Committee of the British Medical Association.
He was a Founder Director of the International Association of Bioethics and a founder
member of the Board of the journal Bioethics. Among his books are The Value of Life
(1985) and Clones, Genes and Immortality (1998) (a revised edition of Wonderwoman and
Superman, 1992), and he is editor of Bioethics (2001) in the Oxford Readings in Phi-
losophy series.
Ross Harrison teaches philosophy at the University of Cambridge, where he is also a
Fellow of King’s College. Among his publications are Bentham (1983), Democracy
(1993) and (as editor and contributor) Henry Sidgwick (2001).
Martin Hollis was Professor of Philosophy at the University of East Anglia, Norwich.
He specialized in the philosophy of social science, especially in topics to do with ratio-
nality. Among his books are Models of Man (1977), The Cunning of Reason (1987), The
Philosophy of Social Science (1994), Reason in Action (1995), Trust Within Reason (1998)
and Pluralism and Liberal Neutrality (1999). The last two volumes were published after
his untimely death in 1998.
Søren Holm is Professor of Clinical Bioethics at the University of Manchester. He is
the author of Ethical Problems of Clinical Practice: The Ethical Reasoning of Health Care
Professionals (1997) and has edited (with Inez de Beaufort and Medard Hilhorst) In the
Eye of the Beholder: Ethics and Medical Change of Appearance (1996) and (with John
Harris) The Future Of Human Reproduction: Ethics, Choice and Regulation (1998).
Michael Inwood is Tutorial Fellow in Philosophy at Trinity College, Oxford. He has
published several books on Hegel. His other interests include ancient philosophy and
Heidegger. He is especially interested in the interconnections between Greek and
German philosophy.
Peter Jones was Professor of Philosophy and Director of the Institute for Advanced
Studies in the Humanities at the University of Edinburgh. He is the author of numer-
ous works, including Hume’s Sentiments (1982).
Robin Le Poidevin is Professor of Metaphysics at the University of Leeds, where he
was Head of the School of Philosophy 1988–2001. He is the author of Change, Cause
and Contradiction: A Defence of the Tenseless Theory of Time (1991) and Arguing for
Atheism: An Introduction to the Philosophy of Religion (1996) and has edited Questions of
Time and Tense (1998) and (with Murray MacBeath) The Philosophy of Time (1993).
William G. Lycan is William Rand Kenan, Jr Professor of Philosophy at the University | Blackwell |
of North Carolina. He has published a number of books, including Consciousness
(1987), Judgement and Justification (1988) and Consciousness and Experience (1996). He
is the editor of Mind and Cognition (1990). His interests are in the philosophy of mind,
the philosophy of language and epistemology.
A. W. Moore is Tutorial Fellow in Philosophy at St Hugh’s College, Oxford. He is the
author of The Infinite (2nd edn 2001) and Points of View (1997). He has also edited two
collections of essays: Meaning and Reference (1993) and Infinity (1993).
xv
CONTRIBUTORS
Richard Norman is Professor of Moral Philosophy at the University of Kent. His pub-
lications include The Moral Philosophers (1983), Free and Equal (1987) and Ethics, Killing
and War (1995).
David Papineau is Professor of Philosophy of Science at King’s College, London. He
has published widely in epistemology, the philosophy of mind and the philosophy of
science. His books include Reality and Representation (1987), Philosophical Naturalism
(1993), Introducing Consciousness (2000) and Thinking About Consciousness (2002).
David Pears is Emeritus Professor of Philosophy at the University of Oxford. His most
recent publications are The False Prison: A Study in the Development of Wittgenstein’s Phi-
losophy (2 vols, 1987 and 1988) and Hume’s System: An Examination of Book I of the
Treatise (1991). His other interests include entomology and the visual arts.
Leon Pompa was Professor of Philosophy at the University of Birmingham. His
research interests include the history of philosophy and the philosophy of history. He
has published a number of articles on the problems of fact, value and narrative in
history and on Descartes, Vico, Kant, Hegel, Marx, Collingwood and Wittgenstein. He
co-edited with W. H. Dray Substance and Form in History: Essays in Philosophy of History
(1981), was editor and translator of Vico: A Study of the ‘New Science’ (2nd edn 1990)
and is the author of Human Nature and Historical Knowledge: Hume, Hegel and Vico
(1990).
Howard Robinson is Professor of Philosophy, Central European University, Budapest.
He was previously Soros Professor of Philosophy at the Eötvös Loránd University,
Budapest and Reader in Philosophy at the University of Liverpool. His main interests
are in the philosophy of mind and in idealism. He is the author of Matter and Sense
(1982) and Perception (1994), and co-author (with John Foster) of Essays on Berkeley
(1985). He edited Objections to Physicalism (1991) and is currently editing Berkeley’s
Principles and Three Dialogues for Oxford University Press’s World Classics series.
Holmes Rolston, III is University Distinguished Professor and Professor of Philosophy
at Colorado State University. He has written seven books, most recently Genes, Genesis
and God (1999), Philosophy Gone Wild (1986), Environmental Ethics (1988), Science and
Religion: A Critical Survey (1987) and Conserving Natural Value (1994). He gave the
Gifford Lectures, University of Edinburgh, 1997–8, has lectured on seven continents,
is featured in Joy A. Palmer’s (ed.) Fifty Key Thinkers on the Environment and is past and
founding president of the International Society for Environmental Ethics.
George MacDonald Ross is Senior Lecturer in the Department of Philosophy at the
University of Leeds and Director of the Philosophical and Religious Studies Subject
Centre of the Learning and Teaching Support Network. He has written extensively on
Leibniz and other seventeenth- and eighteenth-century philosophers, and is the author
of Leibniz (1984).
R. M. Sainsbury is Stebbing Professor of Philosophy at King’s College, London
and was editor of the journal Mind for several years until 2000. He has published
Russell (1979), Paradoxes (1995) and Logical Forms (2000). His main interests are in
philosophical logic and the philosophy of language.
xvi
CONTRIBUTORS
John R. Searle is Mills Professor of Mind and Language at the University of California | Blackwell |
where he has been a faculty member since 1959. Before that, he was a lecturer at Christ
Church, Oxford, and he received all his university degrees from Oxford. Most of his work
is in the philosophy of mind, the philosophy of language, and social philosophy. His
most recently published books are Rationality in Action (2001) and Mind, Language and
Society (1998). He is the author of several other important books, including Speech Acts:
An Essay in the Philosophy of Language (1969), Expression and Meaning: Studies in the
Theory of Speech Acts (1985), Intentionality (1983), Minds, Brains and Science, the 1984
Reith Lectures (1989), The Rediscovery of Mind (1992) and The Construction of Social
Reality (1995).
N. E. Simmonds is a Fellow of Corpus Christi College, Cambridge where he lectures in
law. His interests include the philosophy of law and political philosophy. He has pub-
lished The Decline of Judicial Reason (1984), Central Issues in Jurisprudence (1986) and
numerous articles on the philosophy of law.
John Skorupski is Professor of Moral Philosophy at the University of St Andrews. He
is the author of John Stuart Mill (1989) and English-Language Philosophy 1750–1945
(1993). His most recent book is Ethical Explorations (1999).
Elliott Sober is Hans Reichenbach Professor and Henry Vilas Research Professor at
the University of Wisconsin, Madison and Centennial Professor at the London School
of Economics. He is the author of The Nature of Selection (1984), Reconstructing the Past
(1988), Philosophy of Biology (1993) and (with David S. Wilson) Unto Others: Evolution
and Psychology of Unselfish Behaviour (1988).
Tom Sorell is Professor of Philosophy at the University of Essex. He is the author of
Hobbes (1986), Descartes (1987), Moral Theory and Capital Punishment (1987), Scientism
(1991), (with John Hendry) Business Ethics (1994) and Moral Theory and Anomaly
(2000). He is the editor of The Rise of Modern Philosophy (1993), The Cambridge Com-
panion to Hobbes (1995), Health Care, Ethics, and Insurance (1998), Descartes (1999), and
(with John Rogers) Hobbes and History (2000).
Charles Taliaferro is Professor of Philosophy at St Olaf College, Northfield, Minnesota
and the author of Consciousness and the Mind of God (1994), Contemporary Philosophy
of Religion (1999) and the co-editor of A Companion to Philosophy of Religion (1998).
Mary Tiles is Professor of Philosophy at the University of Hawaii at Manoa. Her inter-
ests include the history and philosophy of mathematics, science and technology and
their interactions with culture (European and Chinese). She has published Living in a
Technological Culture (with Hans Oberdiek) (1995), An Introduction to Historical Episte-
mology (with James Tiles) (1993), Mathematics and the Image of Reason (1991) and
Bachelard: Science and Objectivity (1984).
Eric P. Tsui-James studied as a postgraduate at Oriel College, Oxford. He taught phi-
losophy at St Hilda’s College, Oxford, for two years before moving to the University of
Hong Kong in 1990. He has published work on the metaphysics of mathematics, but
his research interests now centre around the work of William James, especially the
nineteenth-century psychological and physiological contexts of his radical empiricism.
xvii
CONTRIBUTORS
Robert Wardy teaches philosophy and classics at St Catharine’s College, Cambridge.
He has published in the fields of ancient Greek philosophy and rhetoric, Latin litera-
ture, the philosophy of language and Chinese philosophy.
Bernard Williams is Monroe Deutsch Professor of Philosophy, University of Califor-
nia, Berkeley, and was White’s Professor of Moral Philosophy and a Fellow of Corpus
Christi College, Oxford. His works include Morality (1972), Problems of the Self (1973),
Descartes: The Project of Pure Enquiry (1978), Moral Luck (1981), Ethics and the Limits
of Philosophy (1985), Shame and Necessity (1993) and Making Sense of Humanity
(1995).
R. S. Woolhouse is Professor of Philosophy at the University of York. He is the author | Blackwell |
of Locke’s Philosophy of Science and Knowledge (1971), Locke (1983), The Empiricists
(1988) and Descartes, Spinoza, Leibniz: The Concept of Substance in Seventeenth-Century
Philosophy (1993).
xviii
Contemporary Philosophy in
the United States
J O H N R. S E A R L E
Philosophy as an academic discipline in America has considerably fewer practitioners
than do several other subjects in the humanities and the social sciences, such as
sociology, history, English, or economics; but it still shows enormous diversity. This
variety is made manifest in the original research published by professional philosophers,
whose differing points of view are expressed in the large number of books published
each year, as well as in the many professional philosophy journals. There are
over two thousand colleges and universities in the United States, of which nearly
all have philosophy departments, and the number of professional philosophers is
correspondingly large.
Because of this diversity, any generalizations about the discipline as a whole, which
I am about to make, are bound to be misleading. The subject is too vast and complex to
be describable in a single essay. Furthermore, anyone who is an active participant in
the current controversies, as I am, necessarily has a perspective conditioned by his or
her own interests, commitments and convictions. It would be impossible for me to give
an ‘objective’ account. I am not therefore in what follows trying to give a neutral or
disinterested account of the contemporary philosophical scene; rather I am trying to
say what in the current developments seems to me important.
In spite of its enormous variety, there are certain central themes in contemporary
American philosophy. The dominant mode of philosophizing in the United States
is called ‘analytic philosophy’. Without exception, the best philosophy departments
in the United States are dominated by analytic philosophy, and among the leading
philosophers in the United States, all but a tiny handful would be classified as analytic
philosophers. Practitioners of types of philosophizing that are not in the analytic
tradition – such as phenomenology, classical pragmatism, existentialism, or Marxism –
feel it necessary to define their position in relation to analytic philosophy. Indeed,
analytic philosophy is the dominant mode of philosophizing not only in the United
States, but throughout the entire English-speaking world, including Great Britain,
Canada, Australia and New Zealand. It is also the dominant mode of philosophizing in
Scandinavia, and it is also becoming more widespread in Germany, France, Italy and
throughout Latin America. I personally have found that I can go to all of these parts of
the world and lecture on subjects in contemporary analytic philosophy before audiences
who are both knowledgeable and well trained in the techniques of the discipline.
JOHN R. SEARLE
1 Analytic Philosophy
What, then, is analytic philosophy? The simplest way to describe it is to say that it
is primarily concerned with the analysis of meaning. In order to explain this enterprise
and its significance, we need first to say a little bit about its history. Though the
United States now leads the world in analytic philosophy, the origins of this mode of
philosophizing lie in Europe. Specifically, analytic philosophy is based on the work
of Gottlob Frege, Ludwig Wittgenstein, Bertrand Russell and G. E. Moore, as well as
the work done by the logical positivists of the Vienna Circle in the 1920s and 1930s.
Going further back in history, one can also see analytic philosophy as a natural descend-
ant of the empiricism of the great British philosophers Locke, Berkeley and Hume, and
of the transcendental philosophy of Kant. In the works of philosophers as far back as
Plato and Aristotle, one can see many of the themes and presuppositions of the | Blackwell |
methods of analytic philosophy. We can best summarize the origins of modern analytic
philosophy by saying that it arose when the empiricist tradition in epistemology,
together with the foundationalist enterprise of Kant, were tied to the methods of
logical analysis and the philosophical theories invented by Gottlob Frege in the late
nineteenth century. In the course of his work on the foundations of mathematics,
Frege invented symbolic logic in its modern form and developed a comprehensive
and profound philosophy of language. Though many of the details of his views on
language and mathematics have been superseded, Frege’s work is crucial for at least
two reasons. Firstly, by inventing modern logic, specifically the predicate calculus,
he gave us a primary tool of philosophical analysis; and, secondly, he made the
philosophy of
language central to the entire philosophical enterprise. From the
point of view of analytic philosophy, Frege’s work is the greatest single philo-
logical
sophical achievement of
analysis were later augmented by the ordinary language analysis inspired by the
work of Moore and Wittgenstein and are best exemplified by the school of
lin-
guistic philosophy that flourished in Oxford in the l950s. In short, analytic phi-
losophy attempts to combine certain traditional philosophical themes with modern
techniques.
the nineteenth century. Fregean techniques of
Analytic philosophy has never been fixed or stable, because it is intrinsically self-
critical and its practitioners are always challenging their own presuppositions and con-
clusions. However, it is possible to locate a central period in analytic philosophy – the
period comprising, roughly speaking, the logical positivist phase immediately prior to
the 1939–45 war and the postwar phase of linguistic analysis. Both the prehistory and
the subsequent history of analytic philosophy can be defined by the main doctrines of
that central period.
In the central period, analytic philosophy was defined by a belief in two linguistic
distinctions, combined with a research programme. The two distinctions are, firstly, that
between analytic and synthetic propositions, and, secondly, that between descriptive
and evaluative utterances. The research programme is the traditional philosophical
research programme of attempting to find foundations for such philosophically prob-
lematic phenomena as language, knowledge, meaning, truth, mathematics and so on.
2
CONTEMPORARY PHILOSOPHY IN THE UNITED STATES
One way to see the development of analytic philosophy over the past thirty years is to
regard it as the gradual rejection of these two distinctions, and a corresponding rejec-
tion of foundationalism as the crucial enterprise of philosophy. However, in the central
period, these two distinctions served not only to identify the main beliefs of analytic phi-
losophy, but, for those who accepted them and the research programme, they defined
the nature of philosophy itself.
1.1 Analytic versus synthetic
The distinction between analytic and synthetic propositions was supposed to be
the distinction between those propositions that are true or false as a matter of
definition or of the meanings of the terms contained in them (the analytic propo-
sitions) and those that are true or false as a matter of
fact in the world and not
solely in virtue of the meanings of the words (the synthetic propositions). Examples
of analytic truths would be such propositions as ‘Triangles are three-sided plane
figures’, ‘All bachelors are unmarried’, ‘Women are female’, ‘2 + 2 = 4’ and so on.
In each of these, the truth of the proposition is entirely determined by its meaning;
they are true by the definitions of the words that they contain. Such propositions
can be known to be true or false a priori, and in each case they express necessary | Blackwell |
truths. Indeed, it was a characteristic feature of the analytic philosophy of this central
period that terms such as ‘analytic’, ‘necessary’, ‘a priori’ and ‘tautological’ were
taken to be co-extensive. Contrasted with these were synthetic propositions, which, if
they were true, were true as a matter of empirical fact and not as a matter of definition
alone. Thus, propositions such as ‘There are more women than men in the United
States’, ‘Bachelors tend to die earlier than married men’ and ‘Bodies attract each other
according to the inverse square law’ are all said to be synthetic propositions, and, if
they are true, they express a posteriori empirical truths about the real world that are
independent of
language. Such empirical truths, according to this view, are never
necessary; rather, they are contingent. For philosophers holding these views, the terms
‘a posteriori’, ‘synthetic’, ‘contingent’ and ‘empirical’ were taken to be more or less
co-extensive.
It was a basic assumption behind the logical positivist movement that all meaning-
ful propositions were either analytic or empirical, as defined by the conceptions that I
have just stated. The positivists wished to build a sharp boundary between meaningful
propositions of science and everyday life on the one hand, and nonsensical propositions
of metaphysics and theology on the other. They claimed that all meaningful proposi-
tions are either analytic or synthetic: disciplines such as logic and mathematics fall
within the analytic camp; the empirical sciences and much of common sense fall within
the synthetic camp. Propositions that were neither analytic nor empirical propositions,
and which were therefore in principle not verifiable, were said to be nonsensical or
meaningless. The slogan of the positivists was called the verification principle, and, in
a simple form, it can be stated as follows: all meaningful propositions are either
analytic or synthetic, and those which are synthetic are empirically verifiable. This
slogan was sometimes shortened to an even simpler battle cry: the meaning of a
proposition is just its method of verification.
3
JOHN R. SEARLE
1.2 The distinction between evaluative utterances and
descriptive utterances
Another distinction, equally important in the positivist scheme of things, is the dis-
tinction between those utterances that express propositions that can be literally either
true or false and those utterances that are used not to express truths or falsehoods, but
rather, to give vent to our feelings and emotions. An example of a descriptive statement
would be, ‘The incidence of crimes of theft has increased in the past ten years’. An
instance of the evaluative class would be ‘Theft is wrong’. The positivists claimed that
many utterances that had the form of meaningful propositions were used not to state
propositions that were verifiable either analytically or synthetically, but to express emo-
tions and feelings. Propositions of ethics look as if they are cognitively meaningful,
but they are not; they have only ‘emotive’ or ‘evaluative’ meaning. The propositions of
science, mathematics, logic and much of common sense fall in the descriptive class; the
utterances of aesthetics, ethics and much of religion fall in the evaluative class. It is
important to note that on this conception evaluative propositions are not, strictly speak-
ing, either true or false, since they are not verifiable as either analytic or empirical. The
two distinctions are crucially related in that all of the statements that fall on one side
or the other of the analytic–synthetic distinction also fall within the descriptive class of
the descriptive–evaluative distinction.
The importance that these two distinctions had for defining both the character of
the philosophical enterprise and the relationships between language and reality is hard | Blackwell |
to exaggerate. One radical consequence of the distinction between descriptive and
evaluative propositions was that certain traditional areas of philosophy, such as ethics,
aesthetics and political philosophy, were virtually abolished as realms of cognitive
meaningfulness. Propositions in these areas were, for the most part, regarded as non-
sensical expressions of feelings and emotions, because they are not utterances that can
be, strictly speaking, either true or false. Since the aim of philosophers is to state the
truth, and since evaluative utterances cannot be either true or false, it cannot be one
of the aims of philosophy to make any evaluative utterances. Philosophers might
analyse the meaning of evaluative terms, and they might examine the logical rela-
tionships among these terms, but philosophers, qua philosophers, can make no first-
order evaluations in aesthetics, ethics or politics, as these first-order evaluations are
not, strictly speaking, meaningful. They may have a sort of secondary, derivative
meaning, called ‘emotive meaning’, but they lack scientifically acceptable cognitive
meaning.
If the task of philosophy is to state the truth and not to provide evaluations, what
then is the subject matter of philosophy? Since the methods of philosophers are not
those of empirical science – since their methods are a priori rather than a posteriori – it
cannot be their aim to state empirical truths about the world. Such propositions are the
propositions of the special sciences. The aim of philosophers, therefore, is to state
analytic truths concerning logical relations among the concepts of our language. In
this period of philosophy, the task of philosophy was taken to be the task of conceptual
analysis. Indeed, for most philosophers who accepted this view, philosophy and con-
ceptual analysis were the same. Where traditional philosophers had taken their task
to be the discussion of the nature of the good, the true, the beautiful and the just, the
4
CONTEMPORARY PHILOSOPHY IN THE UNITED STATES
positivist and post-positivist analytic philosophers took their task to be the analysis of
the meaning of concepts such as ‘goodness’, ‘truth’, ‘beauty’ and ‘justice’. Ideally the
analysis of these and other philosophically interesting concepts, such as ‘knowledge’,
‘certainty’ and ‘cause’, should give necessary and sufficient conditions for the applica-
tion of these concepts. They saw this as being the legitimate heir of the traditional phi-
losophical enterprise, but an heir purged of the metaphysical nonsense and confusion
that had discredited the traditional enterprise.
If we combine the assumption that philosophy is essentially a conceptual, analytic
enterprise with the assumption that its task is foundational – that is, its task is to provide
secure foundations for such things as knowledge – then the consequence for the posi-
tivists is that philosophical analysis tends in large part to be reductive. That is, the aim
of the analysis is to show, for example, how empirical knowledge is based on, and ulti-
mately reducible to, the data of our experience, to so-called sense data. (This view is
called ‘phenomenalism’.) Similarly, statements about the mind are based on, and there-
fore ultimately reducible to, statements about external behaviour (behaviourism). Nec-
essary truth is similarly based on conventions of language as expressed in definitions
(conventionalism); and mathematics is based on logic, especially set theory (logicism).
In each case, the more philosophically puzzling phenomenon is shown to have a secure
foundation in some less puzzling phenomenon, and indeed, the ideal of such analysis
was to show that the puzzling phenomena could be entirely reduced to less puzzling
phenomena. ‘Phenomenalism’ supposedly gave science a secure foundation because
science could be shown to be founded on the data of our senses. Since the form of the | Blackwell |
reduction was analytic or definitional, it had the consequence that statements about
empirical reality could be translated into statements about sense data. Similarly, accord-
ing to behaviourism, statements about mental phenomena could be translated into
statements about behaviour.
Within the camp of analytic philosophers who thought the aim of philosophy was
conceptual analysis, there were two broad streams. One stream thought ordinary
language was in general quite adequate, both as a tool and as a subject matter of
philosophical analysis. The other stream thought of ordinary language as hopelessly
inadequate for philosophical purposes, and irretrievably confused. The philosophers of
this latter stream thought that we should use the tools of modern mathematical logic
both for analysing traditional philosophical problems and, more importantly, for creat-
ing a logically perfect language, for scientific and philosophical purposes, in which
certain traditional confusions could not even arise. There was never a rigid distinction
between these two streams, but there were certainly two broad trends: one which
emphasized ordinary language philosophy and one which emphasized symbolic logic.
Both streams, however, accepted the central view that the aim of philosophy was con-
ceptual analysis, and that in consequence philosophy was fundamentally different from
any other discipline; they thought that it was a second-order discipline analysing the
logical structure of language in general, but not dealing with first-order truths about
the world. Philosophy was universal in subject matter precisely because it had no
special subject matter other than the discourse of all other disciplines and the discourse
of common sense.
A further consequence of this conception was that philosophy became essentially a
linguistic or conceptual enterprise. For that reason, the philosophy of language was
5
JOHN R. SEARLE
absolutely central to the philosophical task. In a sense, the philosophy of language was
not only ‘first philosophy’; all of philosophy became a form of philosophy of language.
Philosophy was simply the logical investigation of the structure of language as it was
used in the various sciences and in common life.
2 The Rejection of These Two Distinctions and
the Rejection of Foundationalism
Work done in the 1950s and 1960s led to the overcoming of these two distinctions;
and with the rejection of these two distinctions came a new conception of analytic phi-
losophy – a conception that emerged in the 1970s and 1980s and which is still being
developed. The rejection of these two distinctions and of the foundationalist research
programme led to an enormous upheaval in the conception of the philosophical enter-
prise and in the practice of analytic philosophers. The most obvious problem with tra-
ditional analytic philosophy was that the reductionist enterprise failed. In every case,
the attempts to provide reductionist analyses of the sort proposed by the phenomenal-
ists and behaviourists were unsuccessful, and by 1960 the lack of success was obvious.
A series of important theoretical developments also took place at this time, but for the
sake of simplicity I shall concentrate on only five of these: Quine’s rejection of the
analytic–synthetic distinction, Austin’s theory of speech acts, Wittgenstein’s criticism
of
foundationalism, Rawls’s work in political philosophy and the changes in the
philosophy of science due to Kuhn and others.
2.1 Quine’s attack on the analytic–synthetic distinction
Perhaps the most important criticism of the analytic–synthetic distinction was made
by W. V. O. Quine in a famous article entitled ‘Two dogmas of empiricism’ (Quine 1953).
In this article, Quine claimed that no adequate, non-circular definition of analyticity
had ever been given. Any attempt to define analyticity had always been made using
notions that were in the same family as analyticity, such as synonymy and definition, | Blackwell |
and consequently, the attempts to define analyticity were invariably circular. However,
an even more important objection that emerged in Quine’s article was this: the notion
of an analytic proposition is supposed to be a notion of a proposition that is immune
to revision, that is irrefutable. Quine claimed that there were no propositions that were
immune to revision, that any proposition could be revised in the face of recalcitrant
evidence, and that any proposition could be held in the face of recalcitrant evidence,
provided that one was willing to make adjustments in other propositions originally held
to be true. Quine argued that we should think of the language of science as being like
a complex network that was impinged upon by empirical verification only at the edges.
Recalcitrant experiences at the edges of science can produce changes anywhere along
the line, but the changes are not forced on us by purely logical considerations; rather,
we make various pragmatic or practical adjustments in the network of our sentences
or beliefs to accommodate the ongoing character of our experiences. Language, on this
view, is not atomistic. It does not consist of a set of propositions, each of which can be
assessed in isolation. Rather, it consists of a holistic network, and, in this network,
6
CONTEMPORARY PHILOSOPHY IN THE UNITED STATES
propositions as groups confront experience; propositions individually are not simply
assessed as true or false. (This holism of scientific discourse was influenced by the
French philosopher of science, Duhem, and the view is frequently referred to as ‘the
Duhem–Quine thesis’.)
Most philosophers today accept some version or other of Quine’s rejection of the
analytic–synthetic distinction. Not everybody agrees with his actual argument (I, for
one, do not), but now there is general scepticism about our ability to make a strict dis-
tinction between those propositions that are true by definition and those that are true
as a matter of fact. The rejection of the analytic–synthetic distinction has profound
consequences for analytic philosophy, as we shall see in more detail later.
At this point it is important to state that if there is no well-defined class of analytic
propositions, then the philosopher’s propositions cannot themselves be clearly identi-
fied as analytic. The results of philosophical analysis cannot be sharply distinguished
from the results of scientific investigation. On the positivist picture, philosophy was not
one among other sciences; rather, it stood outside the frame of scientific discourse and
analysed the logical relations between, on the one hand, that discourse and its vocabu-
lary and, on the other, experience and reality. Philosophers, so to speak, analysed the
relation between language and reality, but only from the side. If we accept Quine’s rejec-
tion of the analytic–synthetic distinction, then philosophy is not something that can
be clearly demarcated from the special sciences. It is, rather, adjacent to, and overlaps
with, other disciplines. Although philosophy is more general than other disciplines, its
propositions do not have any special logical status or special logical priority with regard
to the other disciplines.
2.2 Austin’s theory of speech acts
The British philosopher J. L. Austin was suspicious of both the distinction between ana-
lytic and synthetic propositions, and the distinction between evaluative and descriptive
utterances. During the 1950s he developed an alternative conception of
language
(Austin 1962). His first observation was that there is a class of utterances that are obvi-
ously perfectly meaningful, but which do not even set out to be either true or false. A
man who says, for example, ‘I promise to come and see you’ or a qualified authority
who says to a couple, ‘I pronounce you man and wife’ is neither reporting on nor
describing a promise or a marriage respectively. Such utterances should be thought | Blackwell |
not as cases of describing or stating, but rather as doing, as acting. Austin baptized
these utterances ‘performatives’ and contrasted them with ‘constatives’. The distinction
between constatives and performatives was supposed to contain three features: con-
statives, but not performatives, could be true or false; performatives, on the other hand,
though they could not be true or false, could be felicitous or infelicitous, depending on
whether or not they were correctly, completely and successfully performed; and finally,
performatives were supposed to be actions, doings or performances, as opposed to mere
sayings or statings. But, as Austin himself saw, the distinctions so drawn did not work.
Many so-called performatives turned out to be capable of being true or false; for
example, warnings could be either true or false. And statements, as well as performa-
tives, could be infelicitous. For example, if one made a statement for which one had
insufficient evidence, one would have made an infelicitous statement. And finally,
7
JOHN R. SEARLE
stating is as much performing an action as promising or ordering or apologizing. The
abandonment of the performative–constative distinction led Austin to a general theory
of speech acts. Communicative utterances in general are actions of a type he called
‘illocutionary acts’.
One great merit of Austin’s theory of speech acts is that it enabled subsequent
philosophers to construe the philosophy of language as a branch of the philosophy of
action. Since speech acts are as much actions as any other actions, the philosophical
analysis of language is part of the general analysis of human behaviour. And since
intentional human behaviour is an expression of mental phenomena, it turns out that
the philosophy of language and the philosophy of action are really just different aspects
of one larger area, namely, the philosophy of mind. On this view, the philosophy of lan-
guage is not ‘first philosophy’; it is a branch of the philosophy of mind. Though Austin
did not live to carry out the research programme implicit in his initial discoveries,
subsequent work, including my own, has carried this research further.
By treating speaking as a species of intentional action we can give a new sense to a
lot of old questions. For example, the old question, ‘How many kinds of utterances are
there?’ is too vague to be answered. But if we ask ‘How many kinds of illocutionary
acts are there?’, we can give a precise answer, since the question asks, ‘How many pos-
sible ways are there for speakers to relate propositional contents to reality in the per-
formance of actions that express illocutionary intentions?’ An analysis of the structure
of those intentions reveals five basic types of illocutionary act: we tell people how things
are (Assertives), we try to get them to do things (Directives), we commit ourselves to
doing things (Commissives), we express our feelings and attitudes (Expressives) and we
bring about changes in the world through our utterances, so that the world is changed
to match the propositional content of the utterance (Declarations). (For details see
Searle 1979 and 1983.)
2.3 Wittgenstein’s rejection of foundationalism
The single most influential analytic philosopher of the twentieth century, and indeed,
the philosopher whom most analytic philosophers would regard as the greatest philoso-
pher of the century, is Ludwig Wittgenstein.
Wittgenstein published only one short book during his lifetime, which represents his
early work, but with the posthumous publication of his Philosophical Investigations in
1953, a series of his later writings began to become available. Now, we have a sizeable
corpus of the work he did in the last twenty years of his life. Through painstaking analy-
sis of the use of language, particularly through analysis of psychological concepts,
Wittgenstein attempted to undermine the idea that philosophy is a foundational enter- | Blackwell |
prise. He asserted, on the contrary, that philosophy is a purely descriptive enterprise,
that the task of philosophy is neither to reform language nor to try to place the various
uses of language on a secure foundation. Rather, philosophical problems are removed
by having a correct understanding of how language actually functions.
A key notion in Wittgenstein’s conception of language is the notion of a language
game. We should think of the words in language as being like the pieces in a game. They
are not to be understood by looking for some associated idea in the mind, or by follow-
ing some procedure of verification, or even by looking at the object for which they stand.
8
CONTEMPORARY PHILOSOPHY IN THE UNITED STATES
Rather, we should think of words in terms of their use, and referring to objects in the
world is only one of many uses that words have. The meaning of a word is given by its
use, and the family of uses that a group of words has constitutes a language game.
Examples include the language game we play in describing our own sensations, or the
language game we play in identifying the causes of events. This conception of language
leads Wittgenstein to the rejection of the conception that the task of philosophical
analysis is reductionist or foundationalist. That is, Wittgenstein rejects the idea that lan-
guage games either have or need a foundation in something else, and he rejects the idea
that certain language games can be reduced to certain other kinds of language games.
The effect, Wittgenstein says, of philosophical analysis is not to alter our existing lin-
guistic practices or to challenge their validity; it is simply to describe them. Language
neither has nor needs a foundation in the traditional sense.
I said that Wittgenstein was the single most influential philosopher in the analytic
tradition, but there is a sense in which it seems to me he has still not been properly
understood, nor had his lessons been fully assimilated by analytic philosophers. I will
have more to say about his influence later.
2.4 Rawls’s theory of justice
The conception of moral philosophy in the positivist and post-positivist phases of ana-
lytic philosophy was extremely narrow. Strictly speaking, according to the positivists,
moral utterances could not be either true or false, so there was nothing that the philoso-
pher could say, qua philosopher, by way of making moral judgements. The task for the
moral philosopher was to analyse moral discourse, to analyse the meaning and use of
moral terms such as ‘good’, ‘ought’, ‘right’, ‘obligation’, etc. It is important to see that
this conception of moral philosophy was a strict logical consequence of the acceptance
of the distinction between evaluative and descriptive utterances. For if evaluative utter-
ances cannot be either true or false, and if first-order moral discourse consists in evalu-
ative utterances, and if the task of the philosopher is to state the truth, it follows that
the philosopher, qua philosopher, cannot make any first-order moral judgements. As a
philosopher, all he or she can do is the second-order task of analysing moral concepts.
Some philosophers of the positivist and post-positivist periods rejected this narrow
conception of moral philosophy, and there were a series of attacks mounted on the
distinction between evaluative and descriptive utterances, including some attacks by
myself in the mid-1960s (Searle 1964). It remained, however, for John Rawls to reopen
the traditional conception of political and moral philosophy with the publication of his
book A Theory of Justice in 1971. For the purposes of the present discussion, the im-
portant thing about Rawls’s work was not that he refuted the traditional dichotomy of
descriptive and evaluative utterances, but that he simply ignored it and proceeded to
develop a theory of political institutions of a sort that has a long philosophical tradi-
tion and which the positivists thought they had overcome. Rawls, in effect, revived the | Blackwell |
social contract theory, which had long been assumed to be completely defunct; but he
did it by an ingenious device: he did not attempt, as some traditional theorists had done,
to show that there might have been an original social contract, nor did he try to show
that the participation of individuals in society involved a tacit contract. Rather, he used
the following thought experiment as an analytic tool: think of the sort of society that
9
JOHN R. SEARLE
rational beings would agree to if they did not know what sort of position they them-
selves would occupy in that society. If we imagine rational beings, hidden behind a veil
of ignorance, who are asked to select and agree on forms of social institutions that
would be fair for all, then we can develop criteria for appraising social institutions on
purely rational grounds.
The importance of Rawls for our present discussion is not whether he succeeded in
developing new foundations for political theory, but the fact that his work gave rise to
a renewed interest in political philosophy, which was soon accompanied by a renewed
interest in the traditional questions of moral philosophy. Moral and political philoso-
phy had been confined to a very small realm by the positivist philosophers, and for that
reason seemed sterile and uninteresting. Very little work was done in that area, but
since the 1970s it has grown enormously, and is now a flourishing branch of analytic
philosophy.
2.5 Post-positivist philosophy of science
Throughout the positivist period the model of empirical knowledge was provided by the
physical sciences, and the general conception was that the empirical sciences proceeded
by the gradual but cumulative growth of empirical knowledge through the systematic
application of scientific method. There were different versions of scientific method,
according to the philosophers of that period, but they all shared the idea that scientific,
empirical propositions are essentially ‘testable’. Initially a proposition was thought
testable if it could be confirmed, but the most influential version of this idea is Popper’s
claim that empirical propositions are testable if they are falsifiable in principle. That is,
in order for a proposition to tell us how the world is as opposed to how it might be or
might have been, there must be some conceivable state of affairs that would render that
proposition false. Propositions of science are, strictly speaking, never verifiable – they
simply survive repeated attempts at falsification. Science is in this sense fallible, but it
is at the same time rational and cumulative.
This picture of the history of science was very dramatically challenged in Thomas
(1962). According to Kuhn, the
Kuhn’s book The Structure of Scientific Revolutions
history of science shows not a gradual and steady accumulation of knowledge but
periodic revolutionary overthrows of previous conceptions of reality. The shift from
Aristotelian physics to Newtonian physics, and the shift from Newtonian physics to
relativistic physics are both illustrations of how one ‘paradigm’ is replaced by another.
When the burden of puzzling cases within one paradigm becomes unbearable, a new
paradigm emerges, which provides not just a new set of truths but a whole new way
of looking at the subject matter. ‘Normal sciences’ always proceed by puzzle-solving
within a paradigm, but revolutionary breakthroughs, rather than puzzle-solving
within a paradigm, are matters of overthrowing one paradigm and replacing it with
another.
Just as Kuhn challenged the picture of science as essentially a matter of a steady
accumulation of knowledge, so Paul Feyerabend challenged the conception of there
being a unitary rational ‘scientific method’ (Feyerabend 1975). Feyerabend tried to
show that the history of science reveals not a single rational method but rather a series
of opportunistic, chaotic, desperate (and sometimes even dishonest) attempts to cope
10
| Blackwell |
CONTEMPORARY PHILOSOPHY IN THE UNITED STATES
with immediate problems. The lesson that Feyerabend draws from this is that we should
abandon the constraining idea of there being such a thing as a single, rational method
that applies everywhere in science; rather, we should adopt an ‘anarchistic’ view,
according to which ‘anything goes’. Reactions to Kuhn and Feyerabend, not surpris-
ingly, differ enormously among analytic philosophers. Kuhn sometimes seems to be
arguing that there is not any such thing as the real world existing independently of our
scientific theories, which it is the aim of our scientific theories to represent. Kuhn, in
short, seems to be denying realism. Most philosophers do not take this denial of realism
at all seriously. Even if Kuhn were right about the structure of scientific revolutions,
this in no way shows that there is no independent reality that science is investigating.
Again, most philosophers would accept Feyerabend’s recognition of a variety of
methods used in the history of science, but very few people take seriously the idea that
there are no rational constraints on investigation whatever. Nonetheless, the effect of
these authors has been important in at least the following respect. The positivists’ con-
ception of science as a steady accumulation of factual knowledge, and of the task of
the philosopher as the conceptual analysis of scientific method, has given way to an
attitude to science that is at once more sceptical and more activist. It is more sceptical
in the sense that few philosophers are looking for the one single method that pervades
every enterprise called ‘science’, but it is more activist in the sense that philosophy of
science interacts more directly with scientific results. For example, recent philosophi-
cal discussions about quantum mechanics, or about the significance of Bell’s theorem
within quantum mechanics, reveal that it is now impossible to say exactly where the
problem in physics ends and the problem in philosophy begins. There is a steady inter-
action and collaboration between philosophy and science on such philosophically
puzzling questions.
3 Some Recent Developments
The results of the changes that I have just outlined are to make analytic philosophy on
the one hand a more interesting discipline, but on the other hand a much less well-
defined research project. In the way that the verification principle formed the core
ideology of the logical positivists and in the way that the conceptual analysis formed
the core research project of the post-positivistic analytic philosopher, there is now no
ideological point of reference that is commonly agreed upon; nor is there a universally
accepted research programme. For example, conceptual analysis thirty years ago was
taken to be the heart of analytic philosophy, but now many philosophers would deny
that it is the central element in the philosophical enterprise. Some philosophers, indeed,
would say that the traditional enterprise of attempting to find logically necessary and
sufficient conditions for the applicability of a concept is misconceived in principle. They
think the possibility of such an enterprise has been refuted by Quine’s refutation of the
analytic–synthetic distinction, as well as Wittgenstein’s observation that many philo-
sophically puzzling concepts have not a central core or essence of meaning, but a
variety of different uses united only by a ‘family resemblance’. Many other philosophers
would say that conceptual analysis is still an essential part of the philosophical enter-
prise, as indeed it has been since the time of Plato’s dialogues, but it is no longer seen
11
JOHN R. SEARLE
to be the whole of the enterprise. Philosophy is now, I believe, a much more interesting
subject than it was a generation ago because it is no longer seen as something separate
from, and sealed off from, other disciplines. In particular, philosophy is now seen by | Blackwell |
most analytic philosophers as being adjacent to and overlapping with the sciences.
My own view, which I feel is fairly widely shared, is that words like ‘philosophy’ and
‘science’ are in many respects misleading, if they are taken to imply the existence of
mutually exclusive forms of knowledge. Rather, it seems to me that there is just knowl-
edge and truth, and that in intellectual enterprises we are primarily aiming at knowl-
edge and truth. These may come in a variety of forms, whether in history, mathematics,
physics, psychology, literary criticism or philosophy. Philosophy tends to be more
general than other subjects, more synoptic in its vision, more conceptually or logically
oriented than other disciplines, but it is not a discipline that is hermetically sealed off
from other subjects. The result is that many areas of investigation which were largely
ignored by analytic philosophers a generation ago have now become thriving branches
of philosophy, including cognitive science, the philosophy of biology and the philoso-
phy of economics. In what follows, I will confine my discussion to five major areas of
philosophical research: cognitive science, the causal theory of reference, intentionalis-
tic theories of meaning, truth-conditional theories of meaning, and Wittgenstein’s
conception of language and mind and his response to scepticism.
3.1 Philosophy and cognitive science
Nowhere is the new period of collaboration between philosophy and other disciplines
more evident than in the new subject of cognitive science. Cognitive science from its
very beginnings has been ‘interdisciplinary’ in character, and is in effect the joint prop-
erty of psychology, linguistics, philosophy, computer science and anthropology. There
is, therefore, a great variety of different research projects within cognitive science, but
the central area of cognitive science, its hardcore ideology, rests on the assumption that
the mind is best viewed as analogous to a digital computer. The basic idea behind cog-
nitive science is that recent developments in computer science and artificial intelligence
have enormous importance for our conception of human beings. The basic inspiration
for cognitive science went something like this: human beings do information process-
ing. Computers are designed precisely to do information processing. Therefore one way
to study human cognition – perhaps the best way to study it – is to study it as a matter
of computational information processing. Some cognitive scientists think that the com-
puter is just a metaphor for the human mind; others think that the human mind is
literally a computer program. But it is fair to say that without the computational
model there would not have been a cognitive science as we now understand it.
This conception of human cognition was ideally suited to the twentieth-century
analytic tradition in philosophy of mind because of the analytic tradition’s resolute
materialism. It was anti-mentalistic and anti-dualistic. The failure of logical behav-
iourism led not to a revival of dualism but to more sophisticated versions of material-
ism. I will now briefly summarize some of the recent developments in materialistic
philosophies of mind that led to the computational theory of the mind.
The logical behaviourists’ thesis was subject to many objections, the most important
being the objection that it ignores internal mental phenomena. In science and common
12
CONTEMPORARY PHILOSOPHY IN THE UNITED STATES
sense it seems more natural to think of human behaviour as being caused by internal
mental states rather than to think of the mental states as simply consisting of the
behaviour. This weakness in behaviourism was corrected by the materialist identity
thesis, sometimes called ‘physicalism’. According to the physicalist identity theory,
mental states are identical with states of the brain. We do not know in detail what these | Blackwell |
identities are, but the progress of the neurosciences makes it seem overwhelmingly
probable that every mental state will be discovered to be identical with some brain state.
In the early version of the identity thesis it was supposed that every type of mental state
would be discovered to be identical with some type of physical state, but after some
debate this began to seem more and more implausible. There is no reason to suppose
that only systems with neurons like ours can have mental states; indeed, there is no
reason to suppose that two human beings who have the same belief must therefore be
in the same neurophysiological state. So, ‘type–type identity theory’ naturally gave way
to ‘token–token identity theory’. The token identity theorists claimed that every par-
ticular mental state is identical with some particular neurophysiological state, even if
there is no type correlation between types of mental states and types of physical states.
But that only leaves open the question, ‘What is it that two different neurophysiologi-
cal states have in common if they are both the same mental state?’ To many analytic
philosophers it seemed obvious that the answer to our question must be that two
neurophysiological states are the same type of mental state if they serve the same
function in the overall ecology of the organism. Mental states on this view can be
defined in terms of their causal relations to input stimuli, to other mental states,
and to external behaviour. This view is called ‘functionalism’ and it is a natural
development from token–token identity theory.
However, the functionalist has to answer a further obvious question: ‘What is it
about the states that gives them the causal relations that they do have?’ If mental states
are defined in terms of their causal relations, then what is it about the structure of dif-
ferent neurophysiological configurations that can give them the same causal relations?
It is at precisely this point that the tradition of materialism in analytic philosophy con-
verges with the tradition of artificial intelligence. The computer provides an obvious
answer to the question that I have just posed. The distinction between the software and
the hardware, the program and the physical system that implements the program, pro-
vides a model for how functionally equivalent elements at a higher level can be realized
in or implemented by different physical systems at a lower level. Just as one and the
same program can be implemented by quite different physical hardware systems, so one
and the same set of mental processes can be implemented in different neurophysiologi-
cal or other forms of hardware implementations. Indeed, on the most extreme version
of this view, the mind is to the brain as the program is to the hardware. This sort of
functionalism came to be called ‘computer functionalism’ or ‘Turing machine func-
tionalism’, and it coincides with the strong version of ‘artificial intelligence’ (Strong
AI), the version that says having a mind just is having a certain sort of program.
I have refuted Strong AI in a series of articles (Searle 1980a, 1980b). The basic idea
of that refutation can be stated quite simply. Minds cannot be equivalent to programs
because programs are defined purely formally or syntactically and minds have mental
contents. The easiest way to see the force of the refutation is to see that a system, say
oneself, could learn to manipulate the formal symbols for understanding a natural
13
JOHN R. SEARLE
language without actually understanding that language. I might have a program that
enables me to answer questions in Chinese simply by matching incoming symbols
with the appropriate processing and output symbols, but nonetheless I still would not
thereby understand Chinese. However, though the project of computer functionalism
is almost certainly a failure, the results of the enterprise are in many respects quite | Blackwell |
useful. Important things can be learned about the mind by pursuing the computer
metaphor, and the research effort has not necessarily been wasted. The most exciting
recent development has been to think of mental processes not on the model of the con-
ventional serial digital computer, but rather to think of brain processes on the model
of parallel distributed processing computers. The most exciting recent development, in
my view, in cognitive science has been the development of such ‘neural net models’ for
human cognition.
In concluding this section, I want to point out that in my view the chief weakness
of analytical philosophy of mind, a weakness it shares with the past 300 years in the
philosophy of mind, has been its assumption that there is somehow an inconsistency
between mentalism and materialism. Analytic philosophers, along with the rest of
the Cartesian tradition, have characteristically assumed that ‘mental’ implies ‘non-
material’ or ‘immaterial’ and that ‘material’ or ‘physical’ implies ‘non-mental’. But if
one reflects on how the brain works, it seems that both of these assumptions are
obviously false. What that shows is that our whole vocabulary, our whole terminology
of the mental and physical, needs wholesale revision.
3.2 The causal theory of reference
A central question in analytic philosophy of language, since Frege (and indeed in phi-
losophy since the time of Plato), has been: How does language relate to the world? How
do words hook on to things? In answering this question, the analytic tradition had char-
acteristically found a connection between the notion of reference and the notion of
truth. An expression, such as a proper name, refers to or stands for or designates an
object because associated with that name is some descriptive content, some concept of
the object in question, and the object in question satisfies or fits that descriptive content.
The expression refers to the object only because the description is true of the object.
This is the standard reading of Frege’s famous distinction between sense and reference,
between Sinn and Bedeutung. Expressions refer to objects in virtue of their sense and
the sense provides a description, a ‘mode of presentation’, of the object in question.
Something analogously applies with general terms: general terms are true of an object
because each general term has associated with it a cluster of features, and the term will
be true of the object if the object in question has those features.
In the 1970s this conception of the relation between language and reality was
attacked by a number of philosophers, most prominently Donnellan (1970), Kripke
(1972) and Putnam (1975). A variety of arguments were mounted against the tradi-
tional conception of meaning and reference, but the common thread running through
these arguments was that the descriptive content associated with a word provided
neither necessary nor sufficient conditions for its application. A speaker might refer to
an object even though the associated description that he or she had was not true of that
object; a speaker might have a description that was satisfied by an object even though
14
CONTEMPORARY PHILOSOPHY IN THE UNITED STATES
that was not the object to which he or she was referring. The most famous version of
this argument was Putnam’s ‘twin earth’ example. Imagine a planet in a distant galaxy
exactly like ours in every respect except that on this planet what they call ‘water’ has
a different chemical composition. It is not composed of H2O but has an extremely com-
plicated formula that we will abbreviate as ‘XYZ’. Prior to 1750, prior to the time that
anyone knew the chemical composition of water, the people on twin earth had in their
minds exactly the same concept of water as the people on earth. Nonetheless our word
‘water’ does not refer to the stuff on twin earth. Our word ‘water’, whether or not we
knew it in 1750, refers to H2O; and this is a matter of objective causal relations in the | Blackwell |
world which are independent of the ideas that people have in their heads. Meanings on
this view are not concepts in people’s heads, but objective relations in the world. Well,
if associated ideas are not sufficient for meaning, what is? The answer given by the three
authors I have mentioned is that there must be some sort of causal connection between
the use of the word and the object or type of entity in the world that it applies to. Thus,
if I use the word ‘Socrates’, it refers to a certain Greek philosopher only because there
is a causal chain connecting that philosopher and my current use of the word. The
word ‘water’ is not defined by any checklist of features; rather, ‘water’ refers to what-
ever stuff in the world was causally related to certain original uses of the word ‘water’,
and these uses subsequently came to be accepted in the community and were then
passed down through a causal chain of communication.
There is a very natural way of connecting the computer functionalist conception of
the mind with the causal theory of reference. If the mind were a computer program,
and if meaning were a matter of causal connections to the world, then the way the
mind acquires meanings is for the system that implements the computer program to be
involved in causal interactions with the world.
3.3 Intentionalistic theories of meaning
Much of the best work in speech act theory done after the publication of Austin’s How
to Do Things with Words in 1962, and my Speech Acts in 1969, attempted to combine
the insights of Paul Grice’s account of meaning with the framework provided by the
theory of speech acts. In a series of articles beginning in the late 1950s (Grice 1957,
1968), Grice had argued that there is a close connection between the speaker’s inten-
tions in the performance of an utterance and the meaning of that utterance. In his
original formulation of this view, Grice analysed the speaker’s meaning in terms of the
intention to produce an effect on the hearer by means of getting the hearer to recog-
nize the intention to produce that very effect. Thus, for example, according to Grice, if
a speaker intends to tell a hearer that it is raining, then in the speaker’s utterance of
the sentence, ‘It is raining’, the speaker’s meaning will consist of his or her intention
to produce in the hearer the belief that it is raining by means of getting the hearer to
recognize his or her intention to produce that very belief. Subsequent work by Grice
altered the details of this account, but the general principle remained the same:
meaning is a matter of a self-referential intention to produce an effect on a hearer
by getting the hearer to recognize the intention to produce that effect. Grice combined
this analysis of meaning with an analysis of certain principles of conversational co-
operation. In conversation, people accept certain tacit principles, which Grice calls
15
JOHN R. SEARLE
‘Maxims of Conversation’ – they accept the principles that the speaker’s remarks will
be truthful and sincere (the maxim of quality), that they will be relevant to the
conversational purposes at hand (the maxim of relation), that the speaker will be clear
(the maxim of manner) and that the speaker will say neither more nor less than is
necessary for the purposes of the conversation (the maxim of quantity).
There has been a great deal of controversy about the details of Grice’s analysis of
meaning, but the basic idea that there is a close connection between meaning and
intention has been accepted and has proved immensely useful in analysing the struc-
ture of certain typical speech act phenomena. My own view is that Grice confuses that
part of meaning which has to do with representing certain states of affairs and certain
illocutionary modes, and that part of meaning that has to do with communicating
those representations to a hearer. Grice, in short, confuses communication with | Blackwell |
representation. However, the combination of an intentionalistic account of meaning,
together with rational principles of co-operation, is immensely fruitful in analysing
such problems as those of ‘indirect speech acts’ and figurative uses of language such
as metaphors. So, for example, in an indirect speech act, a speaker will characteristi-
cally mean something more than what he or she actually says. To take a simple
example, in a dinner table situation a speaker who says ‘Can you pass the salt?’ would
usually not just be asking a question about the salt-passing abilities of the hearer; he
or she would be requesting the hearer to pass the salt. Now the puzzle is this: how is it
that speakers and hearers communicate so effortlessly when there is a big gulf between
what the speaker means and what he or she actually says? In the case of metaphor, a
similar question arises: how does the speaker communicate so effortlessly his or her
metaphorical meaning when the literal meaning of the sentence uttered does not
encode that metaphorical meaning? A great deal of progress has been made on these
and other problems using the apparatus that Grice contributed to the theory of speech
acts.
One of the marks of progress in philosophy is that the results of philosophical analy-
sis tend to be appropriated by other disciplines, and this has certainly happened with
speech act theory. Speech act theory is now a thriving branch of the discipline of lin-
guistics, and the works of Austin and Grice, as well as my own, are as well known
among linguists as they are among philosophers.
3.4 Truth-conditional theories of meaning
Philosophers such as Quine and his former student, Donald Davidson, have always felt
that intentionalistic theories of meaning of the sort proposed by Grice and Searle were
philosophically inadequate, because the intentionalistic notions seemed as puzzling as
the notion of meaning itself and because they could necessarily involve linguistic
meaning in their ultimate analyses. So Quine and Davidson attempted to give accounts
of meaning that did not employ the usual apparatus of intentionality. The most influ-
ential version of this attempt is Davidson’s project of analysing meaning in terms of
truth conditions. The basic idea is that one knows the meaning of a sentence if one
knows under what conditions it is true or false. Thus, one knows the meaning of the
German sentence ‘Schnee ist weiss’ if one knows that it is true if and only if snow is
white. Now since a theory of meaning for a language should be able to state the
16
CONTEMPORARY PHILOSOPHY IN THE UNITED STATES
meaning of each of the sentences of the language, and since the meanings of the sen-
tences of the language are given by truth conditions, and since truth conditions can be
specified independently of the intentionalistic apparatus, it seems to Davidson that a
theory of truth (that is, a theory of the truth conditions of the sentences) of a language
would provide a theory of meaning for that language.
In order to carry out the project of explaining meaning in terms of truth, Davidson
employs the apparatus of Tarski’s semantic definition of truth, a definition that Tarski
had worked out in the 1930s. Tarski points out that it is a condition of adequacy on
any account of truth that for any sentence s and any language L, the account must
have the consequence that
s is true in L if and only if p,
where for s can be substituted the structural description of any sentence whatever, for
L, the name of the language of which s is a part, and for p, the sentence itself or a trans-
lation of it. Thus, for example, in English, the sentence ‘Snow is white’ is true if and
only if snow is white. This condition is usually called ‘convention T’ and the corre-
sponding sentences are called ‘T-sentences’.
Now Davidson notes that convention T employs the fact that the sentence named by
s has the same meaning as the sentence expressed by p, and thus Tarski is using the | Blackwell |
notion of meaning in order to define the notion of truth. Davidson proposes to turn this
procedure around by taking the notion of truth for granted, by taking it as a primitive,
and using it to explain meaning.
Here is how it works. Davidson hopes to get a theory of meaning for a speaker of a
language that would be sufficient to interpret any of the speaker’s utterances by getting
a theory that would provide a set of axioms which would entail all true T-sentences for
that speaker’s language. Thus, if the speaker speaks German, and we use English as a
meta-language in which to state the theory of the speaker’s language, Davidson claims
we would have an adequate theory of the speaker’s language if we could get a set of
axioms which would entail a true T-sentence stated in English for any sentence that the
speaker uttered in German. Thus, for example, our theory of meaning should contain
axioms which entail that the speaker’s utterance ‘Schnee ist weiss’ is true in the
speaker’s language if and only if snow is white. Davidson further claims that we could
make this into an empirical theory of the speaker’s language by proceeding to associ-
ate the speaker’s utterances with the circumstances in which we had empirical evidence
for supposing that the speaker held those utterances to be true. Thus, if we hear the
speaker utter the sentence ‘Es regnet’, we might look around and note that it was
raining in the vicinity, and we might then form the hypothesis that the speaker holds
true the sentence ‘Es regnet’ when it is raining in his or her immediate vicinity. This
would provide the sort of empirical data on which we would begin to construct a theory
of truth for the speaker’s language.
It is important to note that we are to think of this as a thought experiment and not
as an actual procedure that we have to employ when we try to learn German, for
example. The idea is to cash out the notion of meaning in terms of truth conditions,
and then cash out the notion of truth conditions in terms of a truth theory for a lan-
guage, which is a theory that would entail all the true T-sentences of the language. The
17
JOHN R. SEARLE
empirical basis on which the whole system rests is that of the evidence we could get
concerning the conditions under which a speaker holds a sentence to be true. If the
project could in principle be carried out, then we would have given an account of
meaning which employed only one intentionalistic notion, the notion of ‘holding true’
a sentence.
Over the past twenty years there has been quite an extensive literature on the
nature of this project and how it might be applied to several difficult and puzzling
sorts of sentences – for example, indexical sentences, sentences about mental states or
modal sentences. Enthusiasm for this project seems to have waned somewhat in recent
years.
In my view, the central weakness of Davidson’s enterprise is as follows: any theory
of meaning must explain not only what a speaker represents by his or her utterances,
but also how he or she represents them, under what mental aspects the speaker repre-
sents truth conditions. For this reason, a theory of meaning cannot just correlate a
speaker’s utterance with states of affairs in the world; it must explain what is going on
in the speaker’s head which enables the speaker to represent those states of affairs
under certain aspects with the utterances that the speaker makes. Thus, for example,
suppose that snow is composed of H2O molecules in crystalline form, and suppose the
colour white consists of light wave emissions of all wavelengths, then the sentence
‘Schnee ist weiss’ is true if and only if H2O molecules in crystalline form emit light of
all wavelengths. Now this second T-sentence is just as empirically substantiated as the
earlier example, ‘Schnee ist weiss’ is true if and only if snow is white. Indeed, it is a
matter of scientific necessity that the state of affairs described by the former is identi- | Blackwell |
cal with the state of affairs described by the latter. But the former example simply does
not give the speaker’s meaning. The speaker might hold true the sentence ‘Schnee ist
weiss’ under these and only these conditions and not know the slightest thing about
H2O molecules and wavelengths of light. The T-sentence gives the truth conditions, but
the specification of the truth conditions does not necessarily give the meaning of the
sentence, because the specification does not yet tell us how the speaker represents those
truth conditions. Does he or she represent them under the aspect of snow being white,
or what is the same fact in the world, does he or she represent them under the aspect
of frozen H2O crystals emitting light of all wavelengths? Any theory that cannot give
that information is not a theory of meaning.
There are various attempts to meet these sorts of objections, but, in my view, they
are not successful. In the end, all truth definitional accounts of meaning, like the be-
haviourist accounts which preceded them, end up with a certain ‘indeterminacy’ of
meaning. They cannot account in objective terms for all of the subjective details of
meaning, and both Davidson and Quine have acknowledged that their views result in
indeterminacy.
3.5 Wittgenstein’s legacy
Wittgenstein’s work covers such a vast range of topics, from aesthetics to math-
ematics, and covers these topics with so much depth and insight, that it continues to
be a source of ideas and inspiration for analytic philosophers and is likely to continue
to be so for many years to come. I will mention only three areas.
18
CONTEMPORARY PHILOSOPHY IN THE UNITED STATES
3.5.1 Philosophical psychology
One of Wittgenstein’s main areas of research was that of psychological concepts such
as belief, hope, fear, desire, want, expect and sensation concepts such as pain and
seeing. Perhaps his single most controversial claim in this area is that concerning a
private language. He claims that it would be logically impossible for there to be a lan-
guage that was private in the sense that its words could only be understood by the
speaker because they referred to the speaker’s private inner sensations and had no
external definition. Such a language would be absurd, he said, because for the applica-
tion of such words there would be no distinction between what seemed right to the
speaker and what really was right. But unless we can make a distinction between what
seems right and what really is right, we cannot speak of right or wrong at all, and hence
we cannot speak of using a language at all. ‘An inner process’, says Wittgenstein
(1953), ‘stands in need of outward criteria’. Wittgenstein is here attacking the entire
Cartesian tradition, according to which there is a realm of
inner private objects,
our inner mental phenomena, and the meanings of the words that stand for these
entities are entirely defined by private ostensive definitions. No other single claim of
Wittgenstein’s has aroused as much controversy as the ‘private language argument’.
It continues to be a source of fascination to contemporary philosophers, and many
volumes have been written about Wittgenstein’s analysis of psychological concepts.
3.5.2 Following a rule
Wittgenstein is part of a long tradition that emphasizes the distinction between the
modes of explanation of the natural sciences and the modes of explanation of human
behaviour and human cultural and psychological phenomena generally. His analysis
of this problem chiefly deals with the phenomenon of human behaviour which is influ-
enced or determined by mental contents, and, most importantly, by the phenomena
of human beings following a rule. What is it for a human being to follow a rule?
Wittgenstein’s analysis of this stresses the difference between the way that rules
guide human behaviour and the way that natural phenomena are results of causes. | Blackwell |
Wittgenstein throughout emphasizes the difference between causes and reasons, and he
also emphasizes the roles of interpretation and rule following. On the most extreme
interpretation of Wittgenstein’s remarks about following a rule, he is the proponent of
a certain type of scepticism. According to one view of Wittgenstein, he is arguing that
rules do not determine their own application, that anything can be interpreted to accord
with a rule, and consequently that anything can be interpreted to conflict with a rule. If
taken to its extreme, this argument would have the consequence that, logically speaking,
rules do not constrain human behaviour at all. And if that is right, then mental contents,
such as knowledge of meanings of words or principles of action or even beliefs and
desires, do not constrain human behaviour, because they are everywhere subject to an
indefinite range of different interpretations. Wittgenstein’s solution to this scepticism is
to propose that interpretation comes to an end when we simply accept the cultural
practices of the community in which we are imbedded. Interpretation comes to an end,
and we just act on a rule. Acting on a rule is a practice, and it is one that we are brought
up to perform in our culture. The sceptical implications of Wittgenstein’s account of rule
following are resolved by an appeal to a naturalistic solution: we are simply the sort of
beings who follow culturally and biologically conditioned practices.
19
JOHN R. SEARLE
This interpretation of Wittgenstein is largely due to Saul Kripke (1982) and it has
aroused considerable controversy. My own view is that Kripke has misinterpreted
Wittgenstein in certain crucial respects, but whether or not his interpretation is correct,
it has been a source of continuing discussion in contemporary philosophy.
3.5.3 Philosophical scepticism
Important work on philosophical scepticism has been continued by philosophers who
are inspired or provoked by Wittgenstein, notably Thomson Clarke and Barry Stroud.
These philosophers point out that a really serious analysis of our use of epistemic dis-
course shows that the problem of scepticism cannot be simply overcome by the usual
analytic philosopher’s methods of pointing out that the sceptic raises the demand
for justification beyond that which is logically appropriate. Clarke and Stroud claim
that the problem of scepticism goes deeper than this solution will allow. Following
Wittgenstein in investigating the depth grammar of
language, they find that any
solution to the sceptic’s predicament – that is, any justification for our claims to have
knowledge about the world – rests on a much deeper understanding of the difference
between ordinary or plain discourse and philosophical discourse. Work in this line of
research is continuing at present.
4 Overall Assessment
I have not attempted to survey all of the main areas of activity in contemporary ana-
lytic philosophy. Most importantly, I have left out contemporary work in ethics. Perhaps
of comparable importance, I have had nothing to say about purely technical work in
logic. There is, furthermore, a thriving branch of analytic philosophy called ‘action
theory’, which should be mentioned at least in passing. The general aim of analytic
action theory is to analyse the structure of human actions in terms of the causal rela-
tions between such mental states as beliefs, desires and intentions, and the bodily move-
ments which are in some sense constitutive of the actions. Finally, it is worth calling
attention to the fact that among analytic philosophers there has been a great revival of
interest in the history of philosophy. Traditional analytic philosophers thought of the
history of philosophy as mostly the history of mistakes. Some of the history of the
subject could be useful for doing real philosophy; but the overall conception was that
the history of philosophy had no more special relevance to philosophy than the history | Blackwell |
of mathematics to mathematics, or the history of chemistry to chemistry. This attitude
has changed recently, and there is now a feeling of the historical continuity of analytic
philosophy with traditional philosophy in a way that contrasts sharply with the origi-
nal view of analytic philosophers, who thought that they marked a radical, or indeed,
revolutionary break with the philosophical tradition.
It is too early to provide an assessment of the contribution that will be made by work
done in philosophy at the present time, or even in the past few decades. My own view
is that the philosophy of mind and social philosophy will become ever more central to
the entire philosophical enterprise. The idea that the study of language could replace
the study of mind is itself being transformed into the idea that the study of language
is really a branch of the philosophy of mind. Within the philosophy of mind, perhaps
20
CONTEMPORARY PHILOSOPHY IN THE UNITED STATES
the key notion requiring analysis is that of intentionality – that property of the mind
by which it is directed at or about or of objects and states of affairs in the world inde-
pendent of itself. Most of the work done by analytic philosophers in the philosophy of
mind has tended to cluster around the traditional mind–body problem. My own view
is that we need to overthrow this problem: in its traditional version, it was based on the
assumption that mental properties and physical properties were somehow different
from each other, and that therefore, there was some special problem not like other prob-
lems in biology as to how they could both be characteristics of the human person. Once
we see that so-called mental properties really are just higher-level physical properties
of certain biological systems, I believe this problem can be dissolved. Once it is dissolved,
however, we are still left with the task of analysing what is the central problem in the
philosophy of language and in cognitive science, as well as the philosophy of mind,
namely, the way that human representational capacities relate the human organism to
the world. What are called ‘language’, ‘mind’, ‘thinking’, ‘speaking’ and ‘depicting’ are
just different aspects of this mode of relating to reality.
I believe that the causal theory of reference will be seen to be a failure once it is
recognized that all representations must occur under some aspect or other, and that
the extensionality of causal relations is inadequate to capture the aspectual character
of reference. The only kind of causation that could be adequate to the task of reference
is intentional causation or mental causation, but the causal theory of reference cannot
concede that ultimately reference is achieved by some mental device, since the whole
approach behind the causal theory was to try to eliminate the traditional mentalism of
theories of reference and meaning in favour of objective causal relations in the world.
My prediction is that the causal theory of reference, though it is at present by far the
most influential theory of reference, will prove to be a failure for these reasons.
Perhaps the single most disquieting feature of analytic philosophy in the fifty-year
period that I have been discussing is that it has passed from being a revolutionary
minority point of view held in the face of traditionalist objections to becoming itself the
conventional, establishment point of view. Analytic philosophy has become not only
dominant but intellectually respectable, and, like all successful revolutionary move-
ments, it has lost some of its vitality in virtue of its very success. Given its constant
demand for rationality, intelligence, clarity, rigour and self-criticism, it is unlikely that
it can succeed indefinitely, simply because these demands are too great a cost for many
people to pay. The urge to treat philosophy as a discipline that satisfies emotional rather
than intellectual needs is always a threat to the insistence on rationality and intelli- | Blackwell |
gence. However, in the history of philosophy, I do not believe we have seen anything to
equal the history of analytic philosophy for its rigour, clarity, intelligence and, above
all, its intellectual content. There is a sense in which it seems to me that we have been
living through one of the great eras in philosophy.
References
Austin, J. L. 1962: How to do Things with Words. Oxford: Clarendon Press.
Donnellan, K. 1970: Proper Names and Identifying Descriptions. Synthèse , 21, 335–58.
Feyerabend, P. 1975: Against Method. London: Humanities Press.
Grice, H. P. 1957: Meaning. Philosophical Review, 66.
21
JOHN R. SEARLE
—— 1968: Utterer’s Meaning, Sentence-Meaning, and Word-Meaning. Foundations of Language,
4, 1–18.
Kripke, S. 1972: Naming and Necessity. In G. Harman and D. Davidson (eds) Semantics of Natural
Language, Dordrecht: Reidel.
—— 1982: Wittgenstein on Rules and Private Language. Cambridge, MA: Harvard University Press.
Kuhn, T. 1962: The Structure of Scientific Revolutions . Chicago: University of Chicago Press.
Putnam, H. 1975: The Meaning of ‘Meaning’. In his Philosophical Papers, Vol. 2: Mind, Language
and Reality, Cambridge: Cambridge University Press.
Quine, W. V. O. 1953: Two Dogmas of Empiricism. In his From a Logical Point of View, Cambridge,
MA: Harvard University Press.
Rawls, J. 1971: A Theory of Justice. Cambridge, MA: Harvard University Press.
Searle, J. R. 1964: How to Derive ‘Ought’ from ‘Is’. Philosophical Review, 73.
—— 1969: Speech Acts: An Essay in the Philosophy of Language. Cambridge: Cambridge University
Press.
—— 1979: Expression and Meaning. Cambridge: Cambridge University Press.
—— 1980a: Minds, Brains and Programs. Behavioral and Brain Sciences, 3, 417–24.
—— 1980b: Intrinsic Intentionality. Behavioral and Brain Sciences, 3, 450–6.
—— 1983: Intentionality: An Essay in the Philosophy of Mind. Cambridge: Cambridge University
Press.
Wittgenstein, L. 1953: Philosophical Investigations (translated by G. E. M. Anscombe). Oxford:
Blackwell.
22
Contemporary Philosophy: A Second Look
B E R NA R D W I L L I A M S
1 The Identity of Analytical Philosophy
Given the title of John Searle’s essay, this second introduction might have been expected
to complement the first geographically, by dealing with present philosophical develop-
ments in places other than the United States, but this is not in fact what it will try to
do. Philosophy in the United States, in other English-speaking parts of the world, and
in many other countries as well, is now very largely the same. In these places, there is
one philosophical culture, and inasmuch as it contains different approaches, and some
of the philosophy that is done within that culture is distinct from ‘analytical’ philoso-
phy, that itself is not a matter of geographical region.
It is true that ‘analytical’ philosophy, the style of philosophy described in Searle’s
essay and overwhelmingly represented in this volume, is often professionally distin-
guished (in job advertisements, for instance) from ‘continental’ philosophy, and this
does represent, in a clumsy way, something which until recently was true: that the ways
in which philosophy was done in France, Germany and other countries of continental
Europe were typically different from the ‘analytical’ style. To a much more limited
extent, that remains so. (Chapters 40–2 describe the situation in continental Europe.)
However, it is absurd to mark philosophical differences with these two labels. Apart from
involving a strange cross-classification – rather as though one divided cars into front-
wheel drive and Japanese – the labels are seriously misleading, in helping one to forget
lay in continental Europe (notably
that the origins of analytical philosophy itself
so, when its founding father is taken to be Frege and its greatest representative
Wittgenstein), and that the interests of ‘continental’ philosophy are not confined to
the European continent.
Moreover, it is not simply a matter of labelling. It is not that the distinction in itself | Blackwell |
is unproblematical, and only needs more aptly chosen titles to represent it. The
distinctions involved are obscure, and the titles serve to conceal this fact. The term
‘continental’ serves to discourage thought about the possible contrasts to analytical
philosophy, and so about the identity of analytical philosophy itself. At the same time,
the vague geographical resonance of the term does carry a message, that analytical
philosophy is familiar as opposed to exotic, and perhaps – if some older stereotypes
are in play – that it is responsible as opposed to frivolous. This is indeed what many
BERNARD WILLIAMS
analytical philosophers believe about it, and they believe it not so much in contrast to
activities going on in some remote continent, but, in many cases, as opposed to work
done in their local departments of literature. It is not true that work in other styles does
not exist in the heartlands of analytical philosophy; it merely does not exist in depart-
ments of philosophy. The distinctions involved are not geographical but professional,
and what is at issue is the identity of philosophy as a discipline.
In particular, what is at issue is the identity of philosophy as a subject that can sustain
ongoing, cumulative, research. If it can do this, it can make a claim which the human-
ities do not always find it easy to make, except to the extent that they are branches of
history: that there is something to be found out within their disciplines, that they can
add to knowledge. It has been part of the attraction of analytical philosophy that,
without the procedures of the experimental or theoretical sciences and with a more
human subject matter, it can claim to achieve results which command, if not agree-
ment, at least objective discussion, and which represent intellectual progress. It has
achievements that are not arbitrarily personal, and they compare favourably to those of
the social sciences (at least if one leaves aside the quite peculiar case of economics).
I do not think that these claims are empty. I think that the achievements of ana-
lytical philosophy are remarkable, and I agree with Searle that the subject is in various
ways more interesting than it was forty years ago. Its virtues are indeed virtues. I think
it is hard to be in good faith a teacher of philosophy unless you believe that there is
something worth doing that you, in virtue of your experience, can do and which you
can help other people to do. I think that the virtues of good philosophy are to a con-
siderable extent workmanlike. Quite certainly, no philosophy which is to be worthwhile
should lose the sense that there is something to be got right, that it is answerable to
argument and that it is in the business of telling the truth.
These things, I believe, are represented by the best of what is called analytical phi-
losophy, and to that extent I am committed to it. My own work has largely been in its
style. Yet, having now worked in it for a long time, and having, like Searle, seen it
change, I am a great deal more puzzled about it than I once was; in particular, I am
puzzled about the ways in which it must understand itself if it is to have those virtues,
and also about the costs of sustaining those virtues. There is one understanding of
these virtues which is certainly widespread among analytical philosophers, and which
directly serves the promise of ongoing research: that these are indeed the virtues of a
science. Some philosophers who are impressed by this conception of what they are
doing ritualize it into the forms of presentation familiar from the sciences. Sometimes
this is mere scientism, but in other cases it signals the fact that their branch of phi-
losophy is near neighbour to some science, such as quantum mechanics or cognitive
science.
But the virtues of workmanlike truthfulness which analytical philosophy typically | Blackwell |
cultivates are much more important than any attempt to make philosophy look like
a science. With many other branches of philosophy there is no plausible version of
sharing a party wall with science, and yet these virtues are still regarded as virtues. In
fact, even in the case of the more scientific areas of philosophy, it is obvious enough
that these virtues are not recommended only because they are possessed by its scien-
tific neighbours: they are taken to be intellectual virtues, good in the same way for
philosophy and for science.
24
CONTEMPORARY PHILOSOPHY: A SECOND LOOK
But how far can philosophy cultivate just these virtues and remain true to other
aspects of the legacy of the subject, to aims that it has pursued in the past? The sci-
ences aim to make claims that can and should be conveyed in ways that are minimally
expressive; they are not meant to convey feeling, or to display much literary imagina-
tion, or to speak (at least overtly) in a persuasive mode. But if we think in particular of
moral or political philosophy, is this ambition actually true to the traditions of the
subject, even as those are embodied in the historical canon of analytical philosophy
itself? Is it true to a tradition that contains Plato, Hobbes, Hume, Kant and (come to
that) John Stuart Mill? Is it in fact true to any great figure of that tradition, except
perhaps Aristotle? And if we are to take him as our model, we are left with many ques-
tions to consider – whether, for instance, the affectless treatises that we possess do rep-
resent his voice; if so, whether the tone does not represent a quite special view of ethical
life; and whether we should not weigh rather soberly the fact that the closest previous
imitation of Aristotle was to be found in a movement called scholasticism.
Particularly in moral and political philosophy, but not only there, there is a question
of what the procedures typical of analytical philosophy mean. There are many
virtuous and valuable things that they make possible, and at the same time there are
resources of philosophy in the past that they seemingly exclude, and it is important not
to assume that this balance is simply given to us, above all by an unquestionable and
transparent interpretation of the ideals of intellectual responsibility. I do not want to
suggest that the adoption of the analytical style is a mere abdication, a cowardly refusal
to adopt a more imaginative and committed manner which (critics sometimes suggest)
is obviously to hand. Still less is it simply a matter of scientistic camouflage. It is a
feature of our time that the resources of philosophical writing typically available to
analytical philosophy should present themselves so strongly as the responsible way of
going on, the most convincing expression of a philosopher’s claim on people’s atten-
tion. But that is an historical fact, and we should try to understand it as such. I do not
think that I adequately understand it, and, for that reason, I would not like to predict
what other possibilities there may prove to be for a philosophy that preserves the merits
of analytical philosophy.
In the rest of this essay, I shall try to give an outline of some principal concerns of
analytical moral and political philosophy. This will supplement the account that John
Searle has given of the state of the art in other areas, but I hope also that in describing
some of what analytical philosophy has recently done for these subjects, it may en-
courage readers to ask what new things it might be able to do.
2 Meta-ethics
Philosophical studies have often been understood, in the analytical tradition but not
only there, as being higher-order, in the sense that natural science, for instance, will
study natural phenomena, while the philosophy of science will study, from some par-
ticular points of view, the operations of science. Some of moral philosophy (or, as I shall | Blackwell |
also call it, ethics) is certainly a higher-order study. It discusses such things as the
nature of moral judgements, and asks whether they express genuine beliefs, whether
they can be objectively true, and so forth. Such higher-order questions are the concern
25
BERNARD WILLIAMS
of meta-ethics. At one time (thirty to fifty years ago) it was widely thought in analyti-
cal philosophy that ethics consisted only of meta-ethics. A powerful source of this con-
ception was the belief in a firm distinction between fact and value, to which Searle has
referred. However, the idea of ethics as simply meta-ethics does not follow directly from
the distinction between fact and value, and those who used the distinction to support
that idea needed two further assumptions.
One was that philosophy itself should be in the business of ‘fact’ (which, for the pur-
poses of the distinction, included theory) and not of value. This was connected with a
certain conception of philosophy, important to the identity of analytical philosophy, in
which it is taken to derive its authority from its theoretical stance as an abstract intel-
lectual enquiry. Some earlier philosophers, such as G. E. Moore, had indeed believed in
the distinction between fact and value, but had supposed that philosophers, in one way
or another, could have quite a lot to say substantively about values. The journey from
the fact–value distinction to a view of ethics as only meta-ethics involved the assump-
tion that this was impossible or inappropriate.
The second assumption involved in the journey was that meta-ethics itself could be
value-neutral, that the study of the nature of ethical thought did not commit one to
any substantive moral conclusions. A yet further assumption, which was not necessary
to the journey but did often accompany it, was that meta-ethics should be linguistic in
style, and its subject should be ‘the language of morals’. This latter idea has now almost
entirely disappeared, as the purely linguistic conception of philosophical study has
more generally retreated. Beyond that, however, there are now more doubts about the
extent to which meta-ethics can be value-neutral, and, in addition, philosophers simply
feel freer in making their own ethical commitments clear. Meta-ethics remains a part
of ethics, but most writings in philosophical ethics now will declare substantive moral
positions, either in close association with some meta-ethical outlook or in a more
free-standing manner.
Recent meta-ethical discussions have carried on the traditional interest in the objec-
tivity of ethics. In this connection, ‘moral judgements’ are often grouped together and
compared with everyday statements of fact or with scientific statements. Some theories
claim that moral judgements lack some desirable property that factual statements can
attain, such as objectivity or truth, while other theories claim that they can have this
property. These debates, particularly those conducted under the influence of positivism,
have tended to assimilate two different issues. One concerns the prospects of rational
agreement in ethical matters. The other concerns the semantic status of moral judge-
ments: whether they are, typically, statements at all (as opposed, for instance, to
prescriptions), and whether they aim at truth.
Objectivity is best understood in terms of the first kind of question. There clearly are
substantive and systematic disagreements about ethical questions, both between dif-
ferent societies or cultures, and within one society (particularly when, as now, the
culture of one society may be highly pluralist). Some of these disagreements may turn
out to be due to misunderstanding or bad interpretation and dissolve when local prac-
tices are better understood, but this is not true of all of them. Since ancient times it has
been suggested that these disagreements have a status different from disagreements
about facts or about the explanation of natural phenomena. With the latter, if the | Blackwell |
parties understand the question at issue, they see how after further enquiry they may
26
CONTEMPORARY PHILOSOPHY: A SECOND LOOK
end up in one of several positions: they may come to rational agreement on one answer
or another, they may recognize that such evidence as they can obtain underdeter-
mines the answer and leaves them with intelligible room for continued disagreement,
or they may advance in understanding to a point at which they see that the question
in its original form cannot be answered, for instance because it was based on a false
presupposition.
By contrast, it is suggested that we can understand an ethical dispute perfectly well,
and yet it be clear that it need not come out in any of these ways. Disagreeing about
an ethical matter, the parties may radically disagree about the kinds of considerations
that would settle the question, and the suggestion is that at the end of the line there
may be no rational way of arriving at agreement. This is the suggestion that ethical
claims lack objectivity. Some theories have associated this position with a view about
the semantic status of moral utterances. Emotivism, a theory closely associated with
positivism, held that moral utterances were merely expressions of emotion, not far
removed from expletives, and it took this, reasonably enough, not to be an objectivist
theory. In this case, the semantic account and the denial of objectivity went closely
together. However, it is a mistake to think that the two issues are in general simply
related to one another.
A clear illustration of this is Kant’s theory. Kant supposed that moral statements, or
at any rate the most basic of them, were actually prescriptions, and he understood the
fundamental principle of morality to be an imperative. However, when the issue is
expressed in terms of rational agreement or disagreement, Kant is quite certainly an
objectivist: the Categorical Imperative, together in some cases with empirical informa-
tion, determines for any rational agent what morality requires, and all rational agents
are in a position to agree on it. Another example of objectivity which is at least non-
committal about the semantics involved comes from virtue theory. Aristotle believed
that experienced and discriminating agents who had been properly brought up would
reach rational agreement in action, feeling, judgement and interpretation. He believed,
moreover, that this possibility was grounded in the best development and expression of
human nature, and that views about what counted as the best human development
could themselves command rational agreement. This certainly offers a kind of objec-
tivity, but it does not particularly emphasize agreement in belief; no doubt some agree-
ment in belief will matter, but so equally will agreement in feeling and in practical
decision.
However, even if objectivity need not imply rational agreement in belief, it may be
argued that the converse holds: that a theory which represents moral judgements as
basically expressing beliefs must be committed to objectivity. Beliefs, this argument
goes, are true or false. If moral judgements express beliefs, then some of them are true
and there is such a thing as truth in morality. So if people disagree about what to
believe, someone must be wrong. This certainly sounds as though there must be objec-
tivity. The difficulty with this argument is that it seems to be too easy to agree that moral
judgements admit of truth or falsehood. They are certainly called ‘true’ and ‘false’, as
even the emotivists had to concede, and the claim that nevertheless they are not really
true or false needs some deciphering. Emotivism itself offered a semantic analysis in
terms of which such judgements turn out not really to be statements, which certainly
gives some content to the claim that they are not really true or false. However, such
27
BERNARD WILLIAMS
| Blackwell |
analyses run into difficulty precisely because the air of being a statement that sur-
rounds moral judgements is not merely superficial – they behave syntactically just as
other kinds of statements do.
An alternative is to argue that moral judgements can indeed be true or false, but that
nothing interesting follows from this. On some theories of truth, sometimes called
‘redundancy’ theories, to claim that ‘P’ is true is to do no more than to assert that P.
Any theory of truth must accept the equivalence ‘‘‘P’’ is true if and only if P’; the pecu-
liarity of redundancy theories is to claim that this is all there is to the nature of truth.
If this is correct, then the truth or falsehood of moral judgements will follow simply
from their taking a statemental form, which allows them to be asserted or denied. Objec-
tivity will then either be understood as something that necessarily goes along with truth
and falsehood, in which case, on the redundancy theory, it will be no more interesting
or substantive than truth; or it will be more interesting and substantive – implying for
instance the possibility of rational agreement – in which case, on the redundancy view,
it will not follow just from the fact that moral judgements can be true or false.
It is widely, though not universally, agreed that an adequate theory of truth needs
to go beyond the redundancy view, but it is disputed how far it needs to go. Some argue
that if one takes seriously the claim that a given proposition is true, then this does imply
the idea that there could be convergence in belief on the proposition under favourable
circumstances. This approach brings the idea of truth itself nearer to that of objectiv-
ity as that has been introduced here. Others hold that a properly ‘minimalist’ theory of
truth need not bring in such a strong condition.
If objectivism and the mere truth of moral statements have often been assimilated
to one another, realism, equally, is often assimilated to one or both of them. Yet we
should expect realism, if it is an issue at all, to be a further issue. Elsewhere in phi-
losophy, for instance in the philosophy of mathematics or the philosophy of science, it
can be agreed that statements of a certain kind (about numbers, or about subatomic
particles) are capable of truth, and also that they can command rational agreement,
and yet it is thought by many philosophers that this does not answer the question
whether those statements should be interpreted realistically, where this means (very
roughly indeed) that the statements are taken to refer to objects that exist independently
of our thoughts about them. Even if it is not easy to give a determinate sense to such
questions, at any rate one would not expect realism to follow trivially from the claim
that moral statements can be true or false.
Some philosophers, influenced by the late John Mackie and in the line of Hume, deny
realism by claiming that the moral properties of people, actions and so forth, are not
‘in the world’ but are ‘projected’ on to it from our feelings and reactions. According to
the most familiar version of this view, secondary qualities such as colours are also pro-
jected on to the world, and this raises the question whether the metaphor has not mis-
located the most significant issues about moral properties. The theory implies that
ethical outlooks are ‘perspectival’ or related to human experience in ways in which
physical theory (at least) is not, but this does not take us very far: it will not tell us any-
thing very distinctive about ethical realism to know only that ethical concepts are per-
spectival in a sense in which colour concepts, or perhaps psychological concepts, are
also perspectival. An anti-realism that gives moral properties much the same status as
28
CONTEMPORARY PHILOSOPHY: A SECOND LOOK
colours will probably satisfy many moral realists. We need to ask how far the moral
concepts and outlooks of various human groups can intelligibly differ while the rest of | Blackwell |
their ways of describing the world, in particular their psychological concepts for
describing people’s behaviour, remain the same. Again, how far can their psychologi-
cal concepts themselves intelligibly vary, and how should we understand those
variations?
In considering such questions, it is helpful to abandon a very limiting assumption
which has been made up to this point in the discussion, namely, that all ‘moral judge-
ments’ are essentially of the same kind and stand in the same relation to such matters
as truth and objectivity. In considering moral disagreement, philosophers have con-
centrated on cases in which the parties express themselves in terms of the ‘thin’ ethical
concepts such as ‘good’, ‘right’ and ‘ought’. The parties share the same moral and other
concepts, and disagree about whether a given judgement should be asserted or denied:
they disagree, for instance, about whether capital punishment is wrong. To represent
disagreement in this way may seem to isolate in a helpful way its moral focus. But a lot
of moral discussion – to differing degrees in different societies – is conducted in
terms of ‘thick’ concepts, such as ‘brutality’ or ‘betrayal’ or ‘sentimentality’, and it is a
mistake to suppose that such concepts are merely convenient devices for associating a
bunch of empirical considerations with a thin ethical concept. It has been increasingly
accepted in recent discussions that the application of such concepts is guided by their
evaluative point, and that one cannot understand them without grasping that point.
(This does not mean that anyone who understands such a concept must have adopted
it as his or her own, but it does mean that he or she needs to have imaginatively iden-
tified, as an ethnographer does, with those who use it.) At the same time, however, such
concepts apply to some empirical states of affairs and not to others, and there is room
for truth, objectivity and knowledge to be displayed in their application.
If this is correct, then it may be more helpful to consider ethical disagreement, not
at the ultimate cutting edge of the practical judgements about what ought or ought not
to happen, but further back, in the network of more substantive and thicker concepts
that back up such judgements. Such concepts will typically serve more purposes than
expressing bare ‘moral judgements’. They may play a role, for instance, in a scheme of
psychological explanation. The question will then become, rather, why and to what
extent different cultures differ in their ethical concepts, and, more broadly, in the frame-
works of understanding that go with such concepts. Seen in this light, meta-ethical
questions move further away from being questions in the philosophy of language or the
theory of justification or epistemology, and become more like questions in the theory
of cultural understanding. Indeed they may become directly questions of cultural
understanding. The most basic question about objectivity may turn out to be the ques-
tion of the extent to which different human societies share an underlying determinate
framework of ethical concepts. By turning in such a direction, philosophical discussion
becomes more empirical and historical, more richly related to other disciplines, and
more illuminating. At the same time, it means that philosophers have to know about
more things, or people in other disciplines have to take on issues in philosophy. To that
extent, philosophy tends to lose a distinctive subject matter and its identity becomes
blurred.
29
BERNARD WILLIAMS
3 Ethical Theory
I have already said that analytical philosophers are happier than they once were to
recognize that what they say in moral philosophy is likely to have a substantive ethical
content, and that even meta-ethics is likely to have some such consequences. There is
a problem, however, about how this is related to the authority of philosophy. If philoso-
phers are going to offer moral opinions – within their subject, that is to say, and not | Blackwell |
simply as anyone offers moral opinions – they need to have some professional claim to
attention. They are not, as philosophers, necessarily gifted with unusual insight or
imagination, and they may not have a significantly wide experience or knowledge of
the world. Their claim to attention rests on their capacity for drawing distinctions,
seeing the consequences of assumptions, and so forth: in general, on their ability to
develop and control a theoretical structure. If the authority of philosophy lies in its
status as a theoretical subject, the philosopher’s special contribution to substantive
ethical issues is likely to be found in a theoretical approach to them. One of the most
common enterprises in moral philosophy at present is the development of various
ethical theories.
The aim of ethical theory is to cast the content of an ethical outlook into a theoreti-
cal form. An ethical theory must contain some meta-ethics, since it takes one view
rather than another of what the structure and the central concepts of ethical thought
must be, though it need not have an opinion on every meta-ethical issue. It is commit-
ted to putting forward in a theoretical form a substantive ethical outlook. In doing this,
ethical theories are to different degrees revisionary. Some start with a supposedly unde-
niable basis for ethics, and reject everyday moral conclusions that conflict with it. (It is
a good question why the basis should be regarded as undeniable if it has such con-
sequences.) Others, less dogmatically, consider the moral conclusions that would be
delivered by conflicting outlooks, and decide which outlook makes the most coherent
systematic sense of those conclusions that we (that is to say, the author and those
readers who agree with him or her) find most convincing. Unsystematized but carefully
considered judgements about what we would think it right to do in a certain situation,
or would be prepared to say in approval or criticism of people and their actions, are
often called in the context of such a method ‘moral intuitions’. (The term ‘intuition’
has a purely methodological force: it means only that these judgements seem to us, after
consideration, pre-theoretically convincing, not that they are derived from a faculty of
intuiting moral truths.) A preferred method is to seek what John Rawls has called a
‘reflective equilibrium’ between theory and intuitions, modifying the theory to accom-
modate robust intuitions, and discarding some intuitions which clash with the theory,
particularly if one can see how they might be the product of prejudice or confusion.
Moral theories are standardly presented as falling into three basic types, centring
respectively on consequences, rights and virtues. The first are unsurprisingly called
‘consequentialist’, and the last ‘virtue theories’. The second are often called ‘deonto-
logical’, which means that they are centred on duty or obligation, but this is a cross-
classification, since consequentialist theories also give prominence to an obligation,
that of bringing about the best consequences. (In the case of the most familiar and
the consequences is
popular consequentialist theory, utilitarianism, the value of
30
CONTEMPORARY PHILOSOPHY: A SECOND LOOK
expressed in terms of welfare or preference-satisfaction.) In terms of obligations, the
difference is rather that pure consequentialist theories present only one basic obliga-
tion, while the second type of theory has many. A more distinctive mark of difference
is to be found in the idea of a right: the second type of theory grounds many of an
agent’s obligations in the rights of others to expect certain behaviour from that agent,
a kind of consideration that utilitarians and other consequentialists regard as being at
best derivative, and at worst totally spurious.
Another way of understanding the division into three is in terms of what each | Blackwell |
theory sees as most basically bearing ethical value. For the first type of theory, it is good
states of affairs, and right action is understood as action tending to bring about good
states of affairs. For the second type, it is right action; sometimes what makes an action
right is a fact about its consequences, but often it is not – its rightness is determined
rather by respect for others’ rights, or by other obligations that the agent may have.
Virtue theory, finally, puts most emphasis on the idea of a good person, someone who
could be described also as an ethically admirable person. This is an important empha-
sis, and the notion of a virtue is important in ethics. However, once the types of theory
are distinguished in this way, it is hard to see them as all in the same line of business.
Consequentialist and rights theories aim to systematize our principles or rules of action
in ways that will, supposedly, help us to see what to do or recommend in particular
cases. A theory of the virtues can hardly do that: the theory itself, after all, is going to
say that what you basically need in order to do and recommend the right things are
virtues, not a theory about virtues. Moreover, virtuous people do not think always, or
usually, about the virtues. They think about such things as good consequences or
people’s rights, and this makes it clear that ‘virtue theory’ cannot be on the same level
as the other two types of theory.
4 Morality, Politics and Analytical Philosophy
Among moral concepts, that of rights is closest to law and also to politics, and phi-
losophical discussions of them often cross those boundaries. Given these relations, it is
not surprising that the kind of theory most often constructed to articulate the idea of
moral rights is contractualist, invoking the idea of an agreement that might be ratio-
nally arrived at by parties in some hypothetical situation in which they were required
to make rules by which they could live together. The inspiration of contractualist
theories goes back, in particular, to Kant. Kant’s own construction relies on some ideas
that are not shared by many modern theorists, in particular that a commitment to the
basic principle of morality (the so-called ‘Categorical Imperative’) is presupposed by the
very activity of a rational agent. It also involves a very obscure doctrine of freedom.
The modern theories inspired by Kantian ideas are less committed than Kant was to
showing that morality is ultimate rationality, and they allow also more empirical
material into the construction than Kant did.
The leading example of such a theory is that of John Rawls. His model of a set of
contracting parties reaching an agreement behind ‘a veil of ignorance’ has had an
immense influence on thinking about morality. It was designed for a purpose in politi-
cal philosophy, of constructing a theory of social justice. In Rawls’s theory, the veil of
31
BERNARD WILLIAMS
ignorance is introduced to disguise from each contractor his own particular advantages
and disadvantages and his own eventual position in the society that is being designed.
The political theory that uses this thought experiment is liberal, giving a high priority
to liberty and at the same time emphasizing redistribution of resources in the interests
of the disadvantaged. It is significant that when Rawls first produced his theory he saw
it in universalist terms, as offering a construction of social justice which would apply
to any society that met the conditions (very roughly) of being able to think about social
ideals and having the resources to implement them. Now, however, he has moved in
the direction of seeing the theory of justice as one that expresses the aspirations of a
particular social formation, the modern pluralist state.
Much recent political philosophy has centred on this liberal project, of defining terms
of just coexistence for people living in a pluralist society. One interpretation of that aim | Blackwell |
is to look for terms of coexistence that will not presuppose a common conception of the
good. On such an account, citizens can understand themselves as sharing a social exis-
tence although they have as individuals, or as members of communities less extensive
than the state, varying conceptions of a good life. Rawls has interpreted his own
purpose in those terms; it is expressed, for instance, in the fact that the parties in the
original position were supposed to make their decision on (broadly) self-interested
grounds, and not in the light of any antecedent conception of what a good society
would be. The values that they were taken to have were expressed only in the list of
‘primary goods’ in terms of which they made their choice, a list which made it clear
already that they set a high value on liberty, for instance, and did not assess everything
in terms of utility.
In fact, Rawls’s aim of making his theory as independent as possible from substan-
tial claims about the good does put it under some strain. At least in the first version of
the theory, the basic conception of justice included a large-scale commitment to eco-
nomic redistribution, and while this made it very welcome to many liberals (particu-
larly in the American sense of that word), it laid it open to the criticism from others
more inclined to libertarianism, such as Robert Nozick, that the theory incorporated
not only rightful terms of coexistence but a substantial and distinctive conception of
social justice as economic equality. In the later versions of Rawls’s theory, this concep-
tion is less prominent, but even more weight than before is laid on the idea that coex-
istence in a liberal pluralist society is not ‘a mere modus vivendi’ but a condition that
calls on important moral powers of toleration and respect for autonomy. This empha-
sis does seem to express a distinctive conception of the good, of a Kantian kind, to a
point at which it looks as though the condition of pluralism is not simply a contingency
of modern life, but an important vehicle of ethical self-expression.
Others, such as Ronald Dworkin, have pursued liberal theory, and in pluralist terms,
while accepting a commitment to a distinctive conception of the good. Others, again,
have claimed to reject liberal pluralism altogether and have turned in what is some-
times called a ‘communitarian’ direction. It is hard to tell in some cases where writers
of this tendency stand in relation to the politics of liberalism. While a few, notably
Alasdair MacIntyre, despairing of the whole enterprise, try only to diagnose our
condition and store some ethical goods for a better time, others seem to share in the
liberal undertaking but prefer, in opposition to Rawls’s Kantianism, a more Hegelian
type of discourse to express it.
32
CONTEMPORARY PHILOSOPHY: A SECOND LOOK
It is a rather odd feature of communitarian theories, at least if they take a tradi-
tionalist turn, that they recommend a politics that does not sit very easily with the exis-
tence of such theories themselves, except perhaps as a kind of interim measure. They
seek a politics in which people’s relations are formed by shared understandings which
to a considerable extent must be unspoken and taken for granted, and the exchange of
abstract political theories plays no obvious role in this conception of social life. (Hegel,
many of whose concerns are re-enacted in these debates, thought he had an answer to
this, in his conception of a society that could ultimately reconcile abstract under-
standing and concrete practice, but few current disputants are happy to pay the price
of admission to the Hegelian system.)
The liberals, on the other hand, have a conception of modern political life which at
least coherently embraces the existence of their own theories, since they understand
the modern state as a formation in which authority is peculiarly vested in discursive
argument, rather than in traditional or charismatic leadership. It is true that many | Blackwell |
political philosophers in the analytical tradition (unlike Rawls himself, and also unlike
Habermas, who comes from a sociological tradition) do not see the role of their theo-
ries in these terms, but rather as advancing trans-historical views about the demands
of a just political order. But even if liberals do not always recognize the point them-
selves, the role of theory in liberal political philosophy can be given a special justifica-
tion in terms of current political reality. Liberals coherently believe that the project
of political theory makes sense, because they are committed to thinking that in our
circumstances it makes sense to engage in a political activity of explaining the basic
principles of democratic government in such terms.
In this respect, moral philosophy is in an altogether worse situation. It typically lacks
an account of why the project of articulating moral theories makes any sense at all. As
many writers have pointed out, it bears little relation to the psychology of people’s
ethical lives, and inasmuch as it claims that turning morality into a theory makes it
more rational, there is a pressing question of what concept of rationality is being
invoked. To a limited extent, there may be an answer to that question, inasmuch as
some ethical questions, such as those raised by medical ethics, are public questions,
closely tied to politics and the law. In those cases, we need a public discourse to legiti-
mate some answers, since it is a public issue what should be permitted, and mere
appeals to ethical or professional authority will no longer do. But it would be a mistake
to suppose that in such cases we are presented with a pure concept of moral rational-
ity which we then apply to our historical circumstances. Rather, what counts as a ratio-
nal way of discussing such questions is influenced by the historical circumstances, and
above all by the need to give a discursive justification, in something like a legal style, for
procedures which increasingly are adopted in a public domain and can be challenged
in it.
Moreover, many important ethical issues are not of this kind at all. Morality has
always been connected not only with law and politics, but also with the meaning of an
individual’s life and his or her relations to other people. In these connections, the
authority of theory over the moral life remains quite opaque. Certainly, it will not do to
rely on the inference: philosophy must have something to say about the moral life;
the most responsible form of philosophy is analytical philosophy; what analytical phi-
losophy is best at is theory; so philosophy’s contribution to the moral life is theory.
33
BERNARD WILLIAMS
In rejecting this uninviting argument, some will want to attack the second premise.
However, it may be more constructive, and offer more of a challenge to thinking about
what one wants of philosophy, if one reconsiders the third premise, that what analyti-
cal philosophy does best is theory. Analytical philosophy’s own virtues, such as its
unfanatical truthfulness, could encourage it in ethics to remind us of detail rather than
bludgeon us with theory.
Truthfulness in personal life, and even in politics, is not necessarily opposed to the
exercise of the imagination. It is relevant here that an imaginative truthfulness is a
virtue in the arts. Writers in moral philosophy sometimes urge us to extend our ethical
understanding by turning to imaginative literature, in particular the novel. To the
extent that this is good advice, it is not because novels are convenient sources of psy-
chological information, still less because some of them are morally edifying. It is
because imaginative writing can powerfully evoke the strength of ethical considera-
tions in giving sense to someone’s life or to a passage of it, and, equally, present the pos- | Blackwell |
sible complexity, ambivalence and ultimate insecurity of those considerations. Good
literature stands against the isolation of moral considerations from the psychological
and social forces that both empower and threaten them. But this isolation of moral
considerations from the rest of experience is an illusion very much fostered by moral
philosophy itself; indeed, without that illusion some forms of moral philosophy could
not exist at all. So there are lessons here not just for philosophy’s use of other writing,
but for philosophical writing itself. The truthfulness that it properly seeks involves
imaginative honesty and not just argumentative accuracy.
Analytical philosophy, or some recognizable descendant of it, should be able to make
a richer contribution to ethics than has often been the case up to now. If it is to do so,
it will need to hold on to two truths which it tends to forget (not only in ethics, but most
damagingly there): that philosophy cannot be too pure, and must merge with other
kinds of understanding; and that being soberly truthful does not exclude, but may
actually demand, the imagination.
34
PART I
AREAS OF PHILOSOPHY
1
Epistemology
A. C. G R AY L I N G
For most of the modern period of philosophy, from Descartes to the present, epistemol-
ogy has been the central philosophical discipline. It raises questions about the scope and
limits of knowledge, its sources and justification, and it deals with sceptical arguments
concerning our claims to knowledge and justified belief. This chapter firstly considers dif-
ficulties facing attempts to define knowledge and, secondly, explores influential responses
to the challenge of scepticism. Epistemology is closely related to METAPHYSICS (chapter
2), which is the philosophical account of what kinds of entities there are. Epistemologi-
cal questions are also crucial to most of the other areas of philosophy examined in this
volume, from ETHICS (chapter 6) to PHILOSOPHY OF SCIENCE (chapter 9) and PHILOSO-
PHY OF MATHEMATICS (chapter 11) to PHILOSOPHY OF HISTORY (chapter 14). Chapters
on individuals or groups of philosophers from DESCARTES (see chapter 26) to KANT
(chapter 32) discuss classical epistemology, while several chapters about more recent
philosophers also follow epistemological themes.
Introduction
Epistemology, which is also called the theory of knowledge, is the branch of philosophy
concerned with enquiry into the nature, sources and validity of knowledge. Among the
chief questions it attempts to answer are: What is knowledge? How do we get it? Can
our means of getting it be defended against sceptical challenge? These questions are
implicitly as old as philosophy, although their first explicit treatment is to be found in
PLATO (c.427–347 BC) (see chapter 23), in particular in his Theaetetus. But it is primar-
ily in the modern era, from the seventeenth century onwards – as a result of the work
of DESCARTES (1596–1650) (chapter 26) and LOCKE (1632–1704) (chapter 29) in asso-
ciation with the rise of modern science – that epistemology has occupied centre-stage
in philosophy.
One obvious step towards answering epistemology’s first question is to attempt a
definition. The standard preliminary definition has it that knowledge is justified true
belief. This definition looks plausible because, at the very least, it seems that to know
something one must believe it, that the belief must be true, and that one’s reason for
A. C. GRAYLING
believing it must be satisfactory in the light of some criterion – for one could not be said
to know something if one’s reasons for believing it were arbitrary or haphazard. So each
of the three parts of the definition appears to express a necessary condition for knowl-
edge, and the claim is that, taken together, they are sufficient.
There are, however, serious difficulties with this idea, particularly about the nature
of the justification required for true belief to amount to knowledge. Competing pro- | Blackwell |
posals have been offered to meet the difficulties, either by adding further conditions or
by finding a better statement of the definition as it stands. The first part of the follow-
ing discussion considers these proposals.
In parallel with the debate about how to define knowledge is another about how
knowledge is acquired. In the history of epistemology there have been two chief schools
of thought about what constitutes the chief means to knowledge. One is the ‘rational-
ist’ school (see chapters 26 and 27), which holds that reason plays this role. The other
is the ‘empiricist’ (see chapters 29, 30 and 31), which holds that it is experience, prin-
cipally the use of the senses aided when necessary by scientific instruments, which does
so.
The paradigm of knowledge for rationalists is mathematics and logic, where neces-
sary truths are arrived at by intuition and rational inference. Questions about the
nature of reason, the justification of inference, and the nature of truth, especially nec-
essary truth, accordingly press to be answered.
The empiricists’ paradigm is natural science, where observation and experiment are
crucial to enquiry. The history of science in the modern era lends support to empiri-
cism’s case; but precisely for that reason philosophical questions about perception,
observation, evidence and experiment have acquired great importance.
But for both traditions in epistemology the central concern is whether we can trust
the routes to knowledge they respectively nominate. Sceptical arguments suggest that
we cannot simply assume them to be trustworthy; indeed, they suggest that work is
required to show that they are. The effort to respond to scepticism therefore provides a
sharp way of understanding what is crucial in epistemology. Section 2 below is accord-
ingly concerned with an analysis of scepticism and some responses to it.
There are other debates in epistemology about, among other things, memory, judge-
ment, introspection, reasoning, the ‘a priori–a posteriori’ distinction, scientific method
and the methodological differences, if any, between the natural and the social sciences;
however, the questions considered here are basic to them all.
1 Knowledge
1.1 Defining knowledge
There are different ways in which one might be said to have knowledge. One can know
people or places, in the sense of being acquainted with them. That is what is meant
when one says, ‘My father knew Lloyd George’. One can know how to do something,
in the sense of having an ability or skill. That is what is meant when one says, ‘I know
how to play chess’. And one can know that something is the case, as when one says, ‘I
know that Everest is the highest mountain’. This last is sometimes called ‘propositional
knowledge’, and it is the kind epistemologists most wish to understand.
38
EPISTEMOLOGY
The definition of knowledge already mentioned – knowledge as justified true belief
– is intended to be an analysis of knowledge in the propositional sense. The definition
is arrived at by asking what conditions have to be satisfied if we are correctly to describe
someone as knowing something. In giving the definition we state what we hope are the
necessary and sufficient conditions for the truth of the claim ‘S knows that p’, where
‘S’ is the epistemic subject – the putative knower – and ‘p’ a proposition.
The definition carries an air of plausibility, at least as applied to empirical knowl-
edge, because it seems to meet the minimum we can be expected to need from so con-
sequential a concept. It seems right to expect that if S knows that p, then p must at least
be true. It seems right to expect that S must not merely wonder whether or hope that
p is the case, but must have a positive epistemic attitude to it: S must believe that it is
true. And if S believes some true proposition while having no grounds, or incorrect
grounds, or merely arbitrary or fanciful grounds, for doing so, we would not say that S
knows p; which means that S must have grounds for believing p which in some sense | Blackwell |
properly justify doing so.
Of these proposed conditions for knowledge, it is the third that gives most trouble.
The reason is simply illustrated by counter-examples. These take the form of cases in
which S believes a true proposition for what are in fact the wrong reasons, although
they are from his or her own point of view persuasive. For instance, suppose S has
two friends, T and U. The latter is travelling abroad, but S has no idea where. As for T,
S saw him buying and thereafter driving about in a Rolls Royce, and therefore believes
that he owns one. Now, from any proposition p one can validly infer the disjunction
‘p or q’. So S has grounds for believing ‘T owns a Rolls or U is in Paris’, even though, ex
hypothesi, he has no idea of U’s location. But suppose T in fact does not own the
Rolls – he bought it for someone else, on whose behalf he also drives it. And further
suppose that U is indeed, by chance, in Paris. Then S believes, with justification, a true
proposition: but we should not want to call his belief knowledge.
Examples like this are strained, but they do their work; they show that more
needs to be said about justification before we can claim to have an adequate account of
knowledge.
1.2 Justification
A preliminary question concerns whether having justification for believing some p
entails p’s truth, for, if so, counter-examples of the kind just mentioned get no purchase
and we need not seek ways of blocking them. There is indeed a view, called ‘infallibil-
ism’, which offers just such a resource. It states that if it is true that S knows p, then S
cannot be mistaken in believing p, and therefore his justification for believing p guar-
antees its truth. The claim is, in short, that one cannot be justified in believing a false
proposition.
This view is rejected by ‘fallibilists’, who claim that one can indeed have justification
for believing some p although it is false. Their counter to infallibilism turns on identi-
fying a mistake in its supporting argument. The mistake is that whereas the truth of ‘S
knows that p’ indeed rules out the possibility that S is in error, this is far from saying
that S is so placed that he cannot possibly be wrong about p. It is right to say: (1) ‘it is
impossible for S to be wrong about p if he knows p’, but it is not invariably right to say
39
A. C. GRAYLING
(2) ‘if S knows p, then it is impossible for him to be wrong about p’. The mistake turns
on thinking that the correct wide scope reading (1) of ‘it is impossible’ licenses the
narrow scope reading (2) which constitutes infallibilism.
An infallibilist account makes the definition of knowledge look simple: S knows p if
his belief in it is infallibly justified. But this definition renders the notion of knowledge
too restrictive, for it says that S can justifiably believe p only when the possibility of p’s
falsity is excluded. Yet it appears to be a commonplace of epistemic experience that
one can have the very best evidence for believing something and yet be wrong (as the
account of scepticism given below is at pains to show), which is to say that fallibilism
seems the only account of justification adequate to the facts of epistemic life. We need
justification can give us an adequate
therefore to see whether fallibilist theories of
account of knowledge.
The problem for fallibilist accounts is precisely the one illustrated by the Rolls Royce
example above, and others similar to it (so-called ‘Gettier examples’, introduced in
Gettier 1963), namely, that one’s justification for believing p does not connect with the
truth of p in the right way, and perhaps not at all. What is required is an account that
will suitably connect S’s justification both with his belief that p and with p’s truth.
What is needed is a clear picture of ‘justified belief ’. If one can identify what justi-
fies a belief, one has gone all or most of the way to saying what justification is; and en
route one will have displayed the right connection between justification, on the one | Blackwell |
hand, and belief and truth on the other. In this connection there are several standard
species of theory.
Foundationalism
One class of theories of justification employs the metaphor of an edifice. Most of our ordi-
nary beliefs require support from others; we justify a given belief by appealing to another
or others on which it rests. But if the chain of justifying beliefs were to regress without
terminating in a belief that is in some way independently secure, thereby providing a
foundation for the others, we would seem to lack justification for any belief in the chain.
It appears necessary therefore that there should be beliefs which do not need justifica-
tion, or which are in some way self-justifying, to serve as an epistemic underpinning.
On this view a justified belief is one which either is, or is supported by, a founda-
tional belief. The next steps therefore are to make clear the notion of a ‘foundation’ and
to explain how foundational beliefs ‘support’ non-foundational ones. Some way of
understanding foundationalism without reliance on constructional metaphors is
needed.
It is not enough barely to state that a foundational belief is a belief that requires no
justification, for there must be a reason why this is the case. What makes a belief inde-
pendent or self-standing in the required way? It is standardly claimed that such beliefs
justify themselves, or are self-evident, or are indefeasible or incorrigible. These are not
the same things. A belief might be self-justifying without being self-evident (it might take
hard work to see that it justifies itself). Indefeasibility means that no further evidence or
other, competing, beliefs, can render a given belief insecure. Yet this is a property that
the belief might have independently of whether or not it is self-justifying. And so on. But
what these characterizations are intended to convey is the idea that a certain immunity
from doubt, error or revision attaches to the beliefs in question.
40
EPISTEMOLOGY
It might even be unnecessary or mistaken to think that it is belief that provides the
foundations for the edifice of knowledge: some other state might do so. Perceptual states
have been offered as candidates, because they appear to be suitably incorrigible – if one
seems to see a red patch, say, then one cannot be wrong that one seems to see a red patch.
And it appears plausible to say that one’s belief that p needs no further justification or
foundation than that things appear to one as p describes them to be.
These suggestions bristle with difficulties. Examples of self-evident or self-justifying
beliefs tend to be drawn from logic and mathematics – they are of the ‘x is x’ or ‘one plus
one equals two’ variety, which critics are quick to point out give little help in grounding
contingent beliefs. Perceptual states likewise turn out to be unlikely candidates for foun-
dations, on the grounds that perception involves the application of beliefs which them-
selves stand in need of justification – among them beliefs about the nature of things and
the laws they obey. What is most robustly contested is the ‘myth of the given’, the idea
that there are firm, primitive and original data which experience supplies to our minds,
antecedent to and untainted by judgement, furnishing the wherewithal to secure the rest
of our beliefs.
There is a difficulty also about how justification is transmitted from foundational
beliefs to dependent beliefs. It is too strong a claim to say that the latter are deducible from
them. Most if not all contingent beliefs are not entailed by the beliefs that support them;
the evidence I have that I am now sitting at my desk is about as strong as empirical
evidence can be, yet given the standard sceptical considerations (such as, for example,
the possibility that I am now dreaming) it does not entail that I am sitting here.
If the relation is not a deductive one, what is it? Other candidate relations – inductive | Blackwell |
or criterial – are by their nature defeasible, and therefore, unless somehow supplemented,
insufficient to the task of transmitting justification from the foundations to other beliefs.
The supplementation would have to consist of guarantees that the circumstances that
defeat non-deductive justification do not in fact obtain. But if such guarantees – under-
stood, to avoid circularity, as not being part of the putative foundations themselves – were
available to protect non-deductive grounds, then appeal to a notion of foundations looks
simply otiose.
1.3 Coherence
Dissatisfaction with foundationalism has led some epistemologists to prefer saying that
a belief is justified if it coheres with those in an already accepted set. The immediate
task is to specify what coherence is, and to find a way of dealing in a non-circular way
with the problem of how the already accepted beliefs came to be so.
Hard on the heels of this task comes a number of questions. Is coherence a negative
criterion (that is, a belief lacks justification if it fails to cohere with the set) or a posi-
tive one (that is, a belief is justified when it coheres with the set)? And is it to be under-
stood strongly (by which coherence is sufficient for justification) or weakly (by which
coherence is one among other justifying features)?
The concept of coherence has its theoretical basis in the notion of a system, under-
stood as a set whose elements stand in mutual relations of both consistency and (some
kind of) interdependence. Consistency is of course a minimum requirement, and goes
without saying. Dependence is more difficult to specify suitably. It would be far too
41
A. C. GRAYLING
strong – for it would give rise to assertive redundancy – to require that dependence
means mutual entailment among beliefs (this is what some have required, citing
geometry as the closest example). A more diffuse notion has it that a set of beliefs is
coherent if any one of them follows from all the rest, and if no subset of them is
logically independent of the remainder. But this is vague, and anyway seems to require
that the set be known to be complete before one can judge whether a given belief
coheres with it.
A remedy might be to say that a belief coheres with an antecedent set if it can be
inferred from it, or from some significant subset within it, as being the best explanation
in the case. To this someone might object that not all justifications take the form of
explanations. An alternative might be to say that a belief is justified if it survives com-
parison with competitors for acceptance among the antecedent set. But here an objec-
tor might ask how this can be sufficient, since by itself this does not show why the belief
merits acceptance over equally cohering rivals. Indeed, any theory of justification has
to ensure as much for candidate beliefs, so there is nothing about the proposal that
distinctively supports the coherence theory. And these thoughts leave unexamined
the question of the ‘antecedent set’ and its justification, which cannot be a matter of
coherence, for with what is it to cohere in its turn?
1.4 Internalism and externalism
Both the foundationalist and coherence theories are sometimes described as ‘internal-
ist’ because they describe justification as consisting in internal relations among beliefs,
either – as in the former case – from a vertical relation of support between supposedly
basic beliefs and others dependent upon them, or – as in the latter – from the mutual
support of beliefs in an appropriately understood system.
Generally characterized, internalist theories assert or assume that a belief cannot be
justified for an epistemic subject S unless S has access to what provides the justification,
either in fact or in principle. These theories generally involve the stronger ‘in fact’
requirement because S’s being justified in believing p is standardly cashed in terms of
his having reasons for taking p to be true, where having reasons is to be understood in | Blackwell |
an occurrent sense.
Here an objection immediately suggests itself. Any S has only finite access to what
might justify or undermine his beliefs, and that access is confined to his particular view-
point. It seems that full justification for his beliefs would rarely be available, because his
experience would be restricted to what is nearby in space and time, and he would be
entitled to hold only those beliefs which his limited experience licensed.
A related objection is that internalism seems inconsistent with the fact that many
people appear to have knowledge despite not being sophisticated enough to recognize
that thus-and-so is a reason for believing p – that is the case, for example, with
children.
A more general objection still is that relations between beliefs, whether of the foun-
dationalist or coherence type, might obtain without the beliefs in question being true
of anything beyond themselves. One could imagine a coherent fairy tale, say, which in
no point corresponds to some external reality, but in which beliefs are justified never-
theless by their mutual relations.
42
This uneasy reflection prompts the thought that there should be a constraint on
theories of justification, in the form of a demand that there should be some suitable
connection between belief possession and external factors – that is, something other
than the beliefs and their mutual relations – which determines their epistemic value.
This accordingly prompts the idea of an alternative: externalism.
EPISTEMOLOGY
1.5 Reliability, causality and truth-tracking
Externalism is the view that what makes S justified in believing p might not be anything
to which S has cognitive access. It might be that the facts in the world are as S believes
them to be, and that indeed they caused S to believe them to be so by stimulating his
or her sensory receptors in the right kind of way. S need not be aware that this is how
his or her belief was formed. So S could be justified in believing p without it.
One main kind of externalist theory is reliabilism, the thesis – or cluster of theses –
having it that a belief is justified if it is reliably connected with the truth. According to
one influential variant, the connection in question is supplied by reliable belief-forming
processes, ones which have a high success rate in producing true beliefs. An example of
a reliable process might be normal perception in normal conditions.
Much apparent plausibility attaches to theories based on the notion of external
linkage, especially of causal linkage, between a belief and what it is about. An example
of such a theory is Alvin Goldman’s (1986) account of knowledge as ‘appropriately
caused true belief ’, where ‘appropriate causation’ takes a number of forms, sharing
the property that they are processes which are both ‘globally’ and ‘locally’ reliable – the
former meaning that the process has a high success rate in producing true beliefs,
the latter that the process would not have produced the belief in question in some ‘rel-
evant counterfactual situation’ where the belief is false. Goldman’s view is accordingly
a paradigm of a reliabilist theory.
An elegant second-cousin of this view is offered by Robert Nozick (1981). To the
conditions
(1)
p is true
and
(2) S believes p
Nozick adds
(3)
if p were not true, S would not believe p
and
(4)
if p were true, S would believe it.
Conditions (3) and (4) are intended to block Gettier-type counter-examples to the
justified true belief analysis by annexing S’s belief that p firmly to p’s truth. S’s belief
43
A. C. GRAYLING
that p is connected to the world (to the situation described by p) by a relation Nozick
calls ‘tracking’: S’s belief tracks the truth that p. He adds refinements in an attempt to
deflect the counter-examples that philosophers are always ingenious and fertile at
devising.
If these theories seem plausible it is because they accord with our pre-theoretical
views. But as one can readily see, there are plenty of things to object to in them, and a | Blackwell |
copious literature does so. Their most serious flaw, however, is that they are question-
begging. They do not address the question of how S is to be confident that a given belief
is justified; instead they help themselves to two weighty realist assumptions, one about
the domain over which belief ranges and the other about how the domain and S are
connected, so that they can assert that S is justified in believing a given p even if what
justifies him lies outside his own epistemic competence. Whatever else one thinks of
these suggestions, they do not enlighten S, and therefore do not engage the same
problem that internalist theories address.
But worst of all – so an austere critic might say – the large assumptions to which
these theories help themselves are precisely those that epistemology should be exam-
ining. Externalist and causal theories, in whatever guise and combination, are better
done by empirical psychology where the standard assumptions about the external
world and S’s connections with it are premised. Philosophy, surely, is where these
premises themselves come in for scrutiny.
1.6 Knowledge, belief and justification again
Consider this argument: ‘If anyone knows some p, then he or she can be certain that
p. But no one can be certain of anything. Therefore no one knows anything.’ This argu-
ment (advanced in this form by Unger 1975) is instructive. It repeats Descartes’s
mistake of thinking that the psychological state of feeling certain – which someone can
be in with respect to falsehoods, such as the fact that I can feel certain that Arkle will
win the Derby next week, and be wrong – is what we are seeking in epistemology. But
it also exemplifies the tendency in discussions of knowledge as such to make the defi-
nition of knowledge so highly restrictive that little or nothing passes muster. Should
one care if a suggested definition of knowledge is such that, as the argument just quoted
tells us, no one can know anything? Just so long as one has many well-justified beliefs
which work well in practice, can one not be quite content to know nothing? For my
part, I think one can.
This suggests that in so far as the points sketched in preceding paragraphs have inter-
est, it is in connection with the justification of beliefs and not the definition of knowledge
that they do so. Justification is an important matter, not least because in the areas of
application in epistemology where the really serious interest should lie – in questions
about the PHILOSOPHY OF SCIENCE (chapter 9), the PHILOSOPHY OF HISTORY (chapter 14)
or the concepts of evidence and proof in LAW (see chapter 13) – justification is the
crucial problem. That is where epistemologists should be getting down to work. By
comparison, efforts to define ‘knowledge’ are trivial and occupy too much effort in
epistemology. The disagreeable propensity of
the debate generated by Gettier’s
counter-examples – anticipated beautifully in Russell’s review of James (Russell 1910:
95) – to proceed on a chessboard of ‘-isms’, as exemplified above, is a symptom.
44
The general problem with justification is that the procedures we adopt, across all
walks of epistemic life, appear highly permeable to difficulties posed by scepticism. The
problem of justification is therefore in large part the problem of scepticism; which is
precisely why discussion of scepticism is central to epistemology.
EPISTEMOLOGY
2 Scepticism
Introduction
The study and employment of sceptical arguments might in one sense be said to define
epistemology. A chief epistemological aim is to determine how we can be sure that our
means to knowledge (here ‘knowledge’ does duty for ‘justified belief ’) are satisfactory.
A sharp way to show what is required is to look carefully at sceptical challenges to our
epistemic efforts, challenges which suggest ways in which they can go awry. If we are
able not just to identify but to meet these challenges, a primary epistemological aim will | Blackwell |
have been realized.
Scepticism is often described as the thesis that nothing is – or, more strongly, can be –
known. But this is a bad characterization, because if we know nothing, then we do not
know that we know nothing, and so the claim is trivially self-defeating. It is more telling
to characterize scepticism in the way just suggested. It is a challenge directed against
knowledge claims, with the form and nature of the challenge varying according to the
field of epistemic activity in question. In general, scepticism takes the form of a request
for the justification of those knowledge claims, together with a statement of the reasons
motivating that request. Standardly, the reasons are that certain considerations suggest
that the proposed justification might be insufficient. To conceive of scepticism like this is
to see it as being more philosophically troubling and important than if it is described as
a positive thesis asserting our ignorance or incapacity for knowledge.
2.1 Early scepticism
Some among the thinkers of antiquity – Pyrrho of Elis (c.360–c.270 BC) and his school,
and Plato’s successors in his Academy – expressed disappointment at the fact that cen-
turies of enquiry by their philosophical predecessors seemed to have borne little fruit
either in cosmology or ethics (this latter was broadly construed to include politics).
Their disappointment prompted them to sceptical views. The Pyrrhonians argued that
because enquiry is arduous and interminable, one should give up trying to judge what
is true and false or right and wrong; for only thus will we achieve peace of mind.
A less radical form of scepticism overtook Plato’s successors in the Academy. They
agreed with Pyrrho that certainty must elude us, but they tempered their view by
accepting that the practical demands of life must be met. They did not think it a work-
able option to ‘suspend judgement’ as Pyrrho recommended, and therefore argued that
we should accept those propositions or theories which are more PROBABLE (pp. 308–11)
than their competitors. The views of these thinkers, known as Academic sceptics, are
recorded in the work of Sextus Empiricus (c.150–c.225).
45
A. C. GRAYLING
In the later Renaissance – or, which is the same thing, in early modern times – with
religious certainties under attack and new ideas abroad, some of the sceptical argu-
ments of the Academics and Pyrrhonians acquired a special significance, notably as a
result of the use to which René Descartes put them in showing that they are powerful
tools for investigating the nature and sources of knowledge.
In Descartes’s day the same person could be both astronomer and astrologer, chemist
and alchemist, or physician and magician. It was hard to disentangle knowledge from
nonsense; it was even harder to disentangle those methods of enquiry which might
yield genuine knowledge from those that could only deepen ignorance. So there was an
urgent need for some sharp, clean epistemological theorizing. In his Meditations (1986)
Descartes accordingly identified epistemology as an essential preliminary to physics and
mathematics, and attempted to establish the grounds of certainty as a propaedeutic to
science. Descartes’s first step in that task was to adapt and apply some of the traditional
arguments of scepticism. (I shall comment on his use of scepticism again later.)
The Anatomy of Scepticism
Sceptical arguments exploit certain contingent facts about our ways of acquiring,
testing, remembering and reasoning about our beliefs. Any problem that infects the
acquisition and employment of beliefs about a given subject matter, and in particular
any problem that infects our confidence that we hold those beliefs justifiably, threatens
our hold on that subject matter.
The contingent facts in question relate to the nature of perception, the normal human
vulnerability to error and the existence of states of mind – for example, dreaming
and delusion – which can be subjectively indistinguishable from those that we | Blackwell |
End of preview. Expand
in Dataset Viewer.
README.md exists but content is empty.
- Downloads last month
- 77