click here
Popper has argued (I think successfully) that
a scientific idea can never be proven true, because because no
matter how many observations seem to agree with it, it may still
be wrong. On the other hand, a single contrary experiment can
prove a theory forever false.
The Problem of Induction (1953, 1974)
For a brief formulation of the problem of induction we can
turn to Born, who writes: '. . . no observation or
experiment, however extended, can give more than a finite number
of repetitions'; therefore, 'the statement of a law - B
depends on A - always transcends experience. Yet this kind of
statement is made everywhere and all the time, and sometimes from
scanty material. ' 1
In other words, the logical problem of induction arises from
(1) Hume's discovery (so well expressed by Born) that it is
impossible to justify a law by observation or experiment, since
it 'transcends experience'; (2) the fact that science
proposes and uses laws 'everywhere and all the time'.
(Like Hume, Born is struck by the 'scanty material', i.e.
the few observed instances upon which the law may be based.) To
this we have to add (3) the principle of empiricism which asserts
that in science only observation and experiment may decide upon
the acceptance or rejection of scientific statements, including
laws and theories.
These three principles, (1), (2), and (3), appear at first
sight to clash; and this apparent clash constitutes the logical
problem of induction.
Faced with this clash, Born gives up (3), the principle of
empiricism (as Kant and many others, including Bertrand Russell,
have done before him), in favour of what he calls a
'metaphysical principle'; a metaphysical principle which
he does not even attempt to formulate; which he vaguely describes
as a 'code or rule of craft'; and of which I have never
seen any formulation which even looked promising and was not
clearly untenable.
But in fact the principles (1) to (3) do not clash. We can
see this the moment we realize that the acceptance by science of
a law or of a theory is tentative only; which is to say that all
laws and theories are conjectures, or tentative hypotheses (a
position which I have sometimes called 'hypotheticism');
and that we may reject a law or theory on the basis of new
evidence, without necessarily discarding the old evidence which
originally led us to accept it. (I do not doubt that Born and
many others would agree that theories are accepted only
tentatively. But the widespread belief in induction shows that
the far-reaching implications of this view are rarely seen.)
The principle of empiricism (3) can be fully preserved, since
the fate of a theory, its acceptance or rejection, is decided by
observation and experiment - by the results of tests. So long as
a theory stands up to the severest tests we can design, it is
accepted; if it does not, it is rejected. But it is never
inferred, in any sense, from the empirical evidence. There is
neither a psychological nor a logical induction. Only the falsity
of the theory can be inferred from empirical evidence and this
inference is a purely deductive one.
Hume showed that it is not possible to infer a theory from
observation statements; but this does not affect the possibility
of refuting a theory by observation statements. The full
appreciation of this possibility makes the relation between
theories and observations perfectly clear.
This solves the problem of the alleged clash between the
principles (1), (2), and (3), and with it Hume's problem of
induction
Hume's problem of induction has almost always been badly
formulated by what may be called the philosophical tradition. I
will first give a few of these bad formulations, which I shall
call the traditional formulations of the problem of induction. I
shall replace them, however, by what I regard as better
formulations.
Typical examples of formulations of the problem of induction
that are both traditional and bad are the following.
What is the justification for the belief that the future will
resemble the past? What is the justification of so-called
inductive inferences?
By an inductive inference is here meant an inference from
repeatedly observed instances to some as yet unobserved
instances. It is of comparatively minor significance whether such
an inference from the observed to the unobserved is, from the
point of view of time, predictive or retrodictive; whether we
infer that the sun will rise tomorrow or that it did rise 100,000
years ago. Of course, from a pragmatic point of view, one might
say that it is the predictive type of inference which is the more
important. No doubt usually it is.
There are various other philosophers who also regard as
misconceived this traditional problem of induction. Some say that
it is misconceived because no justification is needed for
inductive inference; no more in fact than for deductive
inference. Inductive inference is inductively valid just as
deductive inference is deductively valid. I think it was
Professor Strawson who was the first to say this.
I am of a different opinion. I hold with Hume that there
simply is no such logical entity as an inductive inference; or,
that all so-called inductive inferences are logically invalid -
and even inductively invalid, to put it more sharply [see the end
of this selection]. We have many examples of deductively valid
inferences, and even some partial criteria of deductive validity;
but no example of an inductively valid inference exists. 2 And I
hold, incidentally, that this result can be found in Hume, even
though Hume, at the same time, and in sharp contrast to myself,
believed in the psychological power of induction; not as a valid
procedure, but as a procedure which animals and men successfully
make use of, as a matter of fact and of biological
necessity.
I take it as an important task to make clear, even at the
cost of some repetition, where I agree and where I disagree with
Hume.
I agree with Hume's opinion that induction is invalid and
in no sense justified. Consequently neither Hume nor I can accept
the traditional formulations which uncritically ask for the
justification of induction; such a request is uncritical because
it is blind to the possibility that induction is invalid in every
sense, and therefore unjustifiable.
I disagree with Hume's opinion (the opinion incidentally
of almost all philosophers) that induction is a fact and in any
case needed. I hold that neither animals nor men use any
procedure like induction, or any argument based on the repetition
of instances. The belief that we use induction is simply a
mistake. It is a kind of optical illusion.
What we do use is a method of trial and the examination of
error; however misleadingly this method may look like induction,
its logical structure, if we examine it closely, totally differs
from that of induction. Moreover, it is a method which does not
give rise to any of the difficulties connected with the problem
of induction.
Thus it is not because induction can manage without
justification that I am opposed to the traditional problem; on
the contrary, it would urgently need justification. But the need
cannot be satisfied. Induction simply does not exist, and the
opposite view is a straightforward mistake.
There are many ways to present my own non-inductivist point
of view. Perhaps the simplest is this. I will try to show that
the whole apparatus of induction becomes unnecessary once we
admit the general fallibility of human knowledge or, as I like to
call it, the conjectural character of human knowledge.
Let me point this out first for the best kind of human
knowledge we have; that is, for scientific knowledge. I assert
that scientific knowledge is essentially conjectural or
hypothetical.
Take as an example classical Newtonian mechanics. There never
was a more successful theory. If repeated observational success
could establish a theory, it would have established Newton's
theory. Yet Newton's theory was superseded in the field of
astronomy by Einstein's theory, and in the atomic field by
quantum theory. And almost all physicists think now that
Newtonian classical mechanics is no more than a marvellous
conjecture, a strangely successful hypothesis, and a staggeringly
good approximation to the truth.
I can now formulate my central thesis, which is this. Once we
fully realize the implications of the conjectural character of
human knowledge, then the problem of induction changes its
character completely: there is no need any longer to be disturbed
by Hume's negative results, since there is no need any longer
to ascribe to human knowledge a validity derived from repeated
observations. Human knowledge possesses no such validity. On the
other hand, we can explain all our achievements in terms of the
method of trial and the elimination of error. To put it in a
nutshell, our conjectures are our trial balloons, and we test
them by criticizing them and by trying to replace them - by
trying to show that there can be better or worse conjectures, and
that they can be improved upon. The place of the problem of
induction is usurped by the problem of the comparative goodness
or badness of the rival conjectures or theories that have been
proposed.
The main barrier to accepting the conjectural character of
human knowledge, and to accepting that it contains the solution
of the problem of induction, is a doctrine which may be called
the commonsense theory of human knowledge or the bucket theory of
the human mind. 3
I think very highly of common sense. In fact, I think that
all philosophy must start from commonsense views and from their
critical examination.
For our purposes here I want to distinguish two parts of the
commonsense view of the world and draw attention to the fact that
they clash with one another.
The first is commonsense realism; this is the view that there
is a real world, with real people, animals and plants, cars and
stars in it. I think that this view is true and immensely
important, and I believe that no valid criticism of it has ever
been proposed. [See also selection 17 below.]
A very different part of the commonsense view of the world is
the commonsense theory of knowledge. The problem is the problem
of how we get knowledge about the world. The commonsense solution
is: by opening our eyes and ears. Our senses are the main if not
the only sources of our knowledge of the world.
This second view I regard as thoroughly mistaken, and as
insufficiently criticized (in spite of Leibniz and Kant). I call
it the bucket theory of the mind, because it can be summed up by
the diagram overleaf.
What allegedly enters the bucket through our senses are the
elements, the atoms or molecules, of knowledge. Our knowledge
then consists of an accumulation, a digest, or perhaps a
synthesis of the elements offered to us by our senses.
Both halves of commonsense philosophy, commonsense realism
and the commonsense theory of knowledge, were held by Hume; he
found, as did Berkeley before him, that there is a clash between
them. For the commonsense theory of knowledge is liable to lead
to a kind of anti-realism. If knowledge results from sensations,
then sensations are the only certain elements of knowledge, and
we can have no good reason to believe that anything but sensation
exists.
Hume, Berkeley, and Leibniz were all believers in a principle
of sufficient reason. For Berkeley and Hume the principle took
the form: if you do not have sufficient reasons for holding a
belief, then this fact is itself a sufficient reason for
abandoning this belief. Genuine knowledge consisted for both
Berkeley and Hume essentially of belief, backed by sufficient
reasons: but this led them to the position that knowledge
consists, more or less, of sensations on their own.
Thus for these philosophers the real world of common sense
does not really exist; according to Hume, even we ourselves do
not fully exist. All that exist are sensations, impressions, and
memory images. [See also selection 22, section 1, below.]
This anti-realistic view can be characterized by various
names, but the most usual name seems to be 'idealism'.
Hume's idealism appeared to him to be a strict refutation of
commonsense realism. But though he felt rationally obliged to
regard commonsense realism as a mistake, he himself admitted that
he was in practice quite unable to disbelieve in commonsense
realism for more than an hour.
Thus Hume experienced very strongly the clash between the two
parts of commonsense philosophy: realism, and the commonsense
theory of knowledge. And although he was aware that emotionally
he was unable to give up realism, he looked on this fact as a
mere consequence of irrational custom or habit; he was convinced
that a consistent adherence to the more critical results of the
theory of knowledge ought to make us abandon realism. 4
Fundamentally, Hume's idealism has remained the mainstream of
British empiricism.
Hume's two problems of induction -- the logical problem
and the psychological problem -- can best be presented, I think,
against the background of the commonsense theory of induction.
This theory is very simple. Since all knowledge is supposed to be
the result of past observation, so especially is all
expectational knowledge such as that the sun will rise tomorrow,
or that all men are bound to die, or that bread nourishes. All
this has to be the result of past observation.
It is to Hume's undying credit that he dared to challenge
the commonsense view of induction, even though he never doubted
that it must be largely true. He believed that induction by
repetition was logically untenable - that rationally, or
logically, no amount of observed instances can have the slightest
bearing upon unobserved instances. This is Hume's negative
solution of the problem of induction, a solution which I fully
endorse.
But Hume held, at the same time, that although induction was
rationally invalid, it was a psychological fact, and that we all
rely on it.
Thus Hume's two problems of induction were:
(1) The logical problem: Are we rationally justified in
reasoning from repeated instances of which we have had experience
to instances of which we have had no experience?
Hume's unrelenting answer was: No, we are not justified,
however great the number of repetitions may be. And he added that
it did not make the slightest difference if, in this problem, we
ask for the justification not of certain belief, but of probable
belief. Instances of which we have had experience do not allow us
to reason or argue about the probability of instances of which we
have had no experience, any more than to the certainty of such
instances.
(2) The following psychological question: How is it that
nevertheless all reasonable people expect and believe that
instances of which they have had no experience will conform to
those of which they have had experience? Or in other words, why
do we all have expectations, and why do we hold on to them with
such great confidence, or such strong belief?
Hume's answer to this psychological problem of induction
was: Because of 'custom or habit'; or in other words,
because of the irrational but irresistible power of the law of
association. We are conditioned by repetition; a conditioning
mechanism without which, Hume says, we could hardly survive.
My own view is that Hume's answer to the logical problem
is right and that his answer to the psychological problem is, in
spite of its persuasiveness, quite mistaken.
The answers given by Hume to the logical and psychological
problems of induction lead immediately to an irrationalist
conclusion. According to Hume, all our knowledge, especially all
our scientific knowledge, is just irrational habit or custom, and
it is rationally totally indefensible.
Hume himself thought of this as a form of scepticism; but it
was rather, as Bertrand Russell pointed out, an unintended
surrender to irrationalism. It is an amazing fact that a peerless
critical genius, one of the most rational minds of all ages, not
only came to disbelieve in reason, but became a champion of
unreason, of irrationalism.
Nobody has felt this paradox more strongly than Bertrand
Russell, an admirer and, in many respects, even a late disciple
of Hume. Thus in the Hume chapter in A History of Western
Philosophy, published in 1946, Russell says about Hume's
treatment of induction: 'Hume's philosophy ... represents
the bankruptcy of eighteenth-century reasonableness' and,
'It is therefore important to discover whether there is any
answer to Hume within a philosophy that is wholly or mainly
empirical. If not, there is no intellectual difference between
sanity and insanity. The lunatic who believes that he is a
poached egg is to be condemned solely on the ground that he is in
a minority ....'
Russell goes on to assert that if induction (or the principle
of induction) is rejected, 'every attempt to arrive at
general scientific laws from particular observations is
fallacious, and Hume's scepticism is inescapable for an
empiricist.'
And Russell sums up his view of the situation created by the
clash between Hume's two answers, by the following dramatic
remark:
'The growth of unreason throughout the nineteenth century
and what has passed of the twentieth is a natural sequel to
Hume's destruction of empiricism.' 5
This last quotation of Russell's goes perhaps too far. I
do not wish to overdramatize the situation; and although I
sometimes feel that Russell is right in his emphasis, at other
moments I doubt it.
Yet the following quotation from Professor Strawson seems to
me to support Russell's grave opinion: '[If] . . . there
is a problem of induction, and . . . Hume posed it, it must be
added that he solved it . . . [;] our acceptance of the
"basic canons" [of induction] ... is forced upon us by
Nature.... Reason is, and ought to be the slave of the
passions.' 6
However this may be, I assert that I have an answer to
Hume's psychological problem which completely removes the
clash between the logic and the psychology of knowledge; and with
it, it removes all of Hume's and Strawson's reasoning
against reason.
My own way of avoiding Hume's irrationalist consequences
is very simple. I solve the psychological problem of induction
(and also such formulations as the pragmatic problem) in a manner
which satisfies the following 'principle of the primacy of
the logical solution', or, more briefly, the 'principle
of transference'. The principle runs like this: the solution
of the logical problem of induction, far from clashing with those
of the psychological or pragmatic problems, can, with some care,
be directly transferred to them. As a result, there is no clash,
and there are no irrationalist consequences.
The logical problem of induction itself needs some
reformulation to start with.
First, it must be formulated in terms not only of
'instances' (as by Hume) but of universal regularities or
laws. Regularities or laws are presupposed by Hume's own term
'instance'; for an instance is an instance of something -
of a regularity or of a law. (Or, rather, it is an instance of
many regularities or many laws.)
Secondly, we must widen the scope of reasoning from instances
to laws so that we can take heed also of counterinstances.
In this way, we arrive at a reformulation of Hume's
logical problem of induction along the following lines:
Are we rationally justified in reasoning from instances or
from counterinstances of which we have had experience to the
truth or falsity of the corresponding laws or to instances of
which we have had no experience?
This is a purely logical problem. It is essentially merely a
slight extension of Hume's logical problem of induction
formulated here earlier, in section v.
The answer to this problem is: as implied by Hume, we
certainly are not justified in reasoning from an instance to the
truth of the corresponding law. But to this negative result a
second result, equally negative, may be added: we are justified
in reasoning from a counterinstance to the falsity of the
corresponding universal law (that is, of any law of which it is a
counterinstance). Or in other words, from a purely logical point
of view, the acceptance of one counterinstance to 'All swans
are white' implies the falsity of the law 'All swans are
white' - that law, that is, whose counterinstance we
accepted. Induction is logically invalid; but refutation or
falsification is a logically valid way of arguing from a single
counterinstance to - or, rather, against - the corresponding
law.
This shows that I continue to agree with Hume's negative
logical result; but I extend it.
This logical situation is completely independent of any
question of whether we would, in practice, accept a single
counterinstance - for example, a solitary black swan - in
refutation of a so far highly successful law. I do not suggest
that we would necessarily be so easily satisfied; we might well
suspect that the black specimen before us was not a swan. And in
practice, anyway, we would be most reluctant to accept an
isolated counterinstance. But this is a different question [see
section Iv of selection 10 below]. Logic forces us to reject even
the most successful law the moment we accept one single
counterinstance.
Thus we can say: Hume was right in his negative result that
there can be no logically valid positive argument leading in the
inductive direction. But there is a further negative result;
there are logically valid negative arguments leading in the
inductive direction: a counterinstance may disprove a law.
Hume's negative result establishes for good that all our
universal laws or theories remain for ever guesses, conjectures,
hypotheses. But the second negative result concerning the force
of counterinstances by no means rules out the possibility of a
positive theory of how, by purely rational arguments, we can
prefer some competing conjectures to others.
In fact, we can erect a fairly elaborate logical theory of
preference - preference from the point of view of the search for
truth.
To put it in a nutshell, Russell's desperate remark that
if with Hume we reject all positive induction, 'there is no
intellectual difference between sanity and insanity' is
mistaken. For the rejection of induction does not prevent us from
preferring, say, Newton's theory to Kepler's, or
Einstein's theory to Newton's: during our rational
critical discussion of these theories we may have accepted the
existence of counterexamples to Kepler's theory which do not
refute Newton's, and of counterexamples to Newton's which
do not refute Einstein's. Given the acceptance of these
counterexamples we can say that Kepler's and Newton's
theories are certainly false; whilst Einstein's may be true
or it may be false: that we don't know. Thus there may exist
purely intellectual preferences for one or the other of these
theories; and we are very far from having to say with Russell
that all the difference between science and lunacy disappears.
Admittedly, Hume's argument still stands, and therefore the
difference between a scientist and a lunatic is not that the
first bases his theories securely upon observations while the
second does not, or anything like that. Nevertheless we may now
see that there may be a difference: it may be that the
lunatic's theory is easily refutable by observation, while
the scientist's theory has withstood severe tests.
What the scientist's and the lunatic's theories have
in common is that both belong to conjectural knowledge. But some
conjectures are much better than others; and this is a sufficient
answer to Russell, and it is sufficient to avoid radical
scepticism. For since it is possible for some conjectures to be
preferable to others, it is also possible for our conjectural
knowledge to improve, and to grow. (Of course, it is possible
that a theory that is preferred to another at one time may fall
out of favour at a later time so that the other is now preferred
to it. But, on the other hand, this may not happen.)
We may prefer some competing theories to others on purely
rational grounds. It is important that we are clear what the
principles of preference or selection are.
In the first place they are governed by the idea of truth. We
want, if at all possible, theories which are true, and for this
reason we try to eliminate the false ones.
But we want more than this. We want new and interesting
truth. We are thus led to the idea of the growth of informative
content, and especially of truth content. That is, we are led to
the following principle of preference: a theory with a great
informative content is on the whole more interesting, even before
it has been tested, than a theory with little content.
Admittedly, we may have to abandon the theory with the greater
content, or as I also call it, the bolder theory, if it does not
stand up to tests . But even in this case we may have learned
more from it than from a theory with little content, for
falsifying tests can sometimes reveal new and unexpected facts
and problems. [See also selection 13 below.]
Thus our logical analysis leads us direct to a theory of
method, and especially to the following methodological rule: try
out, and aim at, bold theories, with great informative content;
and then let these bold theories compete, by discussing them
critically and by testing them severely.
My solution of the logical problem of induction was that we
may have preferences for certain of the competing conjectures;
that is, for those which are highly informative and which so far
have stood up to eliminative criticism. These preferred
conjectures are the result of selection, of the struggle for
survival of the hypotheses under the strain of criticism, which
is artificially intensified selection pressure.
The same holds for the psychological problem of induction.
Here too we are faced with competing hypotheses, which may
perhaps be called beliefs, and some of them are eliminated, while
others survive, anyway for the time being. Animals are often
eliminated along with their beliefs; or else they survive with
them. Men frequently outlive their beliefs; but for as long as
the beliefs survive (often a very short time), they form the
(momentary or lasting) basis of action.
My thesis is that this Darwinian procedure of the selection
of beliefs and actions can in no sense be described as
irrational. In no way does it clash with the rational solution of
the logical problem of induction. Rather, it is just the
transference of the logical solution to the psychological field.
(This does not mean, of course, that we never suffer from what
are called 'irrational beliefs'.)
Thus with an application of the principle of transference to
Hume's psychological problem Hume's irrationalist
conclusions disappear.
In talking of preference I have so far discussed only the
theoretician's preference - if he has any; and why it will be
for the 'better', that is, more testable, theory, and for
the better tested one. Of course, the theoretician may not have
any preference: he may be discouraged by Hume's, and my,
'sceptical' solution to Hume's logical problem; he
may say that, if he cannot make sure of finding the true theory
among the competing theories, he is not interested in any method
like the one described - not even if the method makes it
reasonably certain that, if a true theory should be among the
theories proposed, it will be among the surviving the preferred,
the corroborated ones. Yet a more sanguine or more dedicated or
more curious 'pure' theoretician may well be encouraged,
by our analysis, to propose again and again new competing
theories in the hope that one of them may be true - even if we
shall never be able to make sure of any one that it is true.
Thus the pure theoretician has more than one way of action
open to him; and he will choose a method such as the method of
trial and the elimination of error only if his curiosity exceeds
his disappointment at the unavoidable uncertainty and
incompleteness of all our endeavours.
It is different with him qua man of practical action. For a
man of practical action has always to choose between some more or
less definite alternatives, since even inaction is a kind of
action.
But every action presupposes a set of expectations, that is,
of theories about the world. Which theory shall the man of action
choose? Is there such a thing as a rational choice?
This leads us to the pragmatic problems of induction which to
start with, we might formulate thus:
(1) Upon which theory should we rely for practical action,
from a rational point of view?
(2) Which theory should we prefer for practical action, from
a rational point of view?
My answer to (1) is: from a rational point of view, we should
not 'rely' on any theory, for no theory has been shown to
be true, or can be shown to be true (or 'reliable').
My answer to (2) is: we should prefer the best tested theory
as a basis for action.
In other words, there is no 'absolute reliance'; but
since we have to choose, it will be 'rational' to choose
the best tested theory. This will be 'rational' in the
most obvious sense of the word known to me: the best tested
theory is the one which, in the light of our critical discussion,
appears to be the best so far; and I do not know of anything more
'rational' than a well-conducted critical
discussion.
Since this point appears not to have got home I shall try to
restate it here in a slightly new way, suggested to me by David
Miller. Let us forget momentarily about what theories we
'use' or 'choose' or 'base our practical
actions on', and consider only the resulting proposal or
decision (to do X; not to do X; to do nothing; or so on). Such a
proposal can, we hope, be rationally criticized; and if we are
rational agents we will want it to survive, if possible, the most
testing criticism we can muster. But such criticism will freely
make use of the best tested scientific theories in our
possession. Consequently any proposal that ignores these theories
(where they are relevant, I need hardly add) will collapse under
criticism. Should any proposal remain, it will be rational to
adopt it.
This seems to me all far from tautological. Indeed, it might
well be challenged by challenging the italicized sentence in the
last paragraph. Why, it might be asked, does rational criticism
make use of the best tested although highly unreliable theories?
The answer, however, is exactly the same as before. Deciding to
criticize a practical proposal from the standpoint of modern
medicine (rather than, say, in phrenological terms) is itself a
kind of 'practical' decision (anyway it may have
practical consequences). Thus the rational decision is always:
adopt critical methods that have themselves withstood severe
criticism.
There is, of course, an infinite regress here. But it is
transparently harmless.
Now I do not particularly want to deny (or, for that matter,
assert) that, in choosing the best tested theory as a basis for
action, we 'rely' on it, in some sense of the word. It
may therefore even be described as the most 'reliable'
theory available, in some sense of this term. Yet this is not to
say that it is 'reliable'. It is 'unreliable' at
least in the sense that we shall always do well, even in
practical action, to foresee the possibility that something may
go wrong with it and with our expectations.
But it is not merely this trivial caution which we must
derive from our negative reply to the pragmatic problem (1).
Rather, it is of the utmost importance for the understanding of
the whole problem, and especially of what I have called the
traditional problem, that in spite of the 'rationality'
of choosing the best tested theory as a basis of action, this
choice is not 'rational' in the sense that it is based
upon good reasons in favour of the expectation that it will in
practice be a successful choice: there can be no good reasons in
this sense, and this is precisely Hume's result. On the
contrary, even if our physical theories should be true, it is
perfectly possible that the world as we know it, with all its
pragmatically relevant regularities, may completely disintegrate
in the next second. This should be obvious to anybody today; but
I said so 7 before Hiroshima: there are infinitely many possible
causes of local, partial, or total disaster.
From a pragmatic point of view, however, most of these
possibilities are obviously not worth bothering about because we
cannot do anything about them: they are beyond the realm of
action. (I do not, of course, include atomic war among those
disasters which are beyond the realm of human action, although
most of us think in just this way since we cannot do more about
it than about an act of God.)
All this would hold even if we could be certain that our
physical and biological theories were true. But we do not know
it. On the contrary, we have very good reason to suspect even the
best of them; and this adds, of course, further infinities to the
infinite possibilities of catastrophe.
It is this kind of consideration which makes Hume's and
my own negative reply so important. For we can now see very
clearly why we must beware lest our theory of knowledge proves
too much. More precisely, no theory of knowledge should attempt
to explain why we are successful in our attempts to explain
things.
Even if we assume that we have been successful - that our
physical theories are true - we can learn from our cosmology how
infinitely improbable this success is: our theories tell us that
the world is almost completely empty, and that empty space is
filled with chaotic radiation. And almost all places which are
not empty are occupied either by chaotic dust, or by gases, or by
very hot stars - all in conditions which seem to make the
application of any physical method of acquiring knowledge
impossible.
There are many worlds, possible and actual worlds, in which a
search for knowledge and for regularities would fail. And even in
the world as we actually know it from the sciences, the
occurrence of conditions under which life, and a search for
knowledge, could arise - and succeed - seems to be almost
infinitely improbable. Moreover, it seems that if ever such
conditions should appear, they would be bound to disappear again,
after a time which, cosmologically speaking, is very short.
It is in this sense that induction is inductively invalid, as
I said above. That is to say, any strong positive reply to
Hume's logical problem (say, the thesis that induction is
valid) would be paradoxical. For, on the one hand, if induction
is the method of science, then modern cosmology is at least
roughly correct (I do not dispute this); and on the other, modern
cosmology teaches us that to generalize from observations taken,
for the most part, in our incredibly idiosyncratic region of the
universe would almost always be quite invalid. Thus if induction
is 'inductively valid' it will almost always lead to
false conclusions; and therefore it is inductively invalid.
Knowledge without Authority (1960)
This part of my lecture might be described as an attack on
empiricism, as formulated for example in the following classical
statement of Hume's: 'If I ask why you believe any
particular matter of fact . . ., you must tell me some reason;
and this reason will be some other fact, connected with it. But
as you cannot proceed after this manner, in infinitum, you must
at last terminate in some fact, which is present to your memory
or senses; or must allow that your belief is entirely without
foundation.' l
The problem of the validity of empiricism may be roughly put
as follows: is observation the ultimate source of our knowledge
of nature? And if not, what are the sources of our
knowledge?
These questions remain, whatever I may have said about Bacon,
and even if I should have managed to make those parts of his
philosophy on which I have commented somewhat unattractive for
Baconians and for other empiricists.
The problem of the source of our knowledge has recently been
restated as follows. If we make an assertion, we must justify it;
but this means that we must be able to answer the following
questions.
'How do you know? What are the sources of your
assertion?' This, the empiricist holds, amounts in its turn
to the question,
'What observations (or memories of observations) underlie
your assertion?
I find this string of questions quite unsatisfactory.
First of all, most of our assertions are not based upon
observations, but upon all kinds of other sources. 'I read it
in The Times' or perhaps 'I read it in the Encyclopaedia
Britannica' is a more likely and a more definite answer to
the question 'How do you know?' than 'I have observed
it' or 'I know it from an observation I made last
year'.
'But', the empiricist will reply, 'how do you
think that The Times or the Encyclopaedia Britannica got their
information? Surely, if you only carry on your inquiry long
enough, you will end up with reports of the observations of
eyewitnesses (sometimes called "protocol sentences" or-
by yourself- "basic statements"). Admittedly', the
empiricist will continue, 'books are largely made from other
books. Admittedly, a historian, for example, will work from
documents. But ultimately, in the last analysis, these other
books, or these documents, must have been based upon
observations. Otherwise they would have to be described as
poetry, or invention, or lies, but not as testimony. It is in
this sense that we empiricists assert that observation must be
the ultimate source of our knowledge.'
Here we have the empiricist's case, as it is still put by
some of my positivist friends.
I shall try to show that this case is as little valid as
Bacon's; that the answer to the question of the sources of
knowledge goes against the empiricist; and, finally, that this
whole question of ultimate sources - sources to which one may
appeal, as one might to a higher court or a higher authority -
must be rejected as based upon a mistake.
First I want to show that if you actually went on questioning
The Times and its correspondents about the sources of their
knowledge, you would in fact never arrive at all those
observations by eyewitnesses in the existence of which the
empiricist believes. You would find, rather, that with every
single step you take, the need for further steps increases in
snowball-like fashion.
Take as an example the sort of assertion for which reasonable
people might simply accept as sufficient the answer 'I read
it in The Times'; let us say the assertion 'The Prime
Minister has decided to return to London several days ahead of
schedule'. Now assume for a moment that somebody doubts this
assertion, or feels the need to investigate its truth. What shall
he do? If he has a friend in the Prime Minister's office, the
simplest and most direct way would be to ring him up; and if this
friend corroborates the message, then that is that.
In other words, the investigator will, if possible, try to
check, or to examine, the asserted fact itself, rather than trace
the source of the information. But according to the empiricist
theory, the assertion 'I have read it in The Times' is
merely a first step in a justification procedure consisting in
tracing the ultimate source. What is the next step?
There are at least two next steps. One would be to reflect
that 'I have read it in The Times' is also an assertion,
and that we might ask 'What is the source of your knowledge
that you read it in The Times and not, say, in a paper looking
very similar to The Times?' The other is to ask The Times for
the sources of its knowledge. The answer to the first question
may be 'But we have only The Times on order and we always get
it in the morning', which gives rise to a host of further
questions about sources which we shall not pursue. The second
question may elicit from the editor of The Times the answer:
'We had a telephone call from the Prime Minister's
office.' Now according to the empiricist procedure, we should
at this stage ask next: 'Who is the gentleman who received
the telephone call?' and then get his observation report; but
we should also have to ask that gentleman: 'What is the
source of your knowledge that the voice you heard came from an
official in the Prime Minister's office?', and so
on.
There is a simple reason why this tedious sequence of
questions never comes to a satisfactory conclusion. It is this.
Every witness must always make ample use, in his report, of his
knowledge of persons, places, things, linguistic usages, social
conventions, and so on. He cannot rely merely upon his eyes or
ears, especially if his report is to be of use in justifying any
assertion worth justifying. But this fact must of course always
raise new questions as to the sources of those elements of his
knowledge which are not immediately observational.
This is why the programme of tracing back all knowledge to
its ultimate source in observation is logically impossible to
carry through: it leads to an infinite regress. (The doctrine
that truth is manifest cuts off the regress. This is interesting
because it may help to explain the attractiveness of that
doctrine.)
I wish to mention, in parentheses, that this argument is
closely related to another - that all observation involves
interpretation in the light of our theoretical knowledge, or that
pure observational knowledge, unadulterated by theory, would, if
at all possible, be utterly barren and futile. [See the last
paragraph of selection 11, section 1, below. 2]
The most striking thing about the observationalist programme
of asking for sources - apart from its tediousness - is its stark
violation of common sense. For if we are doubtful about an
assertion, then the normal procedure is to test it, rather than
to ask for its sources; and if we find independent corroboration,
then we shall often accept the assertion without bothering at all
about sources.
Of course there are cases in which the situation is
different. Testing an historical assertion always means going
back to sources; but not, as a rule, to the reports of
eyewitnesses.
Clearly, no historian will accept the evidence of documents
uncritically. There are problems of genuineness, there are
problems of bias, and there are also such problems as the
reconstruction of earlier sources. There are, of course, also
problems such as: was the writer present when these events
happened? But this is not one of the characteristic problems of
the historian. He may worry about the reliability of a report,
but he will rarely worry about whether or not the writer of a
document was an eyewitness of the event in question, even
assuming that this event was of the nature of an observable
event. A letter saying 'I changed my mind yesterday on this
question' may be most valuable historical evidence, even
though changes of mind are unobservable (and even though we may
conjecture, in view of other evidence, that the writer was
lying).
As to eyewitnesses, they are important almost exclusively in
a court of law where they can be cross-examined. As most lawyers
know, eyewitnesses often err. This has been experimentally
investigated, with the most striking results. Witnesses most
anxious to describe an event as it happened are liable to make
scores of mistakes, especially if some exciting things happen in
a hurry; and if an event suggests some tempting interpretation,
then this interpretation, more often than not, is allowed to
distort what has actually been seen.
Hume's view of historical knowledge was different:
'... we believe', he writes in the Treatise, 'that
CAESAR was kill'd in the senate-house on the ides of March .
. . because this fact is establish'd on the unanimous
testimony of historians, who agree to assign this precise time
and place to that event. Here are certain characters and letters
present either to our memory or senses; which characters we
likewise remember to have been us'd as the signs of certain
ideas; and these ideas were either in the minds of such as were
immediately present at that action, and receiv'd the ideas
directly from its existence; or they were deriv'd from the
testimony of others, and that again from another testimony . . .
'till we arrive at those who were eye-witnesses and
spectators of the event.' 3
It seems to me that this view must lead to the infinite
regress described above. For the problem is, of course, whether
'the unanimous testimony of historians' is to be
accepted, or whether it is, perhaps, to be rejected as the result
of their reliance on a common yet spurious source. The appeal to
'letters present either to our memory or senses' cannot
have any bearing on this or on any other relevant problem of
historiography.
But what, then, are the sources of our knowledge?
The answer, I think, is this: there are all kinds of sources
of our knowledge; but none has authority.
We may say that The Times can be a source of knowledge, or
the Encyclopaedia Britannica. We may say that certain papers in
the Physical Review about a problem in physics have more
authority, and are more of the character of a source, than an
article about the same problem in The Times or the Encyclopaedia.
But it would be quite wrong to say that the source of the article
in the Physical Review must have been wholly, or even partly,
observation. The source may well be the discovery of an
inconsistency in another paper, or say, the discovery of the fact
that a hypothesis proposed in another paper could be tested by
such and such an experiment; all these non-observational
discoveries are 'sources' in the sense that they all add
to our knowledge.
I do not, of course, deny that an experiment may also add to
our knowledge, and in a most important manner. But it is not a
source in any ultimate sense. It has always to be checked: as in
the example of the news in The Times we do not, as a rule,
question the eyewitness of an experiment, but, if we doubt the
result, we may repeat the experiment, or ask somebody else to
repeat it.
The fundamental mistake made by the philosophical theory of
the ultimate sources of our knowledge is that it does not
distinguish clearly enough between questions of origin and
questions of validity. Admittedly, in the case of historiography,
these two questions may sometimes coincide. The question of the
validity of an historical assertion may be testable only, or
mainly, in the light of the origin of certain sources. But in
general the two questions are different; and in general we do not
test the validity of an assertion or information by tracing its
sources or its origin, but we test it, much more directly, by a
critical examination of what has been asserted - of the asserted
facts themselves.
Thus the empiricist's questions 'How do you know?
What is the source of your assertion?' are wrongly put. They
are not formulated in an inexact or slovenly manner, but they are
entirely misconceived: they are questions that beg for an
authoritarian answer.
The traditional systems of epistemology may be said to result
from yes-answers or no-answers to questions about the sources of
our knowledge. They never challenge these questions or dispute
their legitimacy; the questions are taken as perfectly natural,
and nobody seems to see any harm in them.
This is quite interesting, for these questions are clearly
authoritarian in spirit. They can be compared with that
traditional question of political theory, 'Who should
rule?', which begs for an authoritarian answer such as
'the best', or 'the wisest', or 'the
people', or 'the majority'. (It suggests,
incidentally, such silly alternatives as 'Who should be our
rulers: the capitalists or the workers?', analogous to
'What is the ultimate source of knowledge: the intellect or
the senses?') This political question is wrongly put and the
answers which it elicits are paradoxical [see selection 25
below]. It should be replaced by a completely different question
such as How can we organize our political institutions so that
bad or incompetent rulers (whom we should try not to get, but
whom we so easily might get all the same) cannot do too much
damage? I believe that only by changing our question in this way
can we hope to proceed towards a reasonable theory of political
institutions.
The question about the sources of our knowledge can be
replaced in a similar way. It has always been asked in the spirit
of: 'What are the best sources of our knowledge - the most
reliable ones, those which will not lead us into error, and those
to which we can and must turn, in case of doubt, as the last
court of appeal?' I propose to assume, instead, that no such
ideal sources exist - no more than ideal rulers - and that all
'sources' are liable to lead us into error at times. And
I propose to replace, therefore, the question of the sources of
our knowledge by the entirely different question: 'How can we
hope to detect and eliminate error?'
The question of the sources of our knowledge, like so many
authoritarian questions, is a genetic one. It asks for the origin
of our knowledge, in the belief that knowledge may legitimize
itself by its pedigree. The nobility of the racially pure
knowledge, the untainted knowledge, the knowledge which derives
from the highest authority, if possible from God: these are the
(often unconscious) metaphysical ideas behind the question. My
modified question, 'How can we hope to detect error?' may
be said to derive from the view that such pure, untainted and
certain sources do not exist, and that questions of origin or of
purity should not be confounded with questions of validity, or of
truth. This view may be said to be as old as Xenophanes.
Xenophanes knew that our knowledge is guesswork, opinion - doxa
rather than episteme - as shown by his verses [quoted on p. 31
above]. Yet the traditional question of the authoritative sources
of knowledge is repeated even today - and very often by
positivists and by other philosophers who believe themselves to
be in revolt against authority.
The proper answer to my question 'How can we hope to
detect and eliminate error?' is, I believe, 'By
criticizing the theories or guesses of others and - if we can
train ourselves to do so - by criticizing our own theories or
guesses.' (The latter point is highly desirable, but not
indispensable; for if we fail to criticize our own theories,
there may be others to do it for us.) This answer sums up a
position which I propose to call 'critical rationalism'.
It is a view, an attitude, and a tradition, which we owe to the
Greeks. It is very different from the 'rationalism' or
'intellectualism' of Descartes and his school, and very
different even from the epistemology of Kant. Yet in the field of
ethics, of moral knowledge, it was approached by Kant with his
principle of autonomy- This principle expresses his realization
that we must not accept the command of an authority, however
exalted, as the basis of ethics. For whenever we are faced with a
command by an authority, it is for us to judge, critically,
whether it is moral or immoral to obey. The authority may have
power to enforce its commands, and we may be powerless to resist.
But if we have the physical power of choice, then the ultimate
responsibility remains with us. It is our own critical decision
whether to obey a command; whether to submit to an
authority.
Kant boldly carried this idea into the field of religion:
'. . . in whatever way', he writes, 'the Deity should
be made known to you, and even . . . if He should reveal Himself
to you: it is you . . . who must judge whether you are permitted
to believe in Him, and to worship Him.' 4
In view of this bold statement, it seems strange that in his
philosophy of science Kant did not adopt the same attitude of
critical rationalism, of the critical search for error. I feel
certain that it was only his acceptance of the authority of
Newton's cosmology - a result of its almost unbelievable
success in passing the most severe tests - which prevented Kant
from doing so. If this interpretation of Kant is correct, then
the critical rationalism (and also the critical empiricism) which
I advocate merely put the finishing touch to Kant's own
critical philosophy. And this was made possible by Einstein, who
taught us that Newton's theory may well be mistaken in spite
of its overwhelming success.
So my answer to the questions 'How do you know? What is
the source or the basis of your assertion? What observations have
led you to it?' would be: 'I do not know: my assertion
was merely a guess. Never mind the source, or the sources, from
which it may spring - there are many possible sources, and I may
not be aware of half of them; and origins or pedigrees have in
any case little bearing upon truth. But if you are interested in
the problem which I tried to solve by my tentative assertion, you
may help me by criticizing it as severely as you can; and if you
can design some experimental test which you think might refute my
assertion, I shall gladly, and to the best of my powers, help you
to refute it.'
This answer applies, strictly speaking, only if the question
is asked about some scientific assertion as distinct from an
historical one. If my conjecture was an historical one, sources
(in the non-ultimate sense) will of course come into the critical
discussion of its validity. Yet fundamentally, my answer will be
the same, as we have seen.
It is high time now, I think, to formulate the
epistemological results of this discussion. I will put them in
the form of nine theses.
(1) There are no ultimate sources of knowledge. Every source,
every suggestion, is welcome; and every source, every suggestion,
is open to critical examination. Except in history, we usually
examine the facts themselves rather than the sources of our
information.
(2) The proper epistemological question is not one about
sources; rather, we ask whether the assertion made is true - that
is to say, whether it agrees with the facts. (That we may
operate, without getting involved in antinomies, with the idea of
objective truth in the sense of correspondence to the facts, has
been shown by the work of Alfred Tarski.) And we try to find this
out, as well as we can, by examining or testing the assertion
itself; either in a direct way, or by examining or testing its
consequences.
(3) In connection with this examination, all kinds of
arguments may be relevant. A typical procedure is to examine
whether our theories are consistent with our observations. But we
may also examine, for example, whether our historical sources are
mutually and internally consistent.
(4) Quantitatively and qualitatively by far the most
important source of our knowledge - apart from inborn knowledge -
is tradition. Most things we know we have learnt by example, by
being told, by reading books, by learning how to criticize, how
to take and to accept criticism, how to respect truth.
(5) The fact that most of the sources of our knowledge are
traditional condemns anti-traditionalism as futile. But this fact
must not be held to support a traditionalist attitude: every bit
of our traditional knowledge (and even our inborn knowledge) is
open to critical examination and may be overthrown. Nevertheless,
without tradition, knowledge would be impossible.
(6) Knowledge cannot start from nothing - from a tabula rasa
- nor yet from observation. The advance of knowledge consists,
mainly, in the modification of earlier knowledge. Although we may
sometimes, for example in archaeology, advance through a chance
observation, the significance of the discovery will usually
depend upon its power to modify our earlier theories.
(7) Pessimistic and optimistic epistemologies are about
equally mistaken. The pessimistic cave story of Plato is the true
one, and not his optimistic story of anamnesis (even though we
should admit that all men, like all other animals, and even all
plants, possess inborn knowledge). But although the world of
appearances is indeed a world of mere shadows on the walls of our
cave, we all constantly reach out beyond it; and although, as
Democritus said, the truth is hidden in the deep, we can probe
into the deep. There is no criterion of truth at our disposal,
and this fact supports pessimism. But we do possess criteria
which, if we are lucky, may allow us to recognize error and
falsity. Clarity and distinctness are not criteria of truth, but
such things as obscurity or confusion may indicate error.
Similarly coherence cannot establish truth, but incoherence and
inconsistency do establish falsehood. And, when they are
recognized, our own errors provide the dim red lights which help
us in groping our way out of the darkness of our cave.
(8) Neither observation nor reason is an authority.
Intellectual intuition and imagination are most important, but
they are not reliable: they may show us things very clearly, and
yet they may mislead us. They are indispensable as the main
sources of our theories; but most of our theories are false
anyway. The most important function of observation and reasoning,
and even of intuition and imagination, is to help us in the
critical examination of those bold conjectures which are the
means by which we probe into the unknown.
(9) Every solution of a problem raises new unsolved problems;
the more so the deeper the original problem and the bolder its
solution. The more we learn about the world, and the deeper our
learning, the more conscious, specific, and articulate will be
our knowledge of what we do not know, our knowledge of our
ignorance. For this, indeed, is the main source of our ignorance
- the fact that our knowledge can only be finite, while our
ignorance must necessarily be infinite.
We may get a glimpse of the vastness of our ignorance when we
contemplate the vastness of the heavens: though the mere size of
the universe is not the deepest cause of our ignorance, it is one
of its causes. 'Where I seem to differ from some of my
friends', F.P. Ramsey wrote in a charming passage, 'is in
attaching little importance to physical size. I don't feel
the least humble before the vastness of the heavens. The stars
may be large but they cannot think or love; and these are
qualities which impress me far more than size does. I take no
credit for weighing nearly seventeen stone.' 5 I suspect that
Ramsey's friends would have agreed with him about the
insignificance of sheer physical size; and I suspect that if they
felt humble before the vastness of the heavens, this was because
they saw in it a symbol of their ignorance.
I believe that it would be worth trying to learn something
about the world even if in trying to do so we should merely learn
that we do not know much. This state of learned ignorance might
be a help in many of our troubles. It might be well for all of us
to remember that, while differing widely in the various little
bits we know, in our infinite ignorance we are all equal.
There is a last question I wish to raise.
If only we look for it we can often find a true idea, worthy
of being preserved in a philosophical theory which must be
rejected as false. Can we find an idea like this in one of the
theories of the ultimate sources of our knowledge?
I believe we can; and I suggest that it is one of the two
main ideas which underlie the doctrine that the source of all our
knowledge is super-natural. The first of these ideas is false, I
believe, while the second is true.
The first, the false idea, is that we must justify our
knowledge, or our theories, by positive reasons, that is, by
reasons capable of establishing them, or at least of making them
highly probable; at any rate, by better reasons than that they
have so far withstood criticism. This idea implies, I suggest,
that we must appeal to some ultimate or authoritative source of
true knowledge; which still leaves open the character of that
authority - whether it is human, like observation or reason, or
super-human (and therefore super-natural).
The second idea - whose vital importance has been stressed by
Russell - is that no man's authority can establish truth by
decree; that we should submit to truth; that truth is above human
authority.
Taken together these two ideas almost immediately yield the
conclusion that the sources from which our knowledge derives must
be super-human; a conclusion which tends to encourage
self-righteousness and the use of force against those who refuse
to see the divine truth.
Some who rightly reject this conclusion do not, unhappily,
reject the first idea - the belief in the existence of ultimate
sources of knowledge. Instead they reject the second idea - the
thesis that truth is above human authority. They thereby endanger
the idea of the objectivity of knowledge, and of common standards
of criticism or rationality.
What we should do, I suggest, is to give up the idea of
ultimate sources of knowledge, and admit that all human knowledge
is human: that it is mixed with our errors, our prejudices, our
dreams, and our hopes: that all we can do is to grope for truth
even though it be beyond our reach. We may admit that our groping
is often inspired, but we must be on our guard against the
belief, however deeply felt, that our inspiration carries any
authority, divine or otherwise. If we thus admit that there is no
authority beyond the reach of criticism to be found within the
whole province of our knowledge, however far it may have
penetrated into the unknown, then we can retain, without danger,
the idea that truth is beyond human authority. And we must retain
it. For without this idea there can be no objective standards of
inquiry; no criticism of our conjectures; no groping for the
unknown; no quest for knowledge.
Two Kinds of Definitions (1945)
The chief danger to our philosophy, apart from
laziness and woolliness, is scholasticism, . . . which is
treating what is vague as if it were precise....
F. P. RAMSEY
The problem of definitions and of the 'meaning of
terms' is the most important source of Aristotle's
regrettably still prevailing intellectual influence, of all that
verbal and empty scholasticism that haunts not only the Middle
Ages, but our own contemporary philosophy; for even a philosophy
as recent as that of L. Wittgenstein suffers, as we shall see,
from this influence. The development of thought since Aristotle
could, I think, be summed up by saying that every discipline, as
long as it used the Aristotelian method of definition, has
remained arrested in a state of empty verbiage and barren
scholasticism, and that the degree to which the various sciences
have been able to make any progress depended on the degree to
which they have been able to get rid of this essentialist method.
(This is why so much of our 'social science' still
belongs to the Middle Ages.) The discussion of this method will
have to be a little abstract, owing to the fact that the problem
has been so thoroughly muddled by Plato and Aristotle, whose
influence has given rise to such deep-rooted prejudices that the
prospect of dispelling them does not seem very bright. In spite
of all that, it is perhaps not without interest to analyse the
source of so much confusion and verbiage.
Aristotle followed Plato in distinguishing between knowledge
and opinion. 2 Knowledge, or science, according to Aristotle, may
be of two kinds - either demonstrative or intuitive.
Demonstrative knowledge is also a knowledge of 'causes'.
It consists of statements that can be demonstrated - the
conclusions - together with their syllogistic demonstrations
(which exhibit the 'causes' in their 'middle
terms'). Intuitive knowledge consists in grasping the
'indivisible form' or essence or essential nature of a
thing (if it is 'immediate', i.e. if its 'cause'
is identical with its essential nature); it is the originative
source of all science since it grasps the original basic
premisses of all demonstrations.
Undoubtedly, Aristotle was right when he insisted that we
must not attempt to prove or demonstrate all our knowledge. Every
proof must proceed from premisses; the proof as such, that is to
say, the derivation from the premisses, can therefore never
finally settle the truth of any conclusion, but only show that
the conclusion must be true provided the premisses are true . If
we were to demand that the premisses should be proved in their
turn, the question of truth would only be shifted back by another
step to a new set of premisses, and so on, to infinity. It was in
order to avoid such an infinite regress (as the logicians say)
that Aristotle taught that we must assume that there are
premisses which are indubitably true, and which do not need any
proof; and these he called 'basic premisses'. If we take
for granted the methods by which we derive conclusions from these
basic premisses, then we could say that, according to Aristotle,
the whole of scientific knowledge is contained in the basic
premisses, and that it would all be ours if only we could obtain
an encyclopaedic list of the basic premisses. But how to obtain
these basic premisses? Like Plato, Aristotle believed that we
obtain all knowledge ultimately by an intuitive grasp of the
essences of things. 'We can know a thing only by knowing its
essence', Aristotle writes, and 'to know a thing is to
know its essence'. A 'basic premiss' is, according to
him, nothing but a statement describing the essence of a thing.
But such a statement is just what he calls 3 a definition. Thus
all 'basic premisses of proofs' are definitions.
What does a definition look like? An example of a definition
would be: 'A puppy is a young dog.' The subject of such a
definition sentence, the term 'puppy', is called the term
to be defined (or defined term); the words 'young dog'
are called the defining formula. As a rule, the defining formula
is longer and more complicated than the defined term, and
sometimes very much so. Aristotle considers 4 the term to be
defined as a name of the essence Of a thing, and the defining
formula as the description of that essence. And he insists that
the defining formula must give an exhaustive description of the
essence or the essential properties of the thing in question;
thus a statement like 'A puppy has four legs', although
true, is not a satisfactory definition, since it does not exhaust
what may be called the essence of puppiness, but holds true of a
horse also; and similarly the statement 'A puppy is
brown', although it may be true of some, is not true of all
puppies; and it describes what is not an essential but merely an
accidental property of the defined term.
But the most difficult question is how we can get hold of
definitions or basic premisses, and make sure that they are
correct - that we have not erred, not grasped the wrong essence.
Although Aristotle is not very clear on this point, there can be
little doubt that, in the main, he again follows Plato. Plato
taught 5 that we can grasp the Ideas with the help of some kind
of unerring intellectual intuition; that is to say, we visualise
or look at them with our 'mental eye', a process which he
conceived as analogous to seeing, but dependent purely upon our
intellect, and excluding any element that depends upon our
senses. Aristotle's view is less radical and less inspired
than Plato's, but in the end it amounts to the same. 6 For
although he teaches that we arrive at the definition only after
we have made many observations, he admits that sense experience
does not in itself grasp the universal essence, and that it
cannot, therefore, fully determine a definition. Eventually he
simply postulates that we possess an intellectual intuition, a
mental or intellectual faculty which enables us unerringly to
grasp the essences of things, and to know them. And he further
assumes that if we know an essence intuitively, we must be
capable of describing it and therefore of defining it. (His
arguments in the Posterior Analytics in favour of this theory are
surprisingly weak. They consist merely in pointing out that our
knowledge of the basic premisses cannot be demonstrative, since
this would lead to an infinite regress, and that the basic
premisses must be at least as true and as certain as the
conclusions based upon them. 'It follows from this', he
writes, 'that there cannot be demonstrative knowledge of the
primary premisses; and since nothing but intellectual intuition
can be more true than demonstrative knowledge, it follows that it
must be intellectual intuition that grasps the basic
premisses.' In the De Anima, and in the theological part of
the Metaphysics, we find more of an argument; for here we have a
theory of intellectual intuition - that it comes into contact
with its object, the essence, and that it even becomes one with
its object. 'Actual knowledge is identical with its
object.')
Summing up this brief analysis, we can give, I believe, a
fair description of the Aristotelian ideal of perfect and
complete knowledge if we say that he saw the ultimate aim of all
inquiry in the compilation of an encyclopaedia containing the
intuitive definitions of all essences, that is to say, their
names together with their defining formulae; and that he
considered the progress of knowledge as consisting in the gradual
accumulation of such an encyclopaedia, in expanding it as well as
in filling up the gaps in it and, of course, in the syllogistic
derivation from it of 'the whole body of facts' which
constitute demonstrative knowledge.
Now there can be little doubt that all these essentialist
views stand in the strongest possible contrast to the methods of
modern science. (I have the empirical sciences in mind, not
perhaps pure mathematics.) First, although in science we do our
best to find the truth, we are conscious of the fact that we can
never be sure whether we have got it. We have learnt in the past,
from many disappointments, that we must not expect finality. And
we have learnt not to be disappointed any longer if our
scientific theories are overthrown; for we can, in most cases,
determine with great confidence which of any two theories is the
better one. We can therefore know that we are making progress;
and it is this knowledge that to most of us atones for the loss
of the illusion of finality and certainty. In other words, we
know that our scientific theories must always remain hypotheses,
but that, in many important cases, we can find out whether or not
a new hypothesis is superior to an old one. For if they are
different, then they will lead to different predictions, which
can often be tested experimentally; and on the basis of such a
crucial experiment, we can sometimes find out that the new theory
leads to satisfactory results where the old one breaks down. Thus
we can say that in our search for truth, we have replaced
scientific certainty by scientific progress. And this view of
scientific method is corroborated by the development of science.
For science does not develop by a gradual encyclopaedic
accumulation of essential information, as Aristotle thought) but
by a much more revolutionary method; it progresses by bold ideas,
by the advancement of new and very strange theories (such as the
theory that the earth is not flat, or that 'metrical
space' is not flat), and by the overthrow of the old
ones.
But this view of scientific method [developed in selections
9-14 below] means that in science there is no
'knowledge', in the sense in which Plato and Aristotle
understood the word, in the sense which implies finality; in
science, we never have sufficient reason for the belief that we
have attained the truth. What we usually call 'scientific
knowledge' is, as a rule, not knowledge in this sense, but
rather information regarding the various competing hypotheses and
the way in which they have stood up to various tests; it is,
using the language of Plato and Aristotle, information concerning
the latest, and the best tested, scientific 'opinion'.
This view means, furthermore, that we have no proofs in science
(excepting, of course, pure mathematics and logic). In the
empirical sciences, which alone can furnish us with information
about the world we live in, proofs do not occur, if we mean by
'proof' an argument which establishes once and for ever
the truth of a theory. (What may occur, however, are refutations
of scientific theories.) On the other hand, pure mathematics and
logic, which permit of proofs, give us no information about the
world, but only develop the means of describing it. Thus we could
say (as I have pointed out elsewhere 7 ): 'In so far as a
scientific statement speaks about reality, it must be
falsifiable; and in so far as it is not falsifiable, it does not
speak about reality.' But although proof does not play any
part in the empirical sciences, argument still does; indeed, its
part is at least as important as that played by observation and
experiment.
The role of definitions in science, especially, is also very
different from what Aristotle had in mind. Aristotle taught that
in a definition we have first pointed to the essence - perhaps by
naming it - and that we then describe it with the help of the
defining formula; just as in an ordinary sentence like 'This
puppy is brown', we first point to a certain thing by saying
'this puppy', and then describe it as 'brown'.
And he taught that by thus describing the essence to which the
term points which is to be defined, we determine or explain the
meaning 8 of the term also.
Accordingly, the definition may at one time answer two very
closely related questions. The one is 'What is it?', for
example 'What is a puppy?'; it asks what the essence is
which is denoted by the defined term. The other is 'What does
it mean?', for example, 'What does "puppy"
mean?'; it asks for the meaning of a term (namely, of the
term that denotes the essence). In the present context, it is not
necessary to distinguish between these two questions; rather, it
is important to see what they have in common; and I wish,
especially, to draw attention to the fact that both questions are
raised by the term that stands, in the definition, on the left
side and answered by the defining formula which stands on the
right side. This fact characterizes the essentialist view, from
which the scientific method of definition radically differs.
While we may say that the essentialist interpretation reads a
definition 'normally', that is to say, from the left to
the right, we can say that a definition, as it is normally used
in modern science, must be read back to front, or from the right
to the left; for it starts with the defining formula, and asks
for a short label for it. Thus the scientific view of the
definition 'A puppy is a young dog' would be that it is
an answer to the question 'What shall we call a young
dog?' rather than an answer to the question 'What is a
puppy?' (Questions like 'What is life?' or 'What
is gravity?' do not play any role in science.) The scientific
use of definitions, characterized by the approach 'from the
right to the left', may be called its nominalist
interpretation, as opposed to its Aristotelian or essentialist
interpretation. 9 In modern science, only 10 nominalist
definitions occur, that is to say, shorthand symbols or labels
are introduced in order to cut a long story short. And we can at
once see from this that definitions do not play any very
important part in science. For shorthand symbols can always, of
course, be replaced by the longer expressions, the defining
formulae, for which they stand. In some cases this would make our
scientific language very cumbersome; we should waste time and
paper. But we should never lose the slightest piece of factual
information. Our 'scientific knowledge', in the sense in
which this term may be properly used, remains entirely unaffected
if we eliminate all definitions; the only effect is upon our
language, which would lose, not precision, but merely brevity.
(This must not be taken to mean that in science there cannot be
an urgent practical need for introducing definitions, for
brevity's sake.) There could hardly be a greater contrast
than that between this view of the part played by definitions,
and Aristotle's view. For Aristotle's essentialist
definitions are the principles from which all our knowledge is
derived; they thus contain all our knowledge; and they serve to
substitute a long formula for a short one. As opposed to this,
the scientific or nominalist definitions do not contain any
knowledge whatever, not even any 'opinion'; they do
nothing but introduce new arbitrary shorthand labels; they cut a
long story short.
In practice, these labels are of the greatest usefulness. In
order to see this, we only need to consider the extreme
difficulties that would arise if a bacteriologist, whenever he
spoke of a certain strain of bacteria, had to repeat its whole
description (including the methods of dyeing, etc., by which it
is distinguished from a number of similar species). And we may
also understand, by a similar consideration, why it has so often
been forgotten, even by scientists, that scientific definitions
must be read 'from the right to the left', as explained
above. For most people, when first studying a science, say
bacteriology, must try to find out the meanings of all these new
technical terms with which they are faced. In this way, they
really learn the definition 'from the left to the right',
substituting, as if it were an essentialist definition, a very
long story for a very short one. But this is merely a
psychological accident, and a teacher or writer of a textbook may
indeed proceed quite differently; that is to say, he may
introduce a technical term only after the need for it has
arisen.
So far I have tried to show that the scientific or nominalist
use of definitions is entirely different from Aristotle's
essentialist method of definitions. But it can also be shown that
the essentialist view of definitions is simply untenable in
itself. In order not to prolong this discussion unduly, I shall
criticize two only of the essentialist doctrines; two doctrines
which are of significance because some influential modern schools
are still based upon them. One is the esoteric doctrine of
intellectual intuition, and the other the very popular doctrine
that 'we must define our terms', if we wish to be
precise.
Aristotle held with Plato that we possess a faculty,
intellectual intuition) by which we can visualize essences and
find out which definition is the correct one, and many modern
essentialists have repeated this doctrine. Other philosophers,
following Kant, maintain that we do not possess anything of the
sort. My opinion is that we can readily admit that we possess
something which may be described as 'intellectual
intuition'; or more precisely, that certain of our
intellectual experiences may be thus described Everybody who
'understands' an idea, or a point of view, or an
arithmetical method, for instance, multiplication, in the sense
that he has 'got the feel of it', might be said to
understand that thing intuitively; and there are countless
intellectual experiences of that kind. But I would insist, on the
other hand, that these experiences, important as they may be for
our scientific endeavours, can never serve to establish the truth
of any idea or theory, however strongly somebody may feel,
intuitively, that it must be true, or that it is
'self-evident'." Such intuitions cannot even serve
as an argument, although they may encourage us to look for
arguments. For somebody else may have just as strong an intuition
that the same theory is false. The way of science is paved with
discarded theories which were once declared self-evident; Francis
Bacon, for example, sneered at those who denied the self-evident
truth that the sun and the stars rotated round the earth, which
was obviously at rest. Intuition undoubtedly plays a great part
in the life of a scientist, just as it does in the life of a
poet. It leads him to his discoveries. But it may also lead him
to his failures. And it always remains his private affair, as it
were. Science does not ask how he has got his ideas, it is only
interested in arguments that can be tested by everybody. The
great mathematician, Gauss, described this situation very neatly
once when he exclaimed: 'I have got my result; but I do not
know yet how to get it.' All this applies, of course, to
Aristotle's doctrine of intellectual intuition of so-called
essences, which was propagated by Hegel and in our own time by E.
Husserl and his numerous pupils; and it indicates that the
'intellectual intuition of essences' or 'pure
phenomenology', as Husserl calls it, is a method of neither
science nor philosophy . (The much debated question whether it is
a new invention, as the pure phenomenologists think, or perhaps a
version of Cartesianism or Hegelianism, can be easily decided; it
is a version of Aristotelianism.)
The second doctrine to be criticized has even more important
connections with modern views; and it bears especially upon the
problem of verbalism. Since Aristotle, it has become widely known
that one cannot prove all statements, and that an attempt to do
so would break down because it would lead only to an infinite
regression of proofs. But neither he 12 nor, apparently, a great
many modern writers seems to realize that the analogous attempt
to define the meaning of all our terms must, in the same way,
lead to an infinite regression of definitions. The following
passage from Crossman's Plato To-day is characteristic of a
view which by implication is held by many contemporary
philosophers of repute, for example, by Wittgenstein: 13 '. .
. if we do not know precisely the meaning of the words we use, we
cannot discuss anything profitably. Most of the futile arguments
on which we all waste time are largely due to the fact that we
each have our own vague meaning for the words we use and assume
that our opponents are using them in the same sense. If we
defined our terms to start with, we could have far more
profitable discussions. Again, we have only to read the daily
papers to observe that propaganda (the modern counterpart of
rhetoric) depends largely for its success on confusing the
meaning of the terms. If politicians were compelled by law to
define any term they wished to use, they would lose most of their
popular appeal, their speeches would be shorter, and many of
their disagreements would be found to be purely verbal.' This
passage is very characteristic of one of the prejudices which we
owe to Aristotle, of the prejudice that language can be made more
precise by the use of definitions. Let us consider whether this
can really be done.
First, we can see clearly that if 'politicians' (or
anybody else) 'were compelled by law to define any term they
wished to use', their speeches would not be shorter, but
infinitely long. For a definition cannot establish the meaning of
a term any more than a logical derivation can establish the truth
of a statement; both can only shift this problem back. The
derivation shifts the problem of truth back to the premises, the
definition shifts the problem of meaning back to the defining
terms (i.e., the terms that make up the defining formula). 14 But
these, for many reasons, are likely to be just as vague and
confusing as the terms we started with; and in any case, we
should have to go on to define them in turn; which leads to new
terms which too must be defined. And so on, to infinity. One sees
that the demand that all our terms should be defined is just as
untenable as the demand that all our statements should be
proved.
At first sight this criticism may seem unfair. It may be said
that what people have in mind, if they demand definitions, is the
elimination of the ambiguities so often connected with words such
as 15 'democracy', 'liberty', 'duty',
'religion', etc.; that it is clearly impossible to define
all our terms, but possible to define some of these more
dangerous terms and to leave it at that; and that the defining
terms have just to be accepted, i.e., that we must stop after a
step or two in order to avoid an infinite regression. This
defence, however, is untenable. Admittedly, the terms mentioned
are much misused. But I deny that the attempt to define them can
improve matters. It can only make matters worse. That by
'defining their terms' even once, and leaving the
defining terms undefined, the politicians would not be able to
make their speeches shorter, is clear; for any essentialist
definition, i.e. one that 'defines our terms' (as opposed
to the nominalist one which introduces new technical terms),
means the substitution of a long story for a short one, as we
have seen. Besides, the attempt to define terms would only
increase the vagueness and confusion. For since we cannot demand
that all the defining terms should be defined in their turn, a
clever politician or philosopher could easily satisfy the demand
for definitions. If asked what he means by 'democracy',
for example, he could say 'the rule of the general will'
or 'the rule of the spirit of the people'; and since he
has now given a definition, and so satisfied the highest
standards of precision, nobody will dare to criticize him any
longer. And, indeed, how could he be criticized, since the demand
that 'rule' or 'people' or 'will' or
'spirit' should be defined in their turn, puts us well on
the way to an infinite regression so that everybody would
hesitate to raise it? But should it be raised in spite of all
that, then it can be equally easily satisfied. On the other hand,
a quarrel about the question whether the definition was correct,
or true, can only lead to an empty controversy about words.
Thus the essentialist view of definition breaks down, even if
it does not, with Aristotle, attempt to establish the
'principles' of our knowledge, but only makes the
apparently more modest demand that we should 'define the
meaning of our terms'.
But undoubtedly, the demand that we speak clearly and without
ambiguity is very important, and must be satisfied. Can the
nominalist view satisfy it? And can nominalism escape the
infinite regression?
It can. For the nominalist position there is no difficulty
which corresponds to the infinite regression. As we have seen,
science does not use definitions in order to determine the
meaning of its terms, but only in order to introduce handy
shorthand labels. And it does not depend on definitions; all
definitions can be omitted without loss to the information
imparted. It follows from this that in science, all the terms
that are really needed must be undefined terms. How then do the
sciences make sure of the meanings of their terms? Various
replies to this question have been suggested, 16 but I do not
think that any of them is satisfactory. The situation seems to be
this. Aristotelianism and related philosophies have told us for
such a long time how important it is to get a precise knowledge
of the meaning of our terms that we are all inclined to believe
it. And we continue to cling to this creed in spite of the
unquestionable fact that philosophy, which for twenty centuries
has worried about the meaning of its terms, is not only full of
verbalism but also appallingly vague and ambiguous, while a
science like physics which worries hardly at all about terms and
their meaning, but about facts instead, has achieved great
precision. This, surely, should be taken as indicating that,
under Aristotelian influence, the importance of the meaning of
terms has been grossly exaggerated. But I think that it indicates
even more. For not only does this concentration on the problem of
meaning fail to establish precision; it is itself the main source
of vagueness, ambiguity, and confusion.
In science, we take care that the statements we make should
never depend upon the meaning of our terms. Even where the terms
are defined, we never try to derive any information from the
definition, or to base any argument upon it. This is why our
terms make so little trouble. We do not overburden them. We try
to attach to them as little weight as possible. We do not take
their 'meaning' too seriously. We are always conscious
that our terms are a little vague (since we have learnt to use
them only in practical applications) and we reach precision not
by reducing their penumbra of vagueness, but rather by keeping
well within it, by carefully phrasing our sentences in such a way
that the possible shades of meaning of our terms do not matter.
This is how we avoid quarrelling about words.
The view that the precision of science and of scientific
language depends upon the precision of its terms is certainly
very plausible, but it is none the less, I believe, a mere
prejudice. The precision of a language depends, rather, just upon
the fact that it takes care not to burden its terms with the task
of being precise. A term like 'sand-dune' or
'wind' is certainly very vague. (How many inches high
must a little sand-hill be in order to be called
'sand-dune'? How quickly must the air move in order to be
called 'wind'?) However, for many of the geologist's
purposes, these terms are quite sufficiently precise; and for
other purposes, when a higher degree of differentiation is
needed, he can always say 'dunes between 4 and 30 feet
high' or 'wind of a velocity of between 20 and 40 miles
an hour'. And the position in the more exact sciences is
analogous. In physical measurements, for instance, we always take
care to consider the range within which there may be an error;
and precision does not consist in trying to reduce this range to
nothing, or in pretending that there is no such range, but rather
in its explicit recognition.
Even where a term has made trouble, as for instance the term
'simultaneity' in physics, it was not because its meaning
was imprecise or ambiguous, but rather because of some intuitive
theory which induced us to burden the term with too much meaning,
or with too 'precise' a meaning, rather than with too
little. What Einstein found in his analysis of simultaneity was
that, when speaking of simultaneous events, physicists made a
false assumption which would have been unchallengeable were there
signals of infinite velocity. The fault was not that they did not
mean anything, or that their meaning was ambiguous, or the term
not precise enough; what Einstein found was, rather, that the
elimination of a theoretical assumption, unnoticed so far because
of its intuitive self-evidence, was able to remove a difficulty
which had arisen in science. Accordingly, he was not really
concerned with a question of the meaning of a term, but rather
with the truth of a theory. It is very unlikely that it would
have led to much if someone had started, apart from a definite
physical problem, to improve the concept of simultaneity by
analysing its 'essential] meaning', or even by analysing
what physicists 'really mean' when they speak of
simultaneity.
I think we can learn from this example that we should not
attempt to cross our bridges before we come to them. And I also
think that the preoccupation with questions concerning the
meaning of terms, such as their vagueness or their ambiguity, can
certainly not be justified by an appeal to Einstein's
example. Such a preoccupation rests, rather, on the assumption
that much depends upon the meaning of our terms, and that we
operate with this meaning; and therefore it must lead to
verbalism and scholasticism. From this point of view, we may
criticize a doctrine like that of Wittgenstein, 17 who holds that
while science investigates matters of fact, it is the business of
philosophy to clarify the meaning of terms, thereby purging our
language, and eliminating linguistic puzzles. It is
characteristic of the views of this school that they do not lead
to any chain of argument that could be rationally criticized; the
school therefore addresses its subtle analyses 18 exclusively to
the small esoteric circle of the initiated. This seems to suggest
that any preoccupation with meaning tends to lead to that result
which is so typical of Aristotelianism: scholasticism and
mysticism.
Let us consider briefly how these two typical results of
Aristotelianism have arisen. Aristotle insisted that
demonstration or proof, and definition, are the two fundamental
methods of obtaining knowledge. Considering the doctrine of proof
first, it cannot be denied that it has led to countless attempts
to prove more than can be proved; medieval philosophy is full of
this scholasticism and the same tendency can be observed, on the
Continent, down to Kant. It was Kant's criticism of all
attempts to prove the existence of God which led to the romantic
reaction of Fichte, Schelling, and Hegel. The new tendency is to
discard proofs, and with them, any kind of rational argument.
With the romantics, a new kind of dogmatism becomes fashionable,
in philosophy as well as in the social sciences. It confronts us
with its dictum. And we can take it or leave it. This romantic
period of an oracular philosophy, called by Schopenhauer the
'age of dishonesty', is described by him as follows: 19
'The character of honesty that spirit of undertaking an
inquiry together with the reader, which permeates the works of
all Previous philosophers, disappears here completely. Every page
witnesses that these so-called philosophers do not attempt to
teach, but to bewitch the reader.'
A similar result was produced by Aristotle's doctrine of
definition. First it led to a good deal of hairsplitting. But
later, philosophers began to feel that one cannot argue about
definitions. In this way, essentialism not only encouraged
verbalism, but it also led to the disillusionment with argument,
that is, with reason. Scholasticism and mysticism and despair in
reason, these are the unavoidable results of the essentialism of
Plato and Aristotle. And Plato's open revolt against freedom
becomes, with Aristotle, a secret revolt against reason.
As we know from Aristotle himself, essentialism and the
theory of definition met with strong opposition when they were
first proposed, especially from Socrates's old companion
Antisthenes, whose criticism seems to have been most sensible. 20
But this opposition was unfortunately defeated. The consequences
of this defeat for the intellectual development of mankind can
hardly be overrated.