propositional connectives, which include accounts of
*material implication, *strict implication, and relevant
implication. The Megarians and the Stoics also investi-
gated various logical *antinomies, including the *liar para-
dox. The leading logician of this school was Chrysippus,
credited with over 100 works in logic.
There were few developments in logic in succeeding
periods, other than a number of handbooks, summaries,
translations, and commentaries, usually in a simplified
and combined form. The more influential authors include
Cicero, Porphyry, and Boethius in the later Roman
Empire; the Byzantine scholiast Philoponus; and al-
Fa¯ra¯bı¯, Avicenna, and Averroës in the Arab world.
The next major logician known to us is an innovator of
the first rank: Peter Abelard, who worked in the early
twelfth century. He composed an independent treatise on
logic, the Dialectica, and wrote extensive commentaries.
There are discussions of conversion, opposition, quantity,
quality, tense logic, a reduction of de dicto to *de re modal-
ity, and much else. Abelard also clearly formulates several
semantic principles, including the Tarski biconditional for
the theory of truth, which he rejects. Perhaps most impor-
tant, Abelard is responsible for the clear formulation of a
pair of relevance criteria for logical consequences. (*Rele-
vance logic.) The failure of his criteria led later logicians to
reject relevance implication and to endorse material
implication.
Spurred by Abelard’s teachings and problems he pro-
posed, and by further translations, other logicians began
to grasp the details of Aristotle’s texts. The result, coming
to fruition in the middle of the thirteenth century, was the
first phase of *supposition theory, an elaborate doctrine
about the reference of terms in various propositional con-
texts. Its development is preserved in handbooks by Peter
of Spain, Lambert of Auxerre, and William of Sherwood.
The theory of *obligationes, a part of non-formal logic, was
also invented at this time. Other topics, such as the rela-
tion between time and modality, the conventionality of
semantics, and the theory of *truth, were investigated.
The fourteenth century is the apex of medieval logical
theory, containing an explosion of creative work. Suppo-
sition theory is developed extensively in its second phase
by logicians such as William of Ockham, Jean Buridan,
Gregory of Rimini, and Albert of Saxony. Buridan also
elaborates a full theory of consequences, a cross between
entailments and inference rules. From explicit semantic
principles, Buridan constructs a detailed and extensive
investigation of syllogistic, and offers completeness
proofs. Nor is Buridan an isolated figure. Three new liter-
ary genres emerged: treatises on syncategoremata (logical
particles), which attempted to codify their behaviour and
the inferences they license; treatises on sentences, called
‘sophisms’, that are puzzling or challenging given back-
ground assumptions about logic and language; and trea-
tises on insolubles, such as the liar paradox.
The creative energy that drove the logical inquiries of
the fourteenth century was not sustained. By the middle
of the fifteenth century little if any new work was being
done. There were instead many simplified handbooks and
manuals of logic. The descendants of these textbooks
came to be used in the universities, and the great innova-
tions of medieval logicians were forgotten. Probably the
best of these works is the *Port Royal Logic, by Antoine
Arnauld and Pierre Nicole, which was published in 1662.
When writers refer to ‘traditional logic’, they usually have
this degenerate textbook tradition in mind. (*Logic, tradi-
tional.)
Since the beginning of the modern era most of the con-
tributions to logic have been made by mathematicians.
Leibniz envisioned the development of a universal lan-
guage to be specified with mathematical precision. The
syntax of the words is to correspond to the metaphysical
make-up of the designated entities. The goal, in effect, was
to reduce scientific and philosophical speculation to com-
putation. Although this grandiose project was not devel-
oped very far, and it did not enjoy much direct influence,
the Universal Characteristic is a precursor to much of the
subsequent work in mathematical logic.
In the early nineteenth century Bolzano developed a
number of notions central to logic. Some of these, like
analyticity and logical consequence, are seen to be relative
to a collection of ‘variable’ concepts. For example, a
proposition C is a consequence of a collection P of propo-
sitions relative to a group G of variable items, if every
appropriate uniform substitution for the members of G
that makes every member of P true also makes C true. This
may be the first attempt to characterize consequence in
non-modal terms, and it is the start of a long tradition of
characterizing logical notions in semantic terms, using a
distinction between logical and non-logical terminology.
Toward the end of the nineteenth century one can dis-
tinguish three overlapping traditions in the development
of logic. One of them originates with Boole and includes,
among others, Peirce, Jevons, Schröder, and Venn. This
‘algebraic school’ focused on the relationship between reg-
ularities in correct reasoning and operations like addition
and multiplication. A primary aim was to develop calculi
common to the reasoning in different areas, such as propo-
sitions, classes, and probabilities. The orientation is that of
abstract algebra. One begins with one or more systems of
related operations and articulates a common, abstract
structure. A set of axioms is then formulated which is satis-
fied by each of the system. The system that Boole
developed is quite similar to what is now called Boolean
algebra. Other members of the school developed rudimen-
tary *quantifiers, which were sometimes taken to be
extended, even infinitary, conjunctions and disjunctions.
The aim of the second tradition, the ‘logicist school’,
was to codify the underlying logic of all rational, scientific
discourse into a single system. For them, logic is not the
result of abstractions from the reasoning in particular dis-
ciplines and contexts. Rather, logic concerns the most gen-
eral features of actual precise discourse, features
independent of subject-matter.
The major logicists were Russell, the early Wittgen-
stein perhaps, and the greatest logician since Aristotle,
530 logic, history of
Gottlob Frege. In his Begriffsschrift (translated in van Hei-
jenoort (ed.), From Frege to Gödel), Frege developed a rich
formal language with full mathematical rigour. Despite
the two-dimensional notation, it is easily recognized as a
contemporary *Higher-order logic. Quantifiers are
understood as they are in current logic textbooks, not as
extended conjunctions and disjunctions. Unlike the alge-
braists, Frege did not envision various domains of dis-
course, each of which can serve as an interpretation of the
language. Rather, each (first-order) variable is to range
over all objects whatsoever. Moreover, in contemporary
terms, the systems of the logicists had no non-logical ter-
minology.
Frege made brilliant use of his logical insights when
developing his philosophical programmes concerning
mathematics and language. He held that arithmetic and
analysis are parts of logic (*logicism; mathematics, history
of the philosophy of ), and made great strides in casting
number theory within the system of the Begriffsschrift. To
capture mathematical induction, minimal closures, and a
host of other mathematical notions, he developed and
exploited the *ancestral relation, in purely logical terms.
Unfortunately, the system Frege eventually developed
was shown to be inconsistent. It entails the existence of a
concept R which holds of all and only those extensions
that do not contain themselves. A contradiction, known as
*Russell’s paradox, follows.
A major response was the multi-volume Principia
Mathematica, by Russell and Whitehead, which attempts
to recapture the logicist programme by developing an
elaborate theory of *types. (*Higher- order logic.) Antino-
mies are avoided by enforcing a *‘vicious-circle principle’
that no item may be defined by reference to a totality that
contains the item to be defined. Despite its complexity,
Principia Mathematica enjoyed a wide influence among
logicians and philosophers. An elegant version of the the-
ory, called simple type theory, was introduced by Ramsey.
It violates the vicious-circle principle, but still avoids
formal paradox.
The third tradition dates back to at least Euclid and, in
this period, includes Dedekind, Peano, Hilbert, Pasch,
Veblen, Huntington, Heyting, and Zermelo. The aim of
this ‘mathematical school’ is the axiomatization of partic-
ular branches of mathematics, like geometry, arithmetic,
analysis, and set theory. Zermelo, for example, produced
an axiomatization of set theory in 1908, drawing on
insights of Cantor and others. The theory now known as
Zermelo–Fraenkel set theory is the result of some modifi-
cations and clarifications, due to Skolem, Fraenkel, and
von Neumann, among others.
Unlike Euclid, some members of the mathematical
school thought it important to include an explicit formu-
lation of the rules of inference—the logic—in the
axiomatic development. In some cases, such as Hilbert
and his followers, this was part of a formalist philosophical
agenda, sometimes called the Hilbert programme.
(*Formalism.) Others, like Heyting, produced axiomatic
versions of the logic of *intuitionism and intuitionistic
mathematics, in order to contrast and highlight their revi-
sionist programmes (see Brouwer).
A variation on the mathematical theme took place in
Poland under
Łukasiewicz and others. Logic itself became
the branch of mathematics to be brought within
axiomatic methodology. Systems of propositional logic,
modal logic, tense logic, Boolean algebra, and *mereology
were designed and analysed.
A crucial development occurred when attention was
focused on the languages and the axiomatizations them-
selves as objects for direct mathematical study. Drawing
on the advent of non-Euclidean geometry, mathemati-
cians in this school considered alternative interpretations
of their languages and, at the same time, began to consider
metalogical questions about their systems, including
issues of *independence, *consistency, *categoricity, and
*completeness. Both the Polish school and those pursuing
the Hilbert programme developed an extensive pro-
gramme for such ‘metamathematical’ investigation.
(*Metalanguage; *metalogic.) Eventually, notions about
syntax and proof, such as consistency and derivability,
were carefully distinguished from semantic, or model-
theoretic counterparts, such as satisfiability and logical
consequence.
This metamathematical perspective is foreign to the
logicist school. For them, the relevant languages were
already fully interpreted, and were not to be limited to any
particular subject-matter. Because the languages are com-
pletely general, there is no interesting perspective ‘out-
side’ the system from which to study it. The orientation of
the logicists has been called ‘logic as language’, and that of
the mathematicians and algebraists ‘logic as calculus’.
Despite problems of communication, there was signifi-
cant interaction between the schools. Contemporary
logic is a blending of them.
In 1915 Löwenheim carefully delineated what would
later be recognized as the first-order part of a logical sys-
tem, and showed that if a first-order formula is satisfiable
at all, then it is satisfiable in a countable (or finite) domain.
He was firmly rooted in the algebraic school, using tech-
niques developed there. Skolem went on to generalize
that result in several ways, and to produce more enlight-
ening proofs of them. The results are known as the
Löwenheim–Skolem theorems. (*Skolem’s paradox.)
The intensive work on metamathematical problems
culminated in the achievements of Kurt Gödel, a logician
whose significance ranks with Aristotle and Frege. In his
1929 doctoral dissertation, Gödel showed that a given
first-order sentence is deducible in common deductive
systems for logic if and only if it is logically true in the
sense that it is satisfied by all interpretations. This is
known as Gödel’s completeness theorem. A year later, he
proved that for common axiomatizations of a sufficiently
rich version of arithmetic, there is a sentence which is nei-
ther provable nor refutable therein. This is called Gödel’s
incompleteness theorem, or simply *Gödel’s theorem.
The techniques of Gödel’s theorem appear to be gen-
eral, applying to any reasonable axiomatization that
logic, history of 531
includes a sufficient amount of arithmetic. But what is ‘rea-
sonable’? Intuitively, an axiomatization should be effec-
tive: there should be an *algorithm to determine whether
a given string is a formula, an axiom, etc. But what is an
‘algorithm’? Questions like this were part of the motiva-
tion for logicians to turn their attention to the notions of
computability and effectiveness in the middle of the 1930s.
There were a number of characterizations of computabil-
ity, developed more or less independently, by logicians
like Gödel (recursiveness), Post, Church (lambda-defin-
ability), Kleene, Turing (the *Turing machine), and
Markov (the Markov algorithm). Many of these were by-
products of other research in mathematical logic. It was
shown that all of the characterizations are coextensive,
indicating that an important class had been identified.
Today, it is widely held that an arithmetic function is com-
putable if and only if it is recursive, Turing machine com-
putable, etc. This is known as *Church’s thesis.
Later in the decade Gödel developed the notion of set
theoretic constructibility, as part of his proof that the
axiom of *choice and Cantor’s *continuum hypothesis are
consistent with Zermelo–Fraenkel set theory (formulated
without the axiom of choice). In 1963 Paul Cohen showed
that these statements are independent of Zermelo–
Fraenkel set theory, introducing the powerful technique
of forcing. (*Independence.) There was (and is) a spirited
inquiry among set theorists, logicians, and philosophers,
including Gödel himself, into whether assertions like the
continuum hypothesis have determinate truth-values.
(*Continuum problem; *mathematics, problems of the
philosophy of.)
Alfred Tarski, a pupil of
Łukasiewicz, was one of the
most creative and productive logicians of this, or any
other, period. His influence spreads among a wide range
of philosophical and mathematical schools and locations.
Among philosophers, he is best known for his definitions
of *truth and logical consequence, which introduce the
fruitful semantic notion of *satisfaction. This, however, is
but a small fraction of his work, which illuminates the
methodology of deductive systems, and such central
notions as completeness, decidability, consistency, satisfi-
ability, and definability. His results are the foundation of
several ongoing research programmes.
Alonzo Church was another major influence in both
mathematical and philosophical logic. He and students
such as Kleene and Henkin have developed a wide range
of areas in philosophical and mathematical logic, includ-
ing completeness, definability, computability, and a num-
ber of Fregean themes, such as second-order logic and
sense and reference. Church’s theorem is that the collec-
tion of first-order logical truths is not recursive. It follows
from this and Church’s thesis that there is no algorithm for
determining whether a given first-order formula is a logi-
cal truth. Church was a founder of the Association for
Symbolic Logic and long-time guiding editor of the Journal
of Symbolic Logic, which began publication in 1936. Vol-
umes 1 and 3 contain an extensive bibliography of work in
symbolic logic since antiquity.
The development of logic in the first few decades of this
century is one of the most remarkable events in intellec-
tual history, bringing together many brilliant minds work-
ing on closely related concepts.
Mathematical logic has come to be a central tool of con-
temporary analytic philosophy, forming the backbone of
the work of major figures like Quine, Kripke, Davidson,
and Dummett. Since about the 1950s special topics of
interest to contemporary philosophers, such as modal
logic, tense logic, *many-valued logic (used in the study of
*vagueness), *deontic logic, relevance logic, and nonstan-
dard logic, have been vigorously studied. The field still
attracts talented mathematicians and philosophers, and
there is no sign of abatement. p.k.
s.s.
*logic, traditional; logical laws.
I. M. Bochen´ski, A History of Formal Logic, tr. and ed. Ivo Thomas
(New York, 1956).
Alonzo Church, Introduction to Mathematical Logic (Princeton, NJ,
1956).
Martin Davis (ed.), The Undecidable (New York, 1965).
Jean van Heijenoort (ed.), From Frege to Gödel (Cambridge, Mass.,
1967).
William Kneale and Martha Kneale, The Development of Logic
(Oxford, 1962).
Alfred Tarski, Logic, Semantics and Metamathematics, 2nd edn., tr.
J. H. Woodger, ed. John Corcoran (Indianapolis, 1983).
logic, informal. Informal logic examines the nature and
function of arguments in natural language, stressing the
craft rather than the formal theory of reasoning. It supple-
ments the account of simple and compound statements
offered by *formal logic and, reflecting the character of
arguments in natural language, widens the scope to
include inductive as well as deductive patterns of infer-
ence.
Informal logic’s own account of arguments begins with
assertions—the premisses and conclusions—whose rich
meaning in natural language is largely ignored by formal
logic. Assertions have meaning as statements as well as
actions and often reveal something about the person who
makes them. Not least, they are the main ingredient in
patterns of inference. Apart from the crucial action of
claiming statements to be true, the assertions found in an
argument may play other performative roles, such as war-
ranting a statement’s truth (on one’s own authority or that
of another), conceding its truth, contesting it, or—instead
of asserting it at all—assuming the statement as a hypoth-
esis. Assertions also have an epistemic dimension. It is a
convention of natural language (though hardly a universal
truth) that speakers believe what they assert. Appraising
the full meaning of a premiss or conclusion therefore
involves gauging whether the statement was asserted
merely as a belief or, in addition, as an objective fact or
even as an item of knowledge. Finally, assertions have
an emotive side. Few arguments of natural language
are utterly impersonal. Attitudes and feelings seep from
the language of argument and can easily influence what
direction a sequence of reasoning may take. Because
532 logic, history of
informal logic sees assertions and arguments as woven
into the fabric of discourse, the threads it traces are
extremely varied: embedded but possibly incomplete pat-
terns of deductive and non-deductive inference, hidden
assumptions, conversational implications, vagueness,
rhetorical techniques of persuasion, and, of course, fallac-
ies. Such topics, though important for understanding
arguments in natural language, lead it far from the con-
cerns of formal logic. That informal logic lacks the preci-
sion and elegance of a formal theory is hardly surprising,
therefore, but it probably comes as close as any enterprise
ever will to being a science of argumentation. r.e.t.
I. Copi, Informal Logic (New York, 1986).
F. W. Dauer, Critical Thinking: An Introduction to Reasoning
(Oxford, 1989).
logic, intuitionist: see intuitionist logic.
logic, many-valued: see many-valued logic.
logic, modal: see modal logic.
logic, modern. Logic, whether modern or traditional, is
about sound reasoning and the rules which govern it. In the
mid-nineteenth century (say from 1847, the date of Boole’s
book The Mathematical Analysis of Logic), logic began to be
developed as a rigorous mathematical system. Its develop-
ment was soon speeded along by controversies about the
foundations of mathematics. The resulting discoveries are
now used constantly by mathematicians, philosophers, lin-
guists, computer scientists, electronic engineers, and less
regularly by many others (for example, music composers
and psychologists). Gödel’s incompleteness theorem of
1931 was a high point not only for logic but also for twenti-
eth-century culture. Gödel’s argument showed that there
are absolute limits to what we can achieve by reasoning
within a formal system; but it also showed how powerful
mechanical calculation can be, and so it led almost directly
to the invention of digital computers.
Many arguments are valid because of their form; any
other argument of the same form would be valid too. For
example:
Fifty-pence pieces are large seven-sided coins.
This machine won’t take large coins.
Therefore this machine won’t take fifty-pence pieces.
An auk is a short-necked diving bird.
What Smith saw was not a short-necked bird.
Therefore what Smith saw was not an auk.
Both of these arguments can be paraphrased into the
form:
(1) Every X is a Y and a Z.
No Y is a W.
Therefore no X is a W.
(Thus for the first, X = fifty-pence piece, Y = large coin,
Z = seven-sided object, W = thing that this machine will
take.) This form (1) is an argument schema; it has schematic
letters in it, and it becomes an argument when we trans-
late the letters into phrases. Moreover, every argument
got from the schema in this way is valid: the conclusion
(after ‘Therefore’) does follow from the premisses (the
sentences before ‘Therefore’). So we call (1) a valid argu-
ment schema.
Likewise some statements are true purely by virtue of
their form and hence are logically valid. We can write
down a statement schema to show the form, for example:
If p and q then p.
Here the schematic letters p, q have to be translated into
clauses; but whatever clauses we use, the resulting sen-
tence must be true. Such a schema is logically valid; we
can regard it as a valid argument schema with no pre-
misses.
What does it mean to say that a particular argument,
expressed in English, has a particular argument schema as
its form? Unfortunately this question has no exact answer.
As we saw in the examples above, the words in an argu-
ment can be rearranged or paraphrased to bring out the
form. Words can be replaced by synonyms too; an argu-
ment doesn’t become invalid because it says ‘gramo-
phone’ at one point and ‘record-player’ at another. For the
last 100 years or more, it has been usual to split logic into
an exact part which deals with precisely defined argument
schemas, and a looser part which has to do with translat-
ing arguments into their logical *form.
This looser part has been very influential in philosophy.
One doctrine—we may call it the logical form doctrine—
states that every proposition or sentence has a logical
form, and the logical forms of arguments consist of the
logical forms of the sentences occurring in them. In the
early years of the century Russell and Wittgenstein put
forward this doctrine in a way which led to the pro-
gramme of *analytic philosophy: analysing a proposition
was regarded as uncovering its logical form. Chomsky has
argued that each sentence of a natural language has a
structure which can be analysed at several levels, and one
of these levels is called LF for logical form—roughly speak-
ing, this level carries the meaning of the sentence. How-
ever, Chomsky’s reasons for this linguistic analysis have
nothing to do with the forms of valid arguments, though
his analysis does use devices from logic, such as quantifiers
and variables. One can hope for a general linguistic theory
which gives each natural-language sentence a logical form
that explains its meaning and also satisfies the logical form
doctrine; logicians such as Montague and his student
Kamp have made important suggestions in this direction,
but the goal is still a long way off.
Let us turn to the more exact part of logic. Experience
shows that in valid argument schemas we constantly meet
words such as ‘and’, ‘or’, ‘if’; moreover, the sentences can
be paraphrased so that these words are used to connect
clauses, not single words. For example, the sentence
logic, modern 533
Fifty-pence pieces are large seven-sided coins can be
paraphrased as
Fifty-pence pieces are large coins and fifty-pence
pieces are seven-sided.
We can introduce symbols to replace these words, for
example for ‘and’, ∨ for ‘or’, ¬ for ‘it is not true that . . .
’ and
→ for ‘if . . . then’. Unlike the schematic letters, these
new symbols have a fixed meaning and they can be trans-
lated into English. They are known as *logical constants.
Round about 1880 Frege and Peirce independently sug-
gested another kind of expression for use in argument
schemas. We write
∀ x …x…
to mean that ‘…x…’ is true however x is interpreted. The
expression
∀ x can be read as ‘For all x’. For example, the
sentence
Fifty-pence pieces are large seven-sided coins can be
rewritten as
∀ x (if x is a fifty-pence piece then x is a large seven-
sided coin),
or, using the logical constants,
(2)
∀x (x is a fifty-pence piece → (x is a large
coin x is seven-sided) ).
This last sentence says that whatever thing we consider (as
an interpretation for x), if it’s a fifty-pence piece then it’s a
large coin and it’s seven-sided. The symbol x is not a
schematic letter in (2), because the expression
∀x becomes
nonsense if we give x an interpretation. Instead it is a new
kind of symbol which we call a bound variable. The expres-
sion
∀x has a twin, ∃x, which is read as ‘For some x’. These
two expressions are the main examples of logical *quanti-
fiers.
Quantifiers are somewhere between logical constants
and schematic letters. Like logical constants, they do have
a fixed meaning. But this meaning needs to be filled out by
the context, because we need to known what range of
interpretations of the bound variable is allowed. This
range is called the domain of quantification. (Frege assumed
that the domain of quantification is always the class of all
objects. But in practice when we say ‘everybody’ we usu-
ally mean everybody in the room, or all adults of sound
mind, or some other restricted class of people.)
With the help of the symbols described above, we can
translate English sentences into a *formal language. For
example we can translate (2) into
∀x (A(x) → (B(x) C(x))).
Here A, B, and C are schematic letters which need to be
interpreted as clauses containing x, such as ‘x is a fifty-
pence piece’; this is what the (x) in A(x) indicates. The
grammar of this formal language can be written down in a
mathematical form. By choosing a particular set of sym-
bols and saying exactly what range of interpretations is
allowed for the schematic letters and the quantifiers, we
single out a precise formal language, and we can start to
ask mathematical questions about the valid argument
schemas which are expressible in that language.
For example a first-order language is a formal language
built up from the symbols described above, where all
quantifiers are interpreted as having the same domain of
quantification but this domain can be any non-empty set.
First-order logic is logic based on argument schemas writ-
ten in a first-order language.
What is the dividing-line between valid and invalid
argument schemas? There are two main approaches to
this question. In the first approach, which we may call the
rule-based or syntactic one, we suppose that we can intu-
itively tell when a simple argument is valid, just by looking
at it; we count a complicated argument as valid if it can be
broken down into simple steps which we immediately
recognize as valid. This approach naturally leads us to
write down a set of simple valid argument schemas and
some rules for fitting them together. The result will be a
logical *calculus, i.e. a mathematical device for generating
valid argument schemas. The array of symbols written
down in the course of generating an argument schema by
the rules is called a formal proof of the schema.
Once we have a logical calculus up and running, the
mathematicians may suggest ways of revamping it to
make it easier to teach to undergraduates, or faster to run
on a computer. There is a great variety of logical calcu-
luses for first-order logic, all of them giving the same class
of valid argument schemas. Two well-known examples
are the *natural deduction calculus (Gentzen, 1934),
which breaks down complex arguments into intuitively
‘natural’ pieces, and the tableau or truth-tree calculus
(Beth, 1955) which is very easy to learn and can be thought
of as a systematic search for counter-examples (see the
next paragraph).
There is another approach to defining validity, the
semantic approach. In this approach we count an argu-
ment schema as valid precisely if every interpretation
which makes the premisses true makes the conclusion
true too. To phrase this a little differently, a counter-
example to an argument schema is an interpretation which
turns the premisses into true sentences and the conclusion
into a false sentence; the semantic definition says that an
argument schema is valid if and only if it has no counter-
examples.
At first sight this is a very paradoxical definition; it
makes the following highly implausible argument schema
valid just because the conclusion is true whatever we put
for X:
The Emperor Caligula’s favourite colour was X.
Therefore Omsk today is a town in Siberia with a popu-
lation of over a million and a large petroleum industry,
and X = X.
Nevertheless, one can argue that the semantic approach
works if the language of our logic doesn’t contain any
words (such as ‘Omsk’ or ‘today’) that tie us down to spe-
cific features of our world. This is an untidy view, because
the notion of a specific feature of our world is not sharp;
∨
∨
∨
534 logic, modern
should it include the physical laws of the universe, or the
mathematical properties of sets? One has to answer ques-
tions like these in order to draw a line between logical
necessity and other kinds of necessity (physical or mathe-
matical), and probably there will always be philosophical
debate about how best to do this.
For first-order logic the problem happily doesn’t arise.
One can prove that every first-order argument schema
which is justified by any of the standard logical calculuses
is valid in the semantic sense. This is a mathematical theo-
rem, the soundness theorem for first-order logic. Conversely
if an argument schema is not proved valid by the logical
calculuses, then we can show that there is an interpreta-
tion of the schema which makes the premisses true and
the conclusion false. This again is a mathematical theo-
rem, the *completeness theorem for first-order logic (Gödel,
1930; this is quite different from his incompleteness theo-
rem of 1931). The completeness theorem justifies both the
rule-based approach and the semantic one, in the follow-
ing way. The chief danger with the rule-based approach
was that we might have overlooked some rule that was
needed. The completeness theorem assures us that any
schema not justified by our logical calculus would have a
counter-example, so it certainly wouldn’t be valid. And
conversely the chief danger with the semantic approach
was that it might make some argument schema valid for
spurious reasons (like the example with Omsk above).
The completeness theorem shows that if an argument has
no counter-example, then it is justified by the logical cal-
culus. In this way the valid first-order argument schemas
are trapped securely on both sides, so we can be very con-
fident that we have the dividing-line in the right place.
For other logics the position is less clear. For example,
in monadic second-order logic we have some quantifiers
whose domain of quantification is required to be the fam-
ily of subsets of a particular set. Because of this restriction,
some truths of set theory can be expressed as valid
schemas in this logic, and one consequence is that the
logic doesn’t admit a completeness theorem. In temporal
logics there are logical constants such as ‘until’ or ‘it will
sometime be true that . . . ’; to define validity in these log-
ics, we need to decide what background assumptions we
can make about time, for example whether it is continu-
ous or discrete. For these and other logics, the normal
practice today is to give a precise mathematical definition
of the allowed interpretations, and then use the semantic
definition of validity. The result is an exact notion, even if
some people are unhappy to call it logical validity.
This is the place to mention a muddle in some recent
psychological literature. The question at issue is how
human beings carry out logical reasoning. One often reads
that there are two possible answers: (1) by rules as in a log-
ical calculus, or (2) by models (which are interpretations
stripped down to the relevant essentials) as in the seman-
tic approach. This is a confusion. There is no distinction
between rule-based and semantic ways of reasoning. The
rule-based and semantic approaches are different explana-
tions of what we achieve when we do perform a proof: on
the rule-based view, we correctly follow the rules,
whereas on the semantic view we eliminate counter-
examples.
Can we mechanically test whether a given argument
schema is logically valid, and if so, how? For first-order
logic, half of the answer is positive. Given any standard
logical calculus, we can use it to list in a mechanical way all
possible valid argument schemas; so if an argument
schema is valid, we can prove this by waiting until it
appears in the list. In fact most logical calculi do much bet-
ter than this; we can use them to test the schema system-
atically, and if it is valid they will eventually say ‘Yes’.
The bad news is that there is no possible computer pro-
gram which will tell us when a given first-order argument
schema is invalid. This was proved by Church in 1936,
adapting Gödel’s incompleteness theorem. (Strictly it also
needs Turing’s 1936 analysis of what can be done in prin-
ciple by a computer.) This does not mean that there are
some first-order argument schemas which are undecid-
able, in the sense that it’s impossible for us to tell whether
they are valid or not—that might be true, but it would
need further arguments about the nature of human cre-
ativity. Church’s theorem does mean that there is no
purely mechanical test which will give the right answer in
all cases.
A similar argument, again based on Gödel’s incom-
pleteness theorem, shows that for many other logics
including monadic second-order logic, it is not even pos-
sible to list mechanically the valid argument schemas. On
the other hand there are many less adventurous logics—
for example, the logic of Aristotle’s *syllogisms—for
which we have a decision procedure, meaning that we can
mechanically test any argument schema for validity.
A final question: Is there a particular logical calculus
which can be used to justify all valid reasoning (say, in sci-
ence or mathematics)? For the intuitionist school of
Brouwer, it is an article of faith that the answer is ‘No’. On
the other side, Frege believed that he had given a logical
calculus which was adequate at least for arithmetic; but
*Russell’s paradox showed that Frege’s system was incon-
sistent.
For the moment, the heat has gone out of this question.
In modern mathematics we assume that every argument
can be translated into the first-order language appropriate
for set theory, and that the steps in the argument can all be
justified using a first-order logical calculus together with
the axioms of Zermelo–Fraenkel *set theory. This has
become a criterion of sound mathematical reasoning,
though nobody ever carries out the translation in practice
(it would be horrendously tedious). Versions of this trans-
lation are used to check the correctness of computer soft-
ware, for example where lives may depend on it.
There is a more radical reading of our question. In
many situations we carry out reasoning along quite differ-
ent lines from the logical calculuses mentioned above. For
example, when someone pays us money, we normally
take for granted that it is legal tender and not a forgery,
and so when it adds up correctly we infer that we have
logic, modern 535
been given the correct change. Strictly this is not logical
reasoning, because even when the premisses are true, the
conclusion could be false (and occasionally is). But it is rea-
soning of a kind, and it does follow some rules. Logicians
generally disregarded this kind of reasoning until they
found they needed it to guide intelligent databases. For
this purpose a number of non-monotonic logics have been
proposed; the name refers to the fact that in this kind of
reasoning a valid conclusion may cease to be valid when a
new premiss is added (for example, that the five pound
note has no metal strip).
Several other alternative logics have been suggested,
each for its own purposes. Linear logic tries to formalize the
idea that there is a cost incurred each time we use a pre-
miss, and perhaps we can only afford to use it once. An
older example is intuitionist logic (Heyting, 1930), which
incorporates a *verifiability principle: we can’t claim to
have proved that there is an A until we can show how to
produce an example of an A. Each of these logics must be
justified on its own terms. There is no reason to think that
the list of useful logics is complete yet. w.a.h.
*logic, traditional; quantification.
J. C. Beall and Bas C. von Fraassen, Possibilities and Paradox:
An Introduction to Modal and Many-Valued Logic (Oxford,
2003).
H. D. Ebbinghaus, J. Flum, and W. Thomas, Mathematical Logic,
2nd edn. (New York, 1996).
D. Gabbay, Elementary Logics: A Procedural Perspective (London,
1998).
——and F. Guenthner (eds.), Handbook of Philosophical Logic,
2nd edn. in 18 vols. (Dordrecht, 2001– ).
Wilfrid Hodges, Logic, 2nd edn. (London, 2001).
W. H. Newton-Smith, Logic: An Introductory Course (London,
1985).
W. V. Quine, Philosophy of Logic, 2nd edn. (Cambridge, Mass.,
1986).
A Tarski, Introduction to Logic and to the Methodology of Deductive
Sciences, 4th edn. (New York, 1994).
logic, paraconsistent. A logical system is paraconsistent
if it does not sanction the principle that anything follows
from a contradiction. The rejected inference is called ex
falso quodlibet, and is expressed in symbols thus: p, ¬pq.
Paraconsistent logics have application to the logic of
belief, and other propositional attitudes, especially if one
wants to develop something analogous to *possible
worlds semantics. A person who has contradictory beliefs
is not thereby committed to every proposition what-
soever. A ‘world’ that is ‘compatible’ with one’s beliefs
need not be consistent, but it should not trivially make
every proposition true. Other applications of paraconsis-
tent logic concern reasoning with faulty data and
*dialetheism, the view that some contradictions are true.
Dialetheism is one attempt to deal with paradoxes like the
Liar. Most systems of *relevance logic are paraconsistent.
s.s.
Graham Priest, ‘Paraconsistent Logic’, in Dov M. Gabbay and
F. Guenthner (eds.), Handbook of Philosophical Logic, vi, 2nd
edn. (Dordrecht, 2002).
logic, philosophical: see philosophical logic.
logic, relevance: see relevance logic.
logic, second-order. Consider ‘Socrates is wise’. In a first-
order logic the name ‘Socrates’ may be replaced by a
bound variable to yield ‘something is wise’. It is a further
question whether the predicate in this sentence may also
be replaced by a bound variable. A formal logic that per-
mits this replacement is called ‘second-order’. In the stan-
dard semantics for second-order logic, first-order variables
range over a domain of individuals, whereas second-order
variables range over sets, properties, relations, or func-
tions on the range of the first-order variables. So under-
stood, second-order logic is extremely powerful. It is
*incomplete (there can be no finite deductive system in
which every second-order logical truth is deducible), *cat-
egorical (any two models that satisfy a set S of sentences
are isomorphic), and not compact (even if every finite sub-
set of S has a model, S itself may lack a model). In a non-
standard (Henkin) semantics the second-order variables
range over a separate domain of individuals. So under-
stood, second-order logic is complete, categorical, and
compact. f.m
acb.
*higher-order logic; categoricity; incompleteness.
Stewart Shapiro, Foundations without Foundationalism: A Case for
Second-Order Logic (Oxford, 1991).
logic, traditional. The rough-and-ready title given by
later logicians to the methods and doctrines which once
dominated the universities, but which were supplanted in
the twentieth century by the ‘modern’ or ‘mathematical’
logic with which the names of Frege and Russell are espe-
cially associated. Sometimes called ‘Aristotelian’—or ‘syl-
logistic’, or the ‘logic of terms’—it originated with
Aristotle in the fourth century bc, though it acquired a
great many accretions in the intervening 2,000 years.
The older logic was limited, it is customary to say, by
the uncritical assumption that propositions are of the sub-
ject–predicate form. This contention, however, is mis-
leading; not least because the subject–predicate
distinction is actually quite at odds with the formal system
which is supposed to be based on it.
Most traditional logicians certainly accepted that non-
compound propositions invariably contain *subjects and
predicates. At its vaguest, the idea was perhaps that to
make any judgement at all is to say something about
something. It is easy to drift from this to the more specific
doctrine that every proposition contains two distinct ele-
ments: an element which names or refers to something (a
‘subject-term’), and an element (the ‘predicate-term’)
which expresses what is said about it. Thus, in ‘Socrates is
bald’, the name ‘Socrates’ refers to a person, and the
expression ‘is bald’ says something about this person.
The subject of a proposition in this sense—what it is
about—is not part of the proposition but something to
which part of it refers, not the name ‘Socrates’ but the
T
536 logic, modern
person who bears it. If some traditional logicians failed to
stress the difference, this may have reflected uncertainty
about the status of the predicate. The difference between
‘Socrates’ and Socrates is perfectly clear; not quite so clear
is the difference between ‘is bald’ and is bald.
This asymmetry is one aspect of what is really a very
considerable difference: subjects and predicates belong to
quite distinct categories. Granted that an expression like
‘. . . is bald’ plays a predicative role, a subject is anything of
which this may be said. A subject-term is therefore a word
or expression which fulfils two conditions: it constitutes a
grammatical answer to a question like ‘You said that
something (someone) is bald: of what (whom) did you say
this?’ and it must produce good English when it is substi-
tuted for x in ‘x is bald’. Proper names, referring expres-
sions like ‘Plato’s teacher’, and a variety of other items,
satisfy these conditions; but it is obvious that predicative
expressions cannot themselves be subject-terms, because
‘is bald is bald’ makes no sense at all.
The subject–predicate distinction, then, revolves
around the difference between naming or referring to
something and saying something about it. But no such dis-
tinction can sensibly be applied to the traditional system.
The crowning glory of that system, it is agreed on all sides,
is the doctrine of the syllogism. But this doctrine, as we
shall see, requires—as indeed does the rest of the system—
that what is the predicate of one proposition can be the
subject of another.
Traditional logic was for the most part concerned with
the logical properties of four forms of proposition. More
often than not these were said to be
All S is P.
No S is P.
Some S is P.
Some S is not P.
‘All S is P’ was called the ‘universal affirmative’ or ‘A’ form,
‘No S is P’ the ‘universal negative’ or ‘E’ form, ‘Some Sis P’
the ‘particular affirmative’ or ‘I’ form, and ‘Some S is not P’
the ‘particular negative’ or ‘O’ form. That a proposition is
universal or particular was called its quantity, and that it is
affirmative or negative was called its quality.
A moment’s reflection shows that ‘All S is P’ cannot
properly belong in the same list as the rest, because ‘No
Greek is bald’ is good English, while ‘All Greek is bald’ is
merely good gibberish. This drawback, though, could be
remedied simply by taking ‘Every S is P’ to be the correct
form. A more serious problem concerns the innuendo in
the symbolism, which is in any case frankly espoused by
those who use it, that S and P stand for subjects and predi-
cates. If ‘is bald’ is a predicative expression, P clearly can-
not be a predicate in ‘No S is P’, since ‘No Greek is is bald’
looks like a mere typing error.
The stuttering ‘is’ could be removed in one of at least
two ways. One would be to give up the idea that the predi-
cate is ‘is bald’ in favour of saying that it is merely ‘bald’.
This is no doubt the ulterior motive behind the half-baked
suggestion that pro-positions contain a third element, over
and above the subject and the predicate, namely the copula
(i.e. ‘is’). Another way would be to give up the practice of
writing, for example, ‘No S is P’ in favour of ‘No S P’.
But the difficulties do not end there. We have seen that
a subject-term is anything that takes the place of x in an
expression like ‘x is bald’. According to this criterion,
‘Every man’, ‘No man’, and ‘Some man’ are perfectly
good subject-terms. But substituting them in the standard
forms again produces meaningless repetition: ‘Every
every man is bald’, and so on. Again there are two ways of
coping: one is to say that not ‘Every S is P’ but the simple S
P is the correct form, the other that not ‘Every man’ but
merely ‘man’ is the subject-term.
These different ways of coping led our symbolism in
quite different directions. One leaves us with only two
elements (subject and predicate); the other first with three
elements (subject, predicate, copula), then with four (sub-
ject, predicate, copula, and a sign of quantity). All these
distinct, and mutually inconsistent, ways of analysing
propositions are at least hinted at in the traditional text-
books.
As we saw at the outset, the subject–predicate distinc-
tion arises in the context of singular propositions like
‘Socrates is bald’. In the traditional textbooks, singulars
are treated as universals, on the feeble pretext that in
‘Socrates is bald’ the name ‘Socrates’ refers to everything
it can. This notion was generally expressed in technical
terminology: the name was said to be ‘distributed’ or to
‘refer to its whole extension’. These obscurities presum-
ably reflect a disinclination to say something that is obvi-
ously absurd (that one is talking about the whole of
Socrates), something that is obviously false (that only one
person can be called Socrates), or something that is obvi-
ously vacuous (that the name is here meant to name
everyone it is here meant to name). Be that as it may, it is
worth noticing that the singular propositions which are
paradigmatic in the exposition of the subject–predicate
distinction become quite peripheral in the exposition of
the syllogism. What this indicates is that the subject–pred-
icate distinction is merely a nuisance so far as the formal
system of traditional logic is concerned.
How then should the propositions discussed in tradi-
tional logic be symbolized? The only analysis which is
truly consistent with the traditional system is one in which
propositions are treated as containing two distinct sorts of
elements, but these are not subjects and predicates; they
are logical *constants and *terms. The constants, four in
number, are:
‘All . . . are . . . ’ (A)
‘No . . . are . . . ’ (E)
‘Some . . . are . . . ’ (I)
‘Some . . . are not . . . ’ (O)
These are two-place term-operators, which is to say,
expressions which operate on any two terms to generate
propositions.
What are terms? Given our operators and the require-
ment that a term must be capable of filling either place in
logic, traditional 537
them, this question answers itself. A term is typically a
plural noun—like ‘baldies’—or any expression—like ‘per-
sons challenged in the hair department’—that does the
work of an actual or possible plural noun (‘possible’
because any particular language may or may not have a
single word with the same meaning as a complex expres-
sion). Small letters from the beginning of the alphabet will
be used to stand for terms, i.e. as term-variables, and these
will be written after the term-operator. Thus ‘Anyone
who disagrees with me is a complete fool’ is of the form
Aab, where a =‘persons who disagree with me’ and
b =‘ complete fools’.
The traditional system relied upon two kinds of *nega-
tion. The distinction between ‘Not everything which glis-
ters is gold’ (negating a proposition) and ‘Everything
which glisters is not gold’ (negating a term) is worth fight-
ing for, despite the common practice of using the second to
mean the first. Propositional-negation will be represented
by N (meaning ‘It is not that . . . ’); term-negation by n
(meaning ‘non-’). Term-negation may preface either or
both terms. Thus ‘Everything which doesn’t glister is gold’
is Anab, ‘Everything which glisters isn’t gold’) is Aanb, and
‘Everything which doesn’t glister isn’t gold’ is Ananb.
We need in our symbolism also ways of representing
connections between propositions. Aab & Abc will signify
the conjunction of these two propositions. Aab → Anbna
will signify the (in this case true) assertion that the second
proposition follows from the first, and Aab ≡ Aba the (in
this case false) assertion that these two propositions are
equivalent, i.e. that each follows from the other.
The laws of the traditional system may be classified
under two headings: those which apply to only two
propositions, and those which apply to three or more. The
square of opposition and immediate inference fall under
the first heading, syllogisms and polysyllogisms under the
second.
The *square of opposition depicts various kinds of
‘opposition’ between the four propositional forms. A and
E are contraries, meaning that, if a and b stand for the same
terms in Aab and Eab, these two propositions cannot both
be true but may both be false; hence Aab
→ NEab and Eab
→ NAab. I and O are subcontraries, meaning that they can-
not both be false but may both be true; hence NIab
→ Oab
and NOab
→ Iab. A and O are contradictories, as are E and
I, meaning that one of each pair must be true, the other
false; hence Aab ≡ NOaband Eab ≡ NIab. I is subaltern to A,
as O is to E, meaning that in each instance the second
implies the first; hence Aab
→ Iab and Eab → Oab.
*Immediate inference, which consists in drawing a con-
clusion from a single premiss, encompasses conversion,
obversion, contraposition, and inversion. Conversion
consists in reversing the order of terms. It is valid for E and
I, invalid for A and O; hence Eab → Ebaand Iab → Iba. The
valid inferences Eab → Oba and Aab → Iba are called con-
version per accidens. Obversion consists in negating the
second term of a proposition and changing its quality. It is
valid for all four forms; hence Eab → Aanb, Aab → Eanb,
Oab → Ianb, and Iab → Oanb. Contraposition consists in
negating both terms and reversing their order. It is valid
for A and O; hence Aab → Anbna and Oab → Onbna.
Inversion consists in inferring from a given proposition
another having for its subject the negation of the original
subject. It is valid in the following cases: Eab → Inab,
Eab → Onanb, Aab → Onab, and Aab → Inanb.
*Syllogisms draw a conclusion from two premisses.
They contain three terms: one (the middle term) is com-
mon to the premisses, another is common to the conclu-
sion and one of the premisses, and the third is common to
the conclusion and the other premiss. We will use b to sig-
nify the middle term, a and c to signify what are called the
extreme terms. Perhaps the best-known syllogism (it was
called Barbara) may be illustrated by the following simple
example:
Any workers who voted for that party were voting
for their own unemployment.
Those who vote for their own unemployment are
fools to themselves.
Any workers who voted for that party are fools to
themselves.
Traditionally syllogisms were set out this way, with the
conclusion under the premisses like the lines of a sum.
In our symbolism, this example is of the form (Aab &Abc)
→ Aac.
Polysyllogisms have more than two premisses but may
be reduced to a series of conventional syllogisms:
Some university teachers profess to believe in acade-
mic freedom but do nothing to defend it.
Those who profess such a thing but do nothing about
it are not practising what they preach.
Teachers who fail to practise what they preach are a
disgrace to their profession.
Some university teachers are a disgrace to their
profession.
This has the form (Iab &Abc &Acd) → Iad, but it may be
regarded as the summation of two conventional syllo-
gisms, namely (Iab &Abc) → Iac and (Iac & Acd) → Iad.
It is customary to say that there are 256 forms of
syllogism. This number results from a convention con-
cerning how syllogisms are depicted: the order of terms in
the conclusion is fixed, but that in the premisses is
reversable. The conclusion is thus restricted to taking one
of four forms: Eac, Aac, Oac, or Iac. Each premiss, how-
ever, may take any one of eight forms: one is Eab, Eba,
Aab, Aba, Iab, Iba, Oab, or Oba, and the other is Ebc, Ecb,
Abc, Acb, Ibc, Icb, Obc, or Ocb. The number 256 is simply
4 × 8 × 8.
Syllogisms were classified in the traditional textbooks
according to their mood and figure. The mood of a syllo-
gism is essentially the sequence of term-operators it con-
tains. The mood of Barbara, for example, is AAA (hence
the name). The various moods, 64 in all, are there-
fore EEE, EEA, EEO, EEI, and so on. The figure of a syllo-
gism is determined by the arrangement of terms in its
premisses. Aristotle distinguished three figures; later
538 logic, traditional
logicians, whose conception of figure differed significantly
from his, decreed that there are four:
(1) ab, bc.
(2) ab, cb.
(3) ba, bc.
(4) ba, cb.
The identity of a syllogism is completely specified by its
mood and figure, so the number 256 is the product of 4
(figures) and 64 (moods). Of these 256, 24 are said to be
valid (some authors, for reasons that will be indicated in a
moment, say 19, or even 15). Omitting brackets, the 24,
arranged in their figures, are:
(1)
Aab &Abc → Aac Aab &Abc → Iac Iab &Abc → Iac
Aab &Ebc → Eac Aab &Ebc → Oac Iab &Ebc → Oac
(2)
Aab &Ecb → Eac Aab &Ecb → Oac Eab &Acb → Eac
Eab &Acb → Oac Iab &Ecb → Oac Oab &Acb→Oac
(3)
Aba &Abc → Iac Aba &Ibc →
Iac Iba &Abc → Iac
Aba &Ebc → Oac Aba &Obc → Oac Iba &Ebc → Oac
(4)
Aba &Acb → Iac Eba &Acb → Eac Eba &Acb → Oac
Aba &Ecb → Oac Aba &Icb → Iac Iba &Ecb → Oac
Of these, five are ‘weakened’, meaning that they draw par-
ticular conclusions from premisses that merit a universal
one. If these are omitted, the number of valid forms is 19.
Among these 19, 15 either draw a universal conclusion
from universal premisses or a particular conclusion from
one universal and one particular premiss: these were
sometimes called ‘fundamental’.
But the convention behind the numbers given in the
traditional textbooks is wholly improper. The effect of
reversing the order of terms in E and I propositions is to
produce mere equivalents, while in A and O non-
equivalents are produced. The textbook account there-
fore includes duplication. It excludes from the syllogism,
moreover, the varieties of negation that are permitted
in immediate inferences, and is as a consequence
incomplete.
The traditional system encompassed what were really
eight logically distinct propositional forms:
Eab (Eba, etc.).
Enab (Aba, etc.).
Eanb (Aab, etc.).
Enanb (Anab, etc.).
NEab (Iab, etc.).
NEnab (Oba, etc.).
NEanb (Oab, etc.).
NEnanb (Onab, etc.).
Any one of these eight forms is expressible in eight
ways. Eab, for example, is equivalent to Eba, Aanb, Abna,
NIab, NIba, NOanb and NObna. A proper account of
the syllogism, then, would cover 64 forms of proposition:
the correct number of syllogisms is therefore 262,144
(i.e. 64 × 64 × 64). c.w.
P. T. Geach, ‘History of the Corruptions of Logic’, in Logic Matters
(Oxford, 1972).
J. N. Keynes, Formal Logic, 4th edn. (London, 1906).
J. Łukasiewicz, Aristotle’s Syllogistic, 2nd edn. (Oxford, 1957).
A. N. Prior, Formal Logic, 2nd edn. (Oxford, 1962), pt. 2, ch. 6.
C. Williamson, ‘Traditional Logic as a Logic of Distribution-
Values’, Logique et analyse (1971).
—— ‘How Many Syllogisms Are There?’, History and Philosophy
of Logic (1988).
logic of discovery. *Deduction in the testing of scientific
theories. For example, the exhibiting of logical relations
between the sentences of a theory (such as equivalence,
derivability, consistency, inconsistency) or between a
theory and estabilished theories; the logical inferring of
predictions from a theory. *Popper argues against the
view that scientific theories are conclusively inductively
verifiable but argues for their deductive and empirical
falsifiability. A claim of the form (
∀a) (Fa), ‘Every a is F’,
cannot be confirmed by any finite number of observations
of a’s that are F, because there could always in principle
exist an undiscovered a that is not F, but (
∀a)(Fa) can be
refused by the discovery of just one a that is not F.
*Popper has an evolutionary epistemology of scientific
discovery. The formulation of theories is analogous to
genetic mutation in evolutionary theory. Theories and
mutations arise randomly as putative solutions to
environmental problems, and only those conducive to the
survival of the species in that environment themselves
survive through trial and error. Popper adopts a Platonist
view of logic, on the grounds that proofs are (sometimes
surprising) discoveries not unsurprising inventions. s.p.
Karl R. Popper, The Logic of Scientific Discovery (London, 1980),
ch. 1, sect. 3.
logical atomism: see atomism, logical.
logical constants. An argument’s logical form is shown
by analysing its constituent propositions into constant and
variable parts, constants representing what is common to
proportions, variables their differing content. The con-
stants peculiar to syllogistic logic are ‘All … are …’, ‘No …
are …’, ‘Some … are …’, Some … are not …’; those of
propositional calculus are truth-functional connectives
like implication, conjunction, and disjunction; those of
predicate calculus add the quantifiers ‘For all x …’ and
‘There is an x such that …’. Constants concerning identity,
tense, modality, etc. may be introduced in more complex
systems. c.w.
logical determinism: see determinism, logical.
logical empiricism: see empiricism, logical.
logical form: see form, logical.
logical form 539