Mathematics: Quantifying Aristotle's Fallacies
Mathematics: Quantifying Aristotle's Fallacies
Article
Quantifying Aristotle’s Fallacies
Evangelos Athanassopoulos 1, * and Michael Gr. Voskoglou 2
1 Independent Researcher, Giannakopoulou 39, 27300 Gastouni, Greece
2 Department of Applied Mathematics, Graduate Technological Educational Institute of Western Greece,
22334 Patras, Greece; [email protected] or [email protected]
* Correspondence: [email protected]
Received: 20 July 2020; Accepted: 18 August 2020; Published: 21 August 2020
Abstract: Fallacies are logically false statements which are often considered to be true. In the
“Sophistical Refutations”, the last of his six works on Logic, Aristotle identified the first thirteen of
today’s many known fallacies and divided them into linguistic and non-linguistic ones. A serious
problem with fallacies is that, due to their bivalent texture, they can under certain conditions
disorient the nonexpert. It is, therefore, very useful to quantify each fallacy by determining the
“gravity” of its consequences. This is the target of the present work, where for historical and practical
reasons—the fallacies are too many to deal with all of them—our attention is restricted to Aristotle’s
fallacies only. However, the tools (Probability, Statistics and Fuzzy Logic) and the methods that we use
for quantifying Aristotle’s fallacies could be also used for quantifying any other fallacy, which gives
the required generality to our study.
Keywords: logical fallacies; Aristotle’s fallacies; probability; statistical literacy; critical thinking; fuzzy
logic (FL)
1. Introduction
Fallacies are logically false statements that are often considered to be true. The first fallacies
appeared in the literature simultaneously with the generation of Aristotle’s bivalent Logic. In the
“Sophistical Refutations” (Sophistici Elenchi), the last chapter of the collection of his six works on
logic—which was named by his followers, the Peripatetics, as “Organon” (Instrument)—the great
ancient Greek philosopher identified thirteen fallacies and divided them in two categories, the linguistic
and non-linguistic fallacies [1].
The research on logical fallacies was reanimated during the later Middle Ages (1300–1500 AD)
with the establishment of the first Universities, where the study of Aristotle’s Logic was one of the
first priorities. Many of the now-existing fallacies took their Latin names at that time. A long period
of reduced interest on the subject followed. However, after the end of the Second World War, and in
particular after 1970, the interest in fallacies was renewed with the addition of the cognitive biases
(prejudices) and other distortions of logic. In addition to Aristotle’s fallacies, many other fallacies are
known today. A list of the most important of them is given in [2], while many fallacies are analyzed in
the first of the present authors’ books [3].
Because of their variety of structure and applications, it is difficult to classify the fallacies so as to
satisfy all practitioners. A standard way of classifying them into formal (with deductive arguments
that they are false) and informal fallacies is according to their structure or content [4].
Another big problem with fallacies is that, due to their bivalent texture, they can under certain
conditions disorient the nonexpert. This explains the frequent use of fallacies as rhetorical devices in the
desire to persuade when the focus is more on communicating and gaining common agreement rather
than on the correctness of reasoning. It is, therefore, useful to quantify the fallacies by determining the
“gravity” of their consequences. In the present work, our attention is focused on Aristotle’s fallacies,
but the tools and the methods used for quantifying them could be also used for any other logical fallacy.
The rest of the article is organized as follows: In Section 2 Aristotle’s thirteen fallacies are presented
and described in brief, while their quantification is attempted in Section 3. Representative examples
are also presented. A general discussion follows in Section 4, and the article closes with the final
conclusions, which are stated in Section 5.
11. Wicked circle: Circular reasoning to prove the assumed premise. Hence the premise is not really
proven by the argument. Example: God exists because the Bible says so and the Bible is true
because it comes from God.
12. False cause: A causal relationship between two facts without proof that it actually exists. Example:
Money makes people happy (not all people and not always just money).
13. Affirming the consequent: This assumes that that given an “if A then B” argument, you can also
invert it (false inversion). The A part of such a statement is called the “antecedent” and the B part
is called the “consequent”. Example: All cats have four feet, therefore all the animals having four
feet are cats.
ideas of Janes, like the expert in Algebraic Geometry David Mumford, who believes that Probability
and Statistics are emerging as a better way of building scientific models [7]. Probability and Statistics
are related mathematical topics that have, however, fundamental differences. In fact, Probability is
a branch of theoretical mathematics dealing with the estimation of the likelihood of future events,
whereas Statistics is an applied branch, which tries to make sense by analyzing the frequencies of
past events.
Nevertheless, both Probability and Statistics have been developed on the basis of the principles of
bivalent logic. As a result, they are effectively only tackling cases of uncertainty existing in the real
world that are due to randomness, and not those due to imprecision [8]. For example, the expression
“The probability that Mary is a clever person is 75%” means that Mary is, according to Aristotle’s
law of the excluded middle, either a clever or not a clever person. However, her outlines (heredity,
educational background, etc.) suggest that the probability of being a clever person is high. The problem
here is that there is not an exact criterion available to the observer (e.g., IQ index) enabling him to
decide definitely whether or not Mary is a clever person. In such cases Fuzzy Logic (FL), introduced
during the 1970s [9], comes to bridge the existing gap.
Multi-valued logics, challenging the law of the excluded middle, have been systematically
proposed previously by Lukasiewicz (1878–1956) and Tarski (1901–1983), although their ideas can
already be traced in the philosophical beliefs of the Ionian Greek philosopher Heraclitus (535–475 BC),
who spoke about the “harmony of the opposites” and the Gautama Buddha, who lived in India
around 500 BC. Those followed by Plato (427–377 BC), who used to be the teacher of Aristotle, and by
several other, more recent, philosophers, like Hegel, Marx, Engels, etc. (see [10], Section 2). However,
the electrical engineer of Iranian origin Lofti Zadeh, Professor of Computer Science at the University
of Berkeley, California, was the first to mathematically formulate the infinite-valued FL through the
notion of the fuzzy set (FS) that assigns membership degrees (degrees of truth) in the real interval [0, 1]
to all elements of the universal set [11].
Probabilities and membership degrees, although both are defined in the same interval [0, 1],
are essentially different from each other. For example, the expression “Mary’s membership degree in
the FS of the clever persons is 0.75”, means that Mary is a rather clever person. However, all people
belong to the FS of clever persons with membership degrees varying from 0 (stupid) to 1 (genius)!
A disadvantage of FL is that the definition of the membership function of a FS, although it must
always be based on logical arguments, is not uniquely determined depending on the observer’s
personal criteria and goals. This was the reason of a series of generalizations and related theories
that followed the introduction of FL [12]. All those theories together form an effective framework for
tackling all the forms of uncertainty existing in the real world and science, although none of them
has been proved suitable for solving all the related problems alone. Statistical data or probability
distributions can be used in certain cases to define membership degrees, but this is not the rule in
general. This will become evident in the rest of the paper through our efforts to quantify the inferences
of Aristotle’s fallacies starting from his non-linguistic fallacies.
fallacy is 97%. Consequently, it is crucial for people to avoid hasty generalizations, but at the same
time, they must be careful about unqualified generalizations. Those two fallacies must be examined
simultaneously in order to make the right decision.
The cultivation of statistical literacy is very important, but it alone is not enough; it must be
combined with critical thinking. The great ancient Greek philosopher Socrates (470–399) in his dialogue
with Euthydemus—which was written by his student Plato in 384 BC, i.e., the year of Aristotle’s birth—
tacitly exploited the dicto simpliciter to give the following important example about the importance of
critical thinking in decision-making.
Socrates asked his friend Euthydemus if he thinks that cheating is immoral. Of course it is,
answered Euthydemus. However, what happens, replied Socrates, if your friend, feeling terrible,
wants to commit suicide and you steal his knife? There is no doubt that you cheat him in that case,
but is this immoral? No, said the embarrassed Euthydemus [13]. Here Euthydemus followed the
statistical way of thinking, since in most cases cheating is considered to be an immoral action. Socrates,
however, taught him to combine it with critical thinking. It is recalled that critical thinking is considered
to be a higher mode of thinking by which the individual transcends his subjective self in order to arrive
rationally at conclusions substantiated using valid information (see [14], Section 3). Through critical
thinking, reasoning skills such as analysis, synthesis and evaluation are combined, giving rise to other
skills like inferring, estimating, predicting, generalizing, problem solving, etc. [15].
Note also that the dialogue of Socrates with Euthydemus introduces indirectly the fallacy of
“the purpose justifies the means”. In Socrates’ example, the stealing of the knife (means) was moral,
since it could save a life (purpose). Stealing, however, for your own profit is an immoral action.
The nature of the fallacies of morality in general require the help of FL for their quantification [16].
Let us now transfer the dialogue of Socrates with Euthydemus to the previous case with the two
parents. Imagine that Socrates (if he were alive at that time) met parent B downtown and asked him:
if your child has a particular interest in the lessons taught by the three bad teachers and he is not
interested in the lessons taught by the 97 good teachers, is your decision to choose this school right for
his future? After this, parent B becomes puzzled and thinks that he should reconsider his decision
after discussing it with his child.
In conclusion, these two of Aristotle’s statistical fallacies are connected to the error created by
inductive reasoning [17]. Therefore, quantifying the gravity of those fallacies, one actually quantifies
the inductive error. Nevertheless, the error of dicto simpliciter is much less than that of the secundum
quid, so that many people consider the former as not being actually a fallacy. On the contrary, the latter
is a serious fallacy caused by the lack of statistical literacy and must be avoided in all cases.
(see Section 2.3), one could bring an elephant to the square to stand on the white powder and observe
if it will go away or not. This could be repeated several times in order to obtain conclusions about the
effectiveness or otherwise of the white powder. A similar procedure is usually followed for testing the
effectiveness of a new medicine.
In other cases, however, things are more complicated. Consider for example the case of an
experimental school, where a continuous selection of both teachers and students is made. Everyone
with non-satisfactory performance is replaced. The quality teachers increase the level and interest of
the students; therefore, student demand also increases. This forces teachers to improve their teaching
methods even more, which causes a further improvement of students and so on. Finally, why is this
school a good school? Because of having good students or good teachers? In other words which is the
cause and which is the effect? It is almost impossible to give a definite answer to this question.
The general form of the fallacy of false inversion (no. 13) is: “If A then B” implies that “If B
then A”, where A = the cause and B = the effect. To quantify this fallacy, a shift is needed from
the Aristotelian logic to Bayesian Reasoning, because its degree of truth is equal to the conditional
probability P(A/B). Then the Bayes’ formula [18] gives that
P(B/A)P(A)
P(A/B) = (1)
P(B)
In the example of this fallacy presented in Section 2.2, we have that A = cats and B = animals having
P(A)
four feet; therefore, P(B/A) = 1. Consequently, Equation (1) gives that P(A/B) = P(B) . For example,
if on a farm there are 100 animals in total, 75 of them having four feet (e.g., cats, dogs, goats, cows
and horses), including three cats, and the rest of them have two feet (e.g., chickens), then P(A) = 100 3
,
P(B) = 100 and P(A/B) = 75 = 0.04. Therefore, the degree of truth of the fallacy in this case is only 4%.
75 3
Nevertheless, in many cases, the conditional probability P(B/A) is not equal to 1. Consider, for
example, that A = I have flu and B = I feel pain in my throat. Assume that on a winter day 30% of
the inhabitants of a village feel pain in their throats and that 25% of the inhabitants have flu. Assume
further that the existing statistical data show that 70% of those having flu feel pain in their throats.
Then Equation (1) gives that P(A/B) = 0.7×0.25
0.3 ' 0.583, or 58.3%.
The fallacy of false inversion is also connected to the credibility of medical tests. Assume,
for example, that Mr. G lives in a city where 2% of the inhabitants have been infected by a dangerous
virus. Mr. X does a test for the virus, whose statistical accuracy is 95%. The test is positive. What is the
probability of Mr. X being a carrier of the virus?
To answer this question, let us consider the events: A = The subject is a carrier of the virus and
B = The test is positive. According to the given data, we have that P(A) = 0.02 and P(B/A) = 0.95.
Furthermore, assuming that 100 inhabitants of the city do the test, we should have on average
2 × 95% = 1.9 positive results from the carriers and 98 × 5% = 4.9 from the noncarriers of the
virus. Therefore P(B) = 0.068. Replacing the values of those probabilities in Equation (1), one finds
P(A/B) ≈ 0.2794. Therefore, the probability of Mr. X being a carrier of the virus is only 27.94% and not
95%, as could be thought after a rough first estimation!
It is worth noting that the only information given within the premises of bivalent logic about
this fallacy is that the inversion between cause and effect is false, or otherwise that the conditional
probability P(A/B) is not equal to 1. However, this information is useless in practice, when one wants
to know “what is” (via positiva) and not “what is not” (via negativa). The latter, for example, is a
method that has been followed by religion when failing to define “what is God”. It was decided then
to define instead “what is not God” (Cataphatic and Apophatic Theologies), which is much easier.
From the beginning of the 19th century, several researchers in the area of bivalent logic (Bantham,
Hamilton, De Morgan, Frege, etc.), in their effort to improve the quality of the bivalent inferences,
introduced the universal (∀) and the existential (∃) quantifiers. In this way, the false inversion becomes
valid by saying, for example, “There exist animals with four feet which are cats”, or “Some of the brain
Mathematics 2020, 8, 1399 7 of 10
mechanisms are Bayesian, but it has not been proved that all of them (even the cognitive ones) are” but
the information given by this modified expression still remains very poor.
influence. Examples of cognitive biases include racism, nationalism, religious, linguistic, sexual or
neurological discrimination, sexism, etc. [20]. The Israeli psychologist and Nobel prize winner in
Economics (2002) Daniel Kahneman with his collaborator Amos Tversky contributed significantly to
the study of the cognitive biases related to Economics [21]. The fact that Kahneman is a (the only) Nobel
laureate in Economics who is a psychologist emphasizes the useful role of psychology in quantifying
the cognitive fallacies and the fuzziness of human reasoning.
In general, too many sources of fuzziness exist in real life, creating several types of fallacies,
such as, for example, all adjectives and adverbs in the natural language. There is obviously a need for
determining the gravity of the consequences of all those fallacies in a way analogous to the Aristotle’s
fallacies, which is a good proposal for future research.
4. Discussion
The quantification of fallacies is very important in everyday life, where people want to know not
simply whether something is true or false, but actually the degree of its truth. Nevertheless, as has
been illustrated by the present study, the latter cannot always be achieved with the help of bivalent
logic. One could think about the role of logic in such cases in terms of a new plot. The plot has to be
fenced first (bivalent logic), and then you can watch what happens inside it (FL).
FL does not oppose bivalent logic; on the contrary it extends and complements it [22–24]. The fact
that FL sometimes uses statistical data or probability distributions to define membership degrees does
not mean that it “steals” ideas and methods from those topics. As we saw in Section 3.1, probabilities
and membership degrees are completely different concepts. In addition, FL frequently uses other
innovative techniques, like linguistic variables, the calculus of fuzzy if–then rules, etc.
In an earlier work [17], we provided full evidence that scientific progress is due to the collaboration
of these two equally valuable types of logic. This collaboration is expressed in everyday life by the
method of trial and error and in the human way of thinking through inductive and deductive reasoning.
Inductive reasoning always precedes the tracing of a new scientific idea, while deductive reasoning
only guarantees the validity and correctness of the corresponding theory on the basis of the axioms on
which it has been built. In addition, whereas deduction is purely based on the principles of bivalent
logic, FL, rejecting the principle of the excluded middle, marks out the real value of induction, which is
disregarded by bivalent logic.
Another important point that was illustrated in Section 3.2 of the present study is the essential
role of the conditional probabilities for quantifying the fallacies of cause and effect. The Bayes’
rule—Equation (1)—connects the conditional probability P(A/B) to the inverse of time conditional
probability P(B/A) in terms of the prior probability P(A) and the posterior probability P(B). Thus,
by changing the value of the prior probability P(A), one obtains different values for the conditional
probability P(A/B), representing in this study the degrees of truth of the corresponding fallacy.
The amazing thing, however, is that, although probabilities in general and conditional probabilities
in particular have been defined and developed on the basis of the principles of bivalent logic, the change
of the values of the prior probability P(A) provides multiple values for the conditional probability
P(A/B), introducing in this way a multi-valued logic! Consequently, one could argue that the conditional
probabilities—often called Bayesian probabilities as well—constitute an interface between bivalent
and fuzzy logic.
At first glance, Bayes’ rule is an immediate consequence of the basic theorem calculating the
P(A ∩ B) P(A ∩ B)
value of a conditional probability. In fact, we have that P(A/B) = P(B)
and P(B/A) = P(A)
or
P(B/A)P(A)
P(A ∩ B) = P(B/A) P(A), which gives that P(A/B) = P(B)
.
However, the consequences of this simple rule have been proved to be very important for all
of science, while recent research gives evidence that even the mechanisms under which the human
brain works are Bayesian! [18,25]. As seen in [17], the validation of any scientific theory T can be
expressed by a deductive argument of the form “If H, then T”, where H represents the premises of T
(observations, intuitive conclusions, axioms on which T has been built, etc.), which have been obtained
Mathematics 2020, 8, 1399 9 of 10
by inductive reasoning. Therefore, the inductive error is transferred through H to the deductive
argument. Consequently, the conditional probability P(T/H) expresses the degree of truth of the theory
T. Thus, Sir Harold Jeffreys’—the British mathematician who played an important role in the revival of
the Bayesian view of probability—characterization of the Bayesian rule as the “Pythagorean Theorem
of Probability Theory” [26] is fully justified.
5. Conclusions
The highlights of the present work can be summarized as follows:
• A deep wisdom must be attributed Aristotle for introducing the logical fallacies. The description
of his statistical fallacies was particularly impressive cular, because at that time Statistics was a
completely unknown concept.
• Aristotle’s fallacies and all the other fallacies of bivalent logic contain very poor information
about the gravity of their consequences, which can be enriched by statistical and critical thinking,
as some textbooks in logic suggest (e.g., [27]).
• Probability and Statistics are able to quantify, i.e., to calculate the degree of truth, of the statistical
fallacies and of the fallacies of cause and effect. The Bayesian probabilities, in particular, which have
been proved to be very important for all of science and human cognition, play an essential role
in quantifying the fallacies of cause and effect. The fuzziness of the linguistic fallacies, however,
cannot be handled by probabilistic and statistical methods. In fact, innovative methods of FL,
like the use of linguistic variables, the calculation of fuzzy if–then rules, etc., must be used to
quantify those fallacies. It is of worth noting that in certain cases (e.g., figures of speech) the
degree of truth of the corresponding fallacy is indeterminate.
• The fact that FL sometimes uses statistical data or probability distributions to define membership
degrees does not mean that it “steals” ideas and methods from those topics. In fact, although
probabilities and membership degrees function in the same interval [0, 1], they are completely
different concepts. FL does not oppose bivalent logic; on the contrary, it extends and complements
it. The whole of human scientific progress is due to the collaboration of these two types of logic.
Author Contributions: E.A., methodology, formal analysis, resources, visualization; M.G.V., writing—original
draft preparation, conceptualization, resources, data curation, visualization. All authors have read and agreed to
the published version of the manuscript.
Funding: This research received no external funding.
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Changingminds.org. Full Alphabetic List of Fallacies. Available online: http://changingminds.org/disciplines/
argument/fallacies/fallacies_alpha.htm (accessed on 23 June 2020).
2. Wikipedia.org. List of Fallacies. Available online: https://en.wikipedia.org/wiki/List_of_fallacies (accessed
on 25 June 2020).
3. Athanassopoulos, E. The Fallacies of Logic and the Cognitive Biases; Self-Edition: Patras, Greece, 2019. (In Greek)
4. Pirie, M. How to Win Every Argument: The Use and Abuse of Logic, 2nd ed.; Bloombury Publishing: London,
UK; New Delhi, India; New York, NY, USA; Sydney, Australian, 2015.
5. Dobbeli, R. Die Kunst Des Klaren Denkens; Hanser Publications: Cincinnati, OH, USA, 2011.
6. Janes, E.T. Probability Theory: The Logic of Science, 8th ed.; Cambridge University Press: Cambridge, UK, 2011.
7. Mumford, D. The dawing of the age of stochasticity. In Mathematics: Frontiers and Perspectives; Amoid, V.,
Atiyah, M., Laxand, P., Mazur, B., Eds.; AMS: Providence, RI, USA, 2000; pp. 197–218.
8. Kosko, B. Fuzzy Thinking: The New Science of Fuzzy Logic; Hyperion: New York, NY, USA, 1993.
9. Zadeh, L.A. Outline of a new approach to the analysis of complex systems and decision processes. IEEE Trans.
Syst. Man Cybern. 1973, 3, 28–44. [CrossRef]
Mathematics 2020, 8, 1399 10 of 10
10. Voskoglou, M.G. Methods for assessing human-machine performance under fuzzy conditions. Mathematics
2019, 7, 230. [CrossRef]
11. Zadeh, L.A. Fuzzy sets. Inf. Control 1965, 8, 338–353. [CrossRef]
12. Voskoglou, M.G. Generalizations of fuzzy sets and related theories. In An Essential Guide to Fuzzy Systems;
Voskoglou, M., Ed.; Nova Science Publishers: New York, NY, USA, 2019; pp. 345–352.
13. Warburton, N. A Little History of Philosophy; Yale University Press: New Haven, CT, USA, 2011.
14. Voskoglou, M.G.; Buckley, S. Problem solving and computers in a learning environment. Egypt. Comput.
Sci. J. 2012, 36, 28–46.
15. Halpern, D. Thought and Knowledge: An Introduction to Critical Thinking, 4th ed.; Earlbaum: Mahwah, NJ,
USA, 2003.
16. Athanassopoulos, E. The Fallacies of Morality; Self-Edition: Patras, Greece, 2020. (In Greek)
17. Athanassopoulos, E.; Voskoglou, M.G. A philosophical treatise on the connection of scientific reasoning with
fuzzy logic. Mathematics 2020, 8, 875. [CrossRef]
18. Hoghan, J. Bayes’ Theorem: What Is the Big Deal? January 2015. Available online: http//:blogs.
scientificamerican.com/cross-check/bayes-s-theorem-what-s-the-big-deal (accessed on 15 April 2020).
19. Gardener, J.; Resnik, D. The misuse of statistics, concepts, tools and a research agenda. Account. Res. Policies
Qual. Assur. 2002, 9, 65–74. [CrossRef] [PubMed]
20. Allport, G.W. The Nature of Prejudice; Perseus Books Publishing: New York, NY, USA, 1979.
21. Kahneman, D. Thinking Fast and Slow; Strauss and Giroux: New York, NY, USA, 2013.
22. Zadeh, L.A. Discussion: Probability theory and fuzzy logic are complementary rather than competitive.
Technometrics 1995, 37, 271–276. [CrossRef]
23. Kosko, B. Fuzziness vs. probability. Int. J. Gen. Syst. 1990, 17, 211–240. [CrossRef]
24. Ross, T.J. Fuzzy Logic with Engineering Applications, 2nd ed.; Wiley: Hoboken, NJ, USA, 2004.
25. Bertsch McGrayne, S. The Theory That Would Not Die; Yale University Press: New Haven, CT, USA; London,
UK, 2012.
26. Jeffreys, H. Scientific Inference, 3rd ed.; Cambridge University Press: Cambridge, UK, 1973.
27. Hurley, P.I.; Watson, L. A Concise Introduction to Logic, 13th ed.; Cengage Learning: Boston, MA, USA, 2018.
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).