Inductive reasoning (as opposed to deductive reasoning or abductive reasoning) is a method of reasoning in which the premises are viewed as supplying some evidence for the truth of the conclusion. While the conclusion of a deductive argument is certain, the truth of the conclusion of an inductive argument may be probable, based upon the evidence given.
Many dictionaries define inductive reasoning as the derivation of general principles from specific observations, though some sources disagree with this usage.
The philosophical definition of inductive reasoning is more nuanced than simple progression from particular/individual instances to broader generalizations. Rather, the premises of an inductive logical argument indicate some degree of support (inductive probability) for the conclusion but do not entail it; that is, they suggest truth but do not ensure it. In this manner, there is the possibility of moving from general statements to individual instances (for example, statistical syllogisms, discussed below).
Inductive reasoning is inherently uncertain. It only deals in degrees to which, given the premises, the conclusion is credible according to some theory of evidence. Examples include a many-valued logic, Dempster-Shafer theory, or probability theory with rules for inference such as Bayes' rule. Unlike deductive reasoning, it does not rely on universals holding over a closed domain of discourse to draw conclusions, so it can be applicable even in cases of epistemic uncertainty (technical issues with this may arise however; for example, the second axiom of probability is a closed-world assumption).
An example of an inductive argument:
This argument could have been made every time a new biological life form was found, and would have been correct every time; however, it is still possible that in the future a biological life form not requiring liquid water could be discovered.
As a result, the argument may be stated less formally as:
Unlike deductive arguments, inductive reasoning allows for the possibility that the conclusion is false, even if all of the premises are true. Instead of being valid or invalid, inductive arguments are either strong or weak, which describes how probable it is that the conclusion is true. Another crucial difference is that deductive certainty is impossible in non-axiomatic systems, such as reality, leaving inductive reasoning as the primary route to (probabilistic) knowledge of such systems.
Given that "if A is true then that would cause B, C, and D to be true", an example of deduction would be "A is true therefore we can deduce that B, C, and D are true". An example of induction would be "B, C, and D are observed to be true therefore A might be true". A is a reasonable explanation for B, C, and D being true.
Note however that this is not necessarily the case. Other events also coincide with the extinction of the non-avian dinosaurs. For example, the Deccan Traps in India.
A classical example of an incorrect inductive argument was presented by John Vickers:
The correct conclusion would be, "We expect that all swans are white".
The definition of inductive reasoning described in this article excludes mathematical induction, which is a form of deductive reasoning that is used to strictly prove properties of recursively defined sets. The deductive nature of mathematical induction is based on the non-finite number of cases involved when using mathematical induction, in contrast with the finite number of cases involved in an enumerative induction procedure with a finite number of cases like proof by exhaustion. Both mathematical induction and proof by exhaustion are examples of complete induction. Complete induction is a type of masked deductive reasoning.
For a move from particular to universal, Aristotle in the 300s BCE used the Greek word epagogé, which Cicero translated into the Latin word inductio. In the 300s CE, Sextus Empiricus maintained that all knowledge derives from sensory experience--concluded in his Outlines of Pyrrhonism that acceptance of universal statements as true cannot be justified by induction.
In 1620, early modern philosopher Francis Bacon repudiated mere experience and enumerative induction, and sought to couple those with neutral and minute and many varied observations before[further explanation needed] to uncover the natural world's structure and causal relations beyond the present scope of experience via his method of inductivism, which nonetheless required enumerative induction as a component.
The supposedly radical empiricist David Hume's 1740 stance found enumerative induction to have no rational, let alone logical, basis but to be a custom of the mind and an everyday requirement to live, although observations could be coupled with the principle uniformity of nature--another logically invalid conclusion, thus the problem of induction--to seemingly justify enumerative induction and reason toward unobservables, including causality counterfactually, simply that[further explanation needed] modifying such an aspect prevents or produces such outcome.
Awakened from "dogmatic slumber" by a German translation of Hume's work, Kant sought to explain the possibility of metaphysics. In 1781, Kant's Critique of Pure Reason introduced the distinction rationalism, a path toward knowledge distinct from empiricism. Kant sorted statements into two types. The analytic are true by virtue of their terms' arrangement and meanings--thus are tautologies, merely logical truths, true by necessity--whereas the synthetic arrange meanings to refer to states of facts, contingencies. Finding it impossible to know objects as they truly are in themselves, however, Kant found the philosopher's task not peering behind the veil of appearance to view the noumena, but simply handling phenomena.
Reasoning that the mind must contain its own categories organizing sense data, making experience of space and time possible, Kant concluded uniformity of nature a priori. A class of synthetic statements was not contingent but true by necessity, then, the synthetic a priori. Kant thus saved both metaphysics and Newton's law of universal gravitation, but incidentally discarded scientific realism and developed transcendental idealism. Kant's transcendental idealism prompted the trend German idealism. G F W Hegel's absolute idealism flourished across continental Europe and fueled nationalism.
Developed by Saint-Simon, and promulgated in the 1830s by his former student Comte was positivism, the first late modern philosophy of science. In the French Revolution's aftermath, fearing society's ruin again, Comte opposed metaphysics. Human knowledge had evolved from religion to metaphysics to science, said Comte, which had flowed from mathematics to astronomy to physics to chemistry to biology to sociology--in that order--describing increasingly intricate domains, all of society's knowledge having become scientific, as questions of theology and of metaphysics were unanswerable. Comte found enumerative induction reliable by its grounding on experience available, and asserted science's use as improving human society, not metaphysical truth.
According to Comte, scientific method frames predictions, confirms them, and states laws--positive statements--irrefutable by theology or by metaphysics. Regarding experience to justify enumerative induction by having shown uniformity of nature,Mill welcomed Comte's positivism, but thought laws susceptible to recall or revision, and withheld from Comte's Religion of Humanity. Comte was confident to lay laws as irrefutable foundation of other knowledge, and the churches, honoring eminent scientists, sought to focus public mindset on altruism--a term Comte coined--to apply science for humankind's social welfare via Comte's spearheaded science, sociology.
During the 1830s and 1840s, while Comte and Mill were the leading philosophers of science, William Whewell found enumerative induction not nearly so simple, but, amid the dominance of inductivism, described "superinduction". Whewell proposed recognition of "the peculiar import of the term Induction", as "there is some Conception superinduced upon the facts", that is, "the Invention of a new Conception in every inductive inference". Rarely spotted by Whewell's predecessors, such mental inventions rapidly evade notice. Whewell explained,
"Although we bind together facts by superinducing upon them a new Conception, this Conception, once introduced and applied, is looked upon as inseparably connected with the facts, and necessarily implied in them. Having once had the phenomena bound together in their minds in virtue of the Conception, men can no longer easily restore them back to detached and incoherent condition in which they were before they were thus combined".
These "superinduced" explanations may well be flawed, but their accuracy is suggested when they exhibit what Whewell termed consilience--that is, simultaneously predicting the inductive generalizations in multiple areas--a feat that, according to Whewell, can establish their truth. Perhaps to accommodate prevailing view of science as inductivist method, Whewell devoted several chapters to "methods of induction" and sometimes said "logic of induction"--and yet stressed it lacks rules and cannot be trained.
Originator of pragmatism, C S Peirce who, as did Gottlob Frege independently, in the 1870s performed vast investigations that clarified the basis of deductive inference as mathematical proof, recognized induction but continuously insisted on a third type of inference that Peirce variously termed abduction or retroduction or hypothesis or presumption. Later philosophers gave Peirce's abduction, etc, the synonym inference to the best explanation (IBE).
Having highlighted Hume's problem of induction, John Maynard Keynes posed logical probability as its answer--but then figured not quite.Bertrand Russell found Keynes's Treatise on Probability the best examination of induction, and if read with Jean Nicod's Le Probleme logique de l'induction as well as R B Braithwaite's review of it in the October 1925 issue of Mind, to provide "most of what is known about induction", although the "subject is technical and difficult, involving a good deal of mathematics". Two decades later, Russell proposed enumerative induction as an "independent logical principle". Russell found,
"Hume's skepticism rests entirely upon his rejection of the principle of induction. The principle of induction, as applied to causation, says that, if A has been found very often accompanied or followed by B, then it is probable that on the next occasion on which A is observed, it will be accompanied or followed by B. If the principle is to be adequate, a sufficient number of instances must make the probability not far short of certainty. If this principle, or any other from which it can be deduced, is true, then the casual inferences which Hume rejects are valid, not indeed as giving certainty, but as giving a sufficient probability for practical purposes. If this principle is not true, every attempt to arrive at general scientific laws from particular observations is fallacious, and Hume's skepticism is inescapable for an empiricist. The principle itself cannot, of course, without circularity, be inferred from observed uniformities, since it is required to justify any such inference. It must therefore be, or be deduced from, an independent principle not based on experience. To this extent, Hume has proved that pure empiricism is not a sufficient basis for science. But if this one principle is admitted, everything else can proceed in accordance with the theory that all our knowledge is based on experience. It must be granted that this is a serious departure from pure empiricism, and that those who are not empiricists may ask why, if one departure is allowed, others are forbidden. These, however, are not questions directly raised by Hume's arguments. What these arguments prove--and I do not think the proof can be controverted--is that the induction is an independent logical principle, incapable of being inferred either from experience or from other logical principles, and that without this principle, science is impossible".
In a 1965 paper, Gilbert Harman explained that enumerative induction is not an autonomous phenomenon, but is simply a masked consequence of inference to the best explanation (IBE). IBE is otherwise synonym to C S Peirce's abduction. Many philosophers of science espousing scientific realism have maintained that IBE way that scientists develop approximately true scientific theories about nature.
Inductive reasoning has been criticized by thinkers as far back as Sextus Empiricus. The classic philosophical treatment of the problem of induction was given by the Scottish philosopher David Hume.
Although the use of inductive reasoning demonstrates considerable success, its application has been questionable. Recognizing this, Hume highlighted the fact that our mind draws uncertain conclusions from relatively limited experiences. In deduction, the truth value of the conclusion is based on the truth of the premise. In induction, however, the dependence on the premise is always uncertain. As an example, let's assume "all ravens are black." The fact that there are numerous black ravens supports the assumption. However, the assumption becomes inconsistent with the fact that there are white ravens. Therefore, the general rule of "all ravens are black" is inconsistent with the existence of the white raven. Hume further argued that it is impossible to justify inductive reasoning: specifically, that it cannot be justified deductively, so our only option is to justify it inductively. Since this is circular he concluded that our use of induction is unjustifiable with the help of Hume's Fork.
However, Hume then stated that even if induction were proved unreliable, we would still have to rely on it. So instead of a position of severe skepticism, Hume advocated a practical skepticism based on common sense, where the inevitability of induction is accepted.Bertrand Russell illustrated his skepticism in a story about a turkey, fed every morning without fail, who following the laws of induction concludes this will continue, but then his throat is cut on Thanksgiving Day.
Karl Popper. had declared in 1963, "Induction, i.e. inference based on many observations, is a myth. It is neither a psychological fact, nor a fact of ordinary life, nor one of scientific procedure". Popper's 1972 book Objective Knowledge--whose first chapter is devoted to the problem of induction--opens, "I think I have solved a major philosophical problem: the problem of induction". Within Popper's schema, enumerative induction is "a kind of optical illusion" cast by the steps of conjecture and refutation during the problem shift. An imaginative leap, the tentative solution is improvised, lacking inductive rules to guide it. The resulting, unrestricted generalization is deductive, an entailed consequence of all, included explanatory considerations. Controversy continued, however, with Popper's putative solution not generally accepted.
By now, inductive inference has been shown to exist, but is found rarely, as in programs of machine learning in Artificial Intelligence (AI). Popper's stance on induction is strictly falsified--enumerative induction exists--but is overwhelmingly absent from science. Although much talked of nowadays by philosophers, abduction or IBE lacks rules of inference and the discussants provide nothing resembling such, as the process proceeds by humans' imaginations and perhaps creativity.
Inductive reasoning is also known as hypothesis construction because any conclusions made are based on current knowledge and predictions. As with deductive arguments, biases can distort the proper application of inductive argument, thereby preventing the reasoner from forming the most logical conclusion based on the clues. Examples of these biases include the availability heuristic, confirmation bias, and the predictable-world bias
The availability heuristic causes the reasoner to depend primarily upon information that is readily available to them. People have a tendency to rely on information that is easily accessible in the world around them. For example, in surveys, when people are asked to estimate the percentage of people who died from various causes, most respondents would choose the causes that have been most prevalent in the media such as terrorism, and murders, and airplane accidents rather than causes such as disease and traffic accidents, which have been technically "less accessible" to the individual since they are not emphasized as heavily in the world around them.
The confirmation bias is based on the natural tendency to confirm rather than to deny a current hypothesis. Research has demonstrated that people are inclined to seek solutions to problems that are more consistent with known hypotheses rather than attempt to refute those hypotheses. Often, in experiments, subjects will ask questions that seek answers that fit established hypotheses, thus confirming these hypotheses. For example, if it is hypothesized that Sally is a sociable individual, subjects will naturally seek to confirm the premise by asking questions that would produce answers confirming that Sally is in fact a sociable individual.
The predictable-world bias revolves around the inclination to perceive order where it has not been proved to exist, either at all or at a particular level of abstraction. Gambling, for example, is one of the most popular examples of predictable-world bias. Gamblers often begin to think that they see simple and obvious patterns in the outcomes and, therefore, believe that they are able to predict outcomes based upon what they have witnessed. In reality, however, the outcomes of these games are difficult to predict and highly complex in nature. However, in general, people tend to seek some type of simplistic order to explain or justify their beliefs and experiences, and it is often difficult for them to realise that their perceptions of order may be entirely different from the truth.
There are 20 balls--either black or white--in an urn. To estimate their respective numbers, you draw a sample of four balls and find that three are black and one is white. A good inductive generalization would be that there are 15 black and five white balls in the urn.
How much the premises support the conclusion depends upon (a) the number in the sample group, (b) the number in the population, and (c) the degree to which the sample represents the population (which may be achieved by taking a random sample). The hasty generalization and the biased sample are generalization fallacies.
A statistical syllogism proceeds from a generalization to a conclusion about an individual.
Simple induction proceeds from a premise about a sample group to a conclusion about another individual.
This is a combination of a generalization and a statistical syllogism, where the conclusion of the generalization is also the first premise of the statistical syllogism.
The basic form of inductive inference, simply induction, reasons from particular instances to all instances, and is thus an unrestricted generalization. If one observes 100 swans, and all 100 were white, one might infer a universal categorical proposition of the form All swans are white. As this reasoning form's premises, even if true, do not entail the conclusion's truth, this is a form of inductive inference. The conclusion might be true, and might be thought probably true, yet it can be false. Questions regarding the justification and form of enumerative inductions have been central in philosophy of science, as enumerative induction has a pivotal role in the traditional model of the scientific method.
The process of analogical inference involves noting the shared properties of two or more things, and from this basis inferring that they also share some further property:
Analogical reasoning is very frequent in common sense, science, philosophy and the humanities, but sometimes it is accepted only as an auxiliary method. A refined approach is case-based reasoning.
A causal inference draws a conclusion about a causal connection based on the conditions of the occurrence of an effect. Premises about the correlation of two things can indicate a causal relationship between them, but additional factors must be confirmed to establish the exact form of the causal relationship.
A prediction draws a conclusion about a future individual from a past sample.
As a logic of induction rather than a theory of belief, Bayesian inference does not determine which beliefs are a priori rational, but rather determines how we should rationally change the beliefs we have when presented with evidence. We begin by committing to a prior probability for a hypothesis based on logic or previous experience, and when faced with evidence, we adjust the strength of our belief in that hypothesis in a precise manner using Bayesian logic.
Around 1960, Ray Solomonoff founded the theory of universal inductive inference, the theory of prediction based on observations; for example, predicting the next symbol based upon a given series of symbols. This is a formal inductive framework that combines algorithmic information theory with the Bayesian framework. Universal inductive inference is based on solid philosophical foundations, and can be considered as a mathematically formalized Occam's razor. Fundamental ingredients of the theory are the concepts of algorithmic probability and Kolmogorov complexity.
Some dictionaries define "deduction" as reasoning from the general to specific and "induction" as reasoning from the specific to the general. While this usage is still sometimes found even in philosophical and mathematical contexts, for the most part, it is outdated.
In a typical enumerative induction, the premises list the individuals observed to have a common property, and the conclusion claims that all individuals of the same population have that property.