Frequent Links
Open Access Articles Top Results for Inference
Journal of Biometrics & Biostatistics
Causal Inference in Randomized Trials: A Shift from the Sharp Causal Null Hypothesis to the Weak Causal Null HypothesisJournal of Data Mining in Genomics & Proteomics
Conceptual Aspects of Causal Networks in an Applied ContextJournal of Biometrics & Biostatistics
Properties of Estimators in Exponential Family Settings with Observationbased Stopping RulesJournal of Biometrics & Biostatistics
GeoAdditive Modelling of Family Size in NigeriaInternational Journal of Innovative Research in Computer and Communication Engineering
Location Privacy Context Information Effects Using Bayesian Inference FrameworkInference
This article includes a list of references, but its sources remain unclear because it has insufficient inline citations. (April 2010) 
Inference is the act or process of deriving logical conclusions from premises known or assumed to be true.^{[1]} The conclusion drawn is also called an idiomatic. The laws of valid inference are studied in the field of logic.
Alternatively, inference may be defined as the nonlogical, but rational means, through observation of patterns of facts, to indirectly see new meanings and contexts for understanding. Of particular use to this application of inference are anomalies and symbols. Inference, in this sense, does not draw conclusions but opens new paths for inquiry. (See second set of Examples.) In this definition of inference, there are two types of inference: inductive inference and deductive inference. Unlike the definition of inference in the first paragraph above, meaning of word meanings are not tested but meaningful relationships are articulated.
Human inference (i.e. how humans draw conclusions) is traditionally studied within the field of cognitive psychology; artificial intelligence researchers develop automated inference systems to emulate human inference.
Statistical inference uses mathematics to draw conclusions in the presence of uncertainty. This generalizes deterministic reasoning, with the absence of uncertainty as a special case. Statistical inference uses quantitative or qualitative (categorical) data which may be subject to random variation.
Contents
Examples
Greek philosophers defined a number of syllogisms, correct three part inferences, that can be used as building blocks for more complex reasoning. We begin with a famous example:
 All men are mortal
 Socrates is a man
 Therefore, Socrates is mortal.
The reader can check that the premises and conclusion are true, but Logic is concerned with inference: does the truth of the conclusion follow from that of the premises?
The validity of an inference depends on the form of the inference. That is, the word "valid" does not refer to the truth of the premises or the conclusion, but rather to the form of the inference. An inference can be valid even if the parts are false, and can be invalid even if some parts are true. But a valid form with true premises will always have a true conclusion.
For example, consider the form of the following symbological track:
 All meat comes from animals.
 Beef is a type of meat.
 Therefore, beef comes from an animal.
If the premises are true, then the conclusion is necessarily true, too.
Now we turn to an invalid form.
 All A are B.
 C is a B.
 Therefore, C is an A.
To show that this form is invalid, we demonstrate how it can lead from true premises to a false conclusion.
 All apples are fruit. (Correct)
 Bananas are fruit. (Correct)
 Therefore, bananas are apples. (Wrong)
A valid argument with false premises may lead to a false conclusion:
 All tall people are Greek.
 John Lennon was tall.
 Therefore, John Lennon was Greek. (wrong)
When a valid argument is used to derive a false conclusion from false premises, the inference is valid because it follows the form of a correct inference.
A valid argument can also be used to derive a true conclusion from false premises:
 All tall people are musicians (although wrong)
 John Lennon was tall (right, Valid)
 Therefore, John Lennon was a musician (Right)
In this case we have two false premises that imply a true conclusion.
Example for definition #2
Evidence: It is the early 1950s and you are an American stationed in the Soviet Union. You read in the Moscow newspaper that a soccer team from a small city in Siberia starts winning game after game. The team even defeats the Moscow team. Inference: The small city in Siberia is not a small city anymore. The Soviets are working on their own nuclear or highvalue secret weapons program.
Knowns: The Soviet Union is a command economy: people and material are told where to go and what to do. The small city was remote and historically had never distinguished itself; its soccer season was typically short because of the weather.
Explanation: In a command economy, people and material are moved where they are needed. Large cities might field good teams due to the greater availability of high quality players; and teams that can practice longer (weather, facilities) can reasonably be expected to be better. In addition, you put your best and brightest in places where they can do the most good—such as on highvalue weapons programs. It is an anomaly for a small city to field such a good team. The anomaly (i.e. the soccer scores and great soccer team) indirectly described a condition by which the observer inferred a new meaningful pattern—that the small city was no longer small. Why would you put a large city of your best and brightest in the middle of nowhere? To hide them, of course.
Incorrect inference
An incorrect inference is known as a fallacy. Philosophers who study informal logic have compiled large lists of them, and cognitive psychologists have documented many biases in human reasoning that favor incorrect reasoning.
Automatic logical inference
AI systems first provided automated logical inference and these were once extremely popular research topics, leading to industrial applications under the form of expert systems and later business rule engines. More recent work on automated theorem proving has had a stronger basis in formal logic.
An inference system's job is to extend a knowledge base automatically. The knowledge base (KB) is a set of propositions that represent what the system knows about the world. Several techniques can be used by that system to extend KB by means of valid inferences. An additional requirement is that the conclusions the system arrives at are relevant to its task.
Example using Prolog
Prolog (for "Programming in Logic") is a programming language based on a subset of predicate calculus. Its main job is to check whether a certain proposition can be inferred from a KB (knowledge base) using an algorithm called backward chaining.
Let us return to our Socrates syllogism. We enter into our Knowledge Base the following piece of code:
mortal(X) : man(X). man(socrates).
( Here : can be read as "if". Generally, if P → Q (if P then Q) then in Prolog we would code Q:P (Q if P).)
This states that all men are mortal and that Socrates is a man. Now we can ask the Prolog system about Socrates:
? mortal(socrates).
(where ? signifies a query: Can mortal(socrates). be deduced from the KB using the rules) gives the answer "Yes".
On the other hand, asking the Prolog system the following:
? mortal(plato).
gives the answer "No".
This is because Prolog does not know anything about Plato, and hence defaults to any property about Plato being false (the socalled closed world assumption). Finally
? mortal(X) (Is anything mortal) would result in "Yes" (and in some implementations: "Yes": X=socrates)
Prolog can be used for vastly more complicated inference tasks. See the corresponding article for further examples.
Use with the semantic web
Recently automatic reasoners found in semantic web a new field of application. Being based upon description logic, knowledge expressed using one variant of OWL can be logically processed, i.e., inferences can be made upon it.
Bayesian statistics and probability logic
Philosophers and scientists who follow the Bayesian framework for inference use the mathematical rules of probability to find this best explanation. The Bayesian view has a number of desirable features—one of them is that it embeds deductive (certain) logic as a subset (this prompts some writers to call Bayesian probability "probability logic", following E. T. Jaynes).
Bayesians identify probabilities with degrees of beliefs, with certainly true propositions having probability 1, and certainly false propositions having probability 0. To say that "it's going to rain tomorrow" has a 0.9 probability is to say that you consider the possibility of rain tomorrow as extremely likely.
Through the rules of probability, the probability of a conclusion and of alternatives can be calculated. The best explanation is most often identified with the most probable (see Bayesian decision theory). A central rule of Bayesian inference is Bayes' theorem.
See Bayesian inference for examples.
Nonmonotonic logic^{[2]}
A relation of inference is monotonic if the addition of premises does not undermine previously reached conclusions; otherwise the relation is nonmonotonic. Deductive inference is monotonic: if a conclusion is reached on the basis of a certain set of premises, then that conclusion still holds if more premises are added.
By contrast, everyday reasoning is mostly nonmonotonic because it involves risk: we jump to conclusions from deductively insufficient premises. We know when it is worth or even necessary (e.g. in medical diagnosis) to take the risk. Yet we are also aware that such inference is defeasible—that new information may undermine old conclusions. Various kinds of defeasible but remarkably successful inference have traditionally captured the attention of philosophers (theories of induction, Peirce's theory of abduction, inference to the best explanation, etc.). More recently logicians have begun to approach the phenomenon from a formal point of view. The result is a large body of theories at the interface of philosophy, logic and artificial intelligence.
See also
 Reasoning
 Reasoning System
 Entailment
 Epilogism
 Analogy
 Axiom
 Bayesian inference
 Frequentist inference
 Business rule
 Business rules engine
 Expert system
 Fuzzy logic
 Immediate inference
 Inference engine
 Inferential programming
 Inquiry
 Logic
 Logic of information
 Logical assertion
 Logical graph
 Nonmonotonic logic
 Rule of inference
 List of rules of inference
 Theorem
 Transduction (machine learning)
 Sherlock Holmes
References
 ^ http://www.thefreedictionary.com/inference
 ^ Fuhrmann, André. Nonmonotonic Logic (PDF). Archived from the original (PDF) on 9 December 2003.
Further reading
 Hacking, Ian (2011). An Introduction to Probability and Inductive Logic. Cambridge University Press. ISBN 0521775019.
 Jaynes, Edwin Thompson (2003). Probability Theory: The Logic of Science. Cambridge University Press. ISBN 0521592712.
 McKay, David J.C. (2003). Information Theory, Inference, and Learning Algorithms. Cambridge University Press. ISBN 0521642981.
 Russell, Stuart J.; Norvig, Peter (2003), Artificial Intelligence: A Modern Approach (2nd ed.), Upper Saddle River, New Jersey: Prentice Hall, ISBN 0137903952
 Tijms, Henk (2004). Understanding Probability. Cambridge University Press. ISBN 0521701724.
Inductive inference:
 Carnap, Rudolf; Jeffrey, Richard C., eds. (1971). Studies in Inductive Logic and Probability 1. The University of California Press.
 Jeffrey, Richard C., ed. (1979). Studies in Inductive Logic and Probability 2. The University of California Press.
 Angluin, Dana (1976). An Application of the Theory of Computational Complexity to the Study of Inductive Inference (Ph.D.). University of California at Berkeley.
 Angluin, Dana (1980). "Inductive Inference of Formal Languages from Positive Data" (PDF). Information and Control 45: 117–135. doi:10.1016/s00199958(80)902855.
 Angluin, Dana; Smith, Carl H. (Sep 1983). "Inductive Inference: Theory and Methods" (PDF). Computing Surveys 15 (3): 237–269. doi:10.1145/356914.356918.
 Gabbay, Dov M.; Hartmann, Stephan; Woods, John, eds. (2009). Inductive Logic. Handbook of the History of Logic 10. Elsevier.
 Goodman, Nelson (1973). Fact, Fiction, and Forecast. BobbsMerrill Co. Inc.
Abductive inference:
 O'Rourke, P.; Josephson, J., eds. (1997). Automated abduction: Inference to the best explanation. AAAI Press.
 Psillos, Stathis (2009). Gabbay, Dov M.; Hartmann, Stephan; Woods, John, eds. An Explorer upon Untrodden Ground: Peirce on Abduction (PDF). Handbook of the History of Logic 10. Elsevier. pp. 117–152.
 Ray, Oliver (Dec 2005). Hybrid Abductive Inductive Learning (Ph.D.). University of London, Imperial College. CiteSeerX: 10
.1 ..1 .66 .1877
Psychological investigations about human reasoning:
 deductive:
 JohnsonLaird, Philip Nicholas; Byrne, Ruth M. J. (1992). Deduction. Erlbaum.
 Byrne, Ruth M. J.; JohnsonLaird, P. N. (2009). ""If" and the Problems of Conditional Reasoning" (PDF). Trends in Cognitive Science 13 (7): 282–287. doi:10.1016/j.tics.2009.04.003.
 Knauff, Markus; Fangmeier, Thomas; Ruff, Christian C.; JohnsonLaird, P. N. (2003). "Reasoning, Models, and Images: Behavioral Measures and Cortical Activity" (PDF). Journal of Cognitive Neuroscience 15 (4): 559–573. doi:10.1162/089892903321662949.
 JohnsonLaird, Philip N. (1995). Gazzaniga, M. S., ed. Mental Models, Deductive Reasoning, and the Brain (PDF). MIT Press. pp. 999–1008.
 Khemlani, Sangeet; JohnsonLaird, P. N. (2008). "Illusory Inferences about Embedded Disjunctions". Proceedings of the 30th Annual Conference of the Cognitive Science Society. Washington/DC (PDF). pp. 2128–2133.
 statistical:
 McCloy, Rachel; Byrne, Ruth M. J.; JohnsonLaird, Philip N. (2009). "Understanding Cumulative Risk" (PDF). The Quarterly Journal of Experimental Psychology: 18.
 JohnsonLaird, Philip N. (1994). "Mental Models and Probabilistic Thinking" (PDF). Cognition 50: 189–209. doi:10.1016/00100277(94)900280.,
 analogical:
 Burns, B. D. (1996). "MetaAnalogical Transfer: Transfer Between Episodes of Analogical Reasoning". Journal of Experimental Psychology: Learning, Memory, and Cognition 22 (4): 1032–1048. doi:10.1037/02787393.22.4.1032.
 spatial:
 Jahn, Georg; Knauff, Markus; JohnsonLaird, P. N. (2007). "Preferred mental models in reasoning about spatial relations" (PDF). Memory & Cognition 35 (8): 2075–2087. doi:10.3758/bf03192939.
 Knauff, Markus; JohnsonLaird, P. N. (2002). "Visual imagery can impede reasoning" (PDF). Memory & Cognition 30 (3): 363–371. doi:10.3758/bf03194937.
 Waltz, James A.; Knowlton, Barbara J.; Holyoak, Keith J.; Boone, Kyle B.; Mishkin, Fred S.; de Menezes Santos, Marcia; Thomas, Carmen R.; Miller, Bruce L. (Mar 1999). "A System for Relational Reasoning in Human Prefrontal Cortex" (PDF). Psychological Science 10 (2): 119–125. doi:10.1111/14679280.00118.
 moral:
 Bucciarelli, Monica; Khemlani, Sangeet; JohnsonLaird, P. N. (Feb 2008). "The Psychology of Moral Reasoning" (PDF). Judgment and Decision Making 3 (2): 121–139.
External links
40x40px  Look up inference or infer in Wiktionary, the free dictionary. 


Lua error in Module:Authority_control at line 346: attempt to index field 'wikibase' (a nil value).