Deductive reasoning vs. inductive reasoning
Here's a look at the differences between deductive reasoning and inductive reasoning, with examples of each type of scientific reasoning.
- Deductive reasoning
- Inductive reasoning

Deductive reasoning examples
Inductive reasoning examples.
- Abductive reasoning
Additional resources
You don't have to be Sherlock Holmes to use your powers of deductive reasoning … or would that be inductive reasoning?
So what's the difference between inductive and deductive reasoning?
During the scientific process, deductive reasoning is used to reach a logical and true conclusion. Another type of reasoning, inductive, is also commonly used. People often confuse deductive reasoning with inductive reasoning; however, important distinctions separate these two pathways to a logical conclusion.
What is deductive reasoning?
Deductive reasoning, also known as deduction, is a basic form of reasoning. It starts out with a general statement, or hypothesis, and examines the possibilities to reach a specific, logical conclusion, according to Norman Herr , a professor of secondary education at California State University in Northridge. The scientific method uses deduction to test hypotheses and theories, which predict certain outcomes if they are correct, said Dr. Sylvia Wassertheil-Smoller , a researcher and professor emerita at Albert Einstein College of Medicine.
"We go from the general — the theory — to the specific — the observations," Wassertheil-Smoller told Live Science.
Sylvia Wassertheil-Smoller is a distinguished university professor emerita, Department of Epidemiology & Population Health (Epidemiology) at the Albert Einstein College of Medicine in New York. She's led large national studies on women's health, heart disease and stroke prevention, and has published over 300 scientific articles, as well as a book on medical research methods.
In deductive reasoning there is a first premise, then a second premise and finally an inference (a conclusion based on reasoning and evidence). A common form of deductive reasoning is the syllogism, in which two statements — a major premise and a minor premise — together reach a logical conclusion. For example, the major premise "Every A is B" could be followed by the minor premise, "This C is A." Those statements would lead to the conclusion "This C is B." Syllogisms are considered a good way to test deductive reasoning to make sure the argument is valid.
For example, "All spiders have eight legs. A tarantula is a spider. Therefore, tarantulas have eight legs." For deductive reasoning to be sound, the hypothesis must be correct. It is assumed that the statements, "All spiders have eight legs" and "a tarantula is a spider" are true. Therefore, the conclusion is logical and true. In deductive reasoning, if something is true of a class of things in general, it is also true for all members of that class.
Deductive conclusions are reliable provided the premises are true, according to Herr. The argument, "All bald men are grandfathers. Harold is bald. Therefore, Harold is a grandfather," is valid logically, but it is untrue because the original premise is false.
What is inductive reasoning
While deductive reasoning begins with a premise that is proven through observations, inductive reasoning extracts a likely (but not certain) premise from specific and limited observations. There is data, and then conclusions are drawn from the data; this is called inductive logic, according to the University of Illinois in Springfield.
"In inductive inference, we go from the specific to the general. We make many observations, discern a pattern, make a generalization, and infer an explanation or a theory," Wassertheil-Smoller told Live Science. "In science, there is a constant interplay between inductive inference (based on observations) and deductive inference (based on theory), until we get closer and closer to the 'truth,' which we can only approach but not ascertain with complete certainty."
In other words, the reliability of a conclusion made with inductive logic depends on the completeness of the observations. For instance, let's say that you have a bag of coins; you pull three coins from the bag, and each coin is a penny. Using inductive logic, you might then propose that all of the coins in the bag are pennies."Even though all of the initial observations — that each coin taken from the bag was a penny — are correct, inductive reasoning does not guarantee that the conclusion will be true.
Here's another example: "Penguins are birds. Penguins can't fly. Therefore, all birds can't fly." The conclusion does not follow logically from the statements.
Nevertheless, inductive reasoning has its place in the scientific method , and scientists use it to form hypotheses and theories . Deductive reasoning then allows them to apply the theories to specific situations.
Here are some examples of deductive reasoning:
Major premise: All mammals have backbones. Minor premise: Humans are mammals. Conclusion: Humans have backbones.
Major premise: All birds lay eggs. Minor premise: Pigeons are birds. Conclusion: Pigeons lay eggs.
Major premise: All plants perform photosynthesis. Minor premise: A cactus is a plant. Conclusion: A cactus performs photosynthesis.
Here are some examples of inductive reasoning:
Data: I see fireflies in my backyard every summer. Hypothesis: This summer, I will probably see fireflies in my backyard.
Data: I tend to catch colds when people around me are sick. Hypothesis: Colds are infectious.
Data: Every dog I meet is friendly.
Hypothesis: Most dogs are usually friendly.
What is abductive reasoning
Another form of scientific reasoning that diverges from inductive and deductive reasoning is abductive. Abductive reasoning usually starts with an obviously incomplete set of observations and proceeds to the likeliest possible explanation for the data, a ccording to Butte College in Oroville, California. It is based on making and testing hypotheses using the best information available. It often entails making an educated guess after observing a phenomenon for which there is no clear explanation.
For example, a person walks into their living room and finds torn-up papers all over the floor. The person's dog has been alone in the apartment all day. The person concludes that the dog tore up the papers because it is the most likely scenario. It's possible that a family member with a key to the apartment destroyed the papers, or it may have been done by the landlord, but the dog theory is the most likely conclusion.
Abductive reasoning is useful for forming hypotheses to be tested. Abductive reasoning is often used by doctors who make a diagnosis based on test results, and by jurors who make decisions based on the evidence presented to them.
- This guide from Scholastic provides ideas for teaching younger kids all about scientific reasoning. PBS has put together some video clips and games about deductive and inductive reasoning.This book written by Christopher Moore provides information on how to use scientific reasoning in the classroom.
Live Science newsletter
Stay up to date on the latest science news by signing up for our Essentials newsletter.
Mindy Weisberger is a Live Science editor for the channels Animals and Planet Earth. She also reports on general science, covering climate change, paleontology, biology, and space. Mindy studied film at Columbia University; prior to Live Science she produced, wrote and directed media for the American Museum of Natural History in New York City. Her videos about dinosaurs, astrophysics, biodiversity and evolution appear in museums and science centers worldwide, earning awards such as the CINE Golden Eagle and the Communicator Award of Excellence. Her writing has also appeared in Scientific American, The Washington Post and How It Works Magazine.
New York's first dinosaur museum was trashed before it even opened. The culprit is not who we thought.
Did Noah's flood really happen?
Watch an octopus waking up from what scientists think could have been a nightmare
Most Popular
By James Bentley April 19, 2023
By Fran Ruiz April 19, 2023
By Marc Lallanilla April 18, 2023
By Erin Macdonald April 18, 2023
By James O’Malley April 18, 2023
By Anna Gora April 17, 2023
By Anna Gora, Cari Nierenberg April 14, 2023
By James O’Malley April 14, 2023
By James Abbott April 13, 2023
By Dr. David Warmflash April 13, 2023
By Dr. David Warmflash April 06, 2023
- 2 Why are insects attracted to artificial lights?
- 3 Perfectly preserved 7,000-year-old skeleton unearthed during renovation in Poland
- 4 China's mysterious space plane released an unidentified 'object' in orbit, US intelligence reveals
- 5 James Webb Telescope finds evidence of 'celestial monster' stars the size of 10,000 suns lurking at the dawn of time
- 2 Rhino-like 'thunder beasts' grew massive in the evolutionary blink of an eye after dinos died off
- 3 5,400-year-old tomb discovered in Spain perfectly captures the summer solstice
- 4 Long-extinct Tasmanian tiger may still be alive and prowling the wilderness, scientists claim
- 5 What is the deadliest spider in the world?
- More from M-W
- To save this word, you'll need to log in. Log In
Definition of deduction
Frequently asked questions.
What is the difference between deduction and induction ?
Deductive reasoning, or deduction, is making an inference based on widely accepted facts or premises. If a beverage is defined as "drinkable through a straw," one could use deduction to determine soup to be a beverage. Inductive reasoning, or induction , is making an inference based on an observation, often of a sample. You can induce that the soup is tasty if you observe all of your friends consuming it. Read more on the difference between deduction and induction
What is the difference between abduction and deduction ?
Abductive reasoning, or abduction , is making a probable conclusion from what you know. If you see an abandoned bowl of hot soup on the table, you can use abduction to conclude the owner of the soup is likely returning soon. Deductive reasoning, or deduction , is making an inference based on widely accepted facts or premises. If a meal is described as "eaten with a fork" you may use deduction to determine that it is solid food, rather than, say, a bowl of soup.
What is the difference between deduction and adduction ?
Adduction is "the action of drawing (something, such as a limb) toward or past the median axis of the body," and "the bringing together of similar parts." Deduction may be "an act of taking away," or "something that is subtracted." Both words may be traced in part to the Latin dūcere, meaning "to lead."
Example Sentences
These examples are programmatically compiled from various online sources to illustrate current usage of the word 'deduction.' Any opinions expressed in the examples do not represent those of Merriam-Webster or its editors. Send us feedback about these examples.
Word History
15th century, in the meaning defined at sense 1a
Phrases Containing deduction
- payroll deduction
Articles Related to deduction

'Deduction' vs. 'Induction' vs....
'Deduction' vs. 'Induction' vs. 'Abduction'
Learn the differences between these three types of reasoning

10 Sherlock Holmes Words Worth...
10 Sherlock Holmes Words Worth Investigating
The language in Sherlock Holmes is as intriguing as the stories.
Dictionary Entries Near deduction
deduction new for old
Cite this Entry
“Deduction.” Merriam-Webster.com Dictionary , Merriam-Webster, https://www.merriam-webster.com/dictionary/deduction. Accessed 25 May. 2023.
Kids Definition
Kids definition of deduction, legal definition, legal definition of deduction, more from merriam-webster on deduction.
Nglish: Translation of deduction for Spanish Speakers
Britannica English: Translation of deduction for Arabic Speakers
Britannica.com: Encyclopedia article about deduction
Subscribe to America's largest dictionary and get thousands more definitions and advanced search—ad free!

Can you solve 4 words at once?
Word of the day.
See Definitions and Examples »
Get Word of the Day daily email!

You've used more than you might think

When 'thingamajig' and 'thingamabob' just won't do

Look up any year to find out

How 'literally' can mean 'figuratively'

A simple way to keep them apart. (Most of the time.)

And who put it there, anyway?

A simple trick to keep them separate

Test your visual vocabulary!
Take the quiz

Prove you're the best of the nest.

Test your vocabulary with our 10-question quiz!

Can you outdo past winners of the National Spelli...
Deductive Reasoning (Definition + Examples)
Practical Psychology
December 6, 2022
At the age of 11 or 12, children reach what famed psychologist Jean Piaget called the formal operational stage. It’s the final stage in the child’s development. During this stage, children start to think abstractly and even apply those ideas to problem-solving. They also learn a process called deductive reasoning.
Deductive reasoning is a process in which we draw conclusions about the world around us. It’s also one of the basic ideas introduced to students who are learning about logic and how to form an argument. Deductive reasoning can help us discover the truth, but as you’ll see in the video, sometimes this process is done so quickly because it’s obvious.
On this page, I’m going to talk about deductive reasoning, how we use it in everyday life, and how it differs from inductive reasoning. Understanding deductive and inductive reasoning are essential building blocks for understanding how we make sense of the world and how we make decisions.

What is Deductive Reasoning?
Deductive reasoning, or deduction, is the process of using a group of true premises to draw a conclusion that is also true. This is also known as “top-down logic” because it takes broad statements and uses them to create more narrow statements.
Here’s an example of deductive reasoning.
Premise A says that all dogs are good boys.
Premise B says that Kevin is a dog.
The conclusion that we draw from deductive reasoning says that Kevin is a good boy.
Of course, that example is just silly, but it shows how we can use two ideas and deductive reasoning to form an argument or a statement. Other examples of premises like this include “all dogs are mammals” or “every human embryo is made from sperm and an egg.”
Premise A is typically a very broad and general statement. Premise B is a more narrow statement that relates back to Premise A. The conclusion states a narrow truth that relates to both Premise A and Premise B.
Characteristics of deductive reasoning
In order to start the deductive reasoning process, you must use a statement that we all know to be true. If the statement is not true, or true some of the time, you may still be able to form a conclusion through induction. But in order to use deductive reasoning, that truth needs to be as solid as concrete.
It will also have to funnel down to make a more narrow conclusion through entailment. Premise A and Premise B must be related in a way that Premise C can exist. Let’s go back to our example.
In both Premise A and Premise B, dogs are mentioned. Premise C grabs a conclusion from both of these premises in a logical, relevant way. When any of these parts of the deduction don’t follow the rules, problems may ensue.

The rules of deductive reasoning are airtight. If you’re not following them, you’re not using deductive reasoning. This may not change the validity of the premises or the conclusions that you draw from your premises, but it does change whether or not it falls under the category of deductive reasoning.
If any of the following exist, you might end up coming to a false conclusion:
- False premises
- Lack of entailment
- A narrow truth
False Premises
Let’s go back to the idea that all dogs are good boys. In this case, one can unfortunately argue that not all dogs are good boys. This would automatically make the conclusion untrue. A conclusion is only considered the truth when the premises that precede it are true.
Notice here that we said that the conclusion is untrue. You may come back and argue that Kevin is a good boy, even though not all dogs are. That simply means that the conclusion is valid. In philosophy, validity and truth are not the same thing.
So while some dogs are good boys, Kevin is a dog, and Kevin is a good boy, this is not a conclusion that you can draw through deductive reasoning as it was laid out by ancient philosophers.
Lack of Entailment

Here’s another problem with deductive reasoning that we run into a lot. In order for a conclusion to be true, the premises that precede it directly support and lead to the conclusion.
Here’s an example of how failing to use this rule can create a weak conclusion. (Let’s go back to pretending that “all dogs are good boys” is a known fact.)
The conclusion drawn from this is that Kevin has blue eyes.
Kevin could very well have blue eyes, but just because the conclusion is valid doesn’t mean that it is true, because we have nothing to support the idea that Kevin’s eyes are blue.
Remember, you have to reach this conclusion through entailment. No premise has anything to do with the color of Kevin’s eyes or the color of any dog’s eyes. So we can’t come to that conclusion based on the premises that have been given to us.
Narrow Truth
Think of all of the things that you know as true. Surprisingly, these broad and general facts are not easy to come by. And when they do, they always seem too obvious to use in an example.
So deductive reasoning also seems very obvious, and outside from being the basis of forming an argument, it’s not useful in everyday life.
Let’s use another example of deductive reasoning, shall we?
Premise A says that all humans live on land.
Premise B says that Megan is a human.
The conclusion that you would get from deductive reasoning says that Megan lives on land.
Well, yeah. Duh. She’s a human, after all.
Deductive reasoning comes naturally to us. We do it without thinking. To figure out that a human lives on land or that a dog is a mammal is a quick process when you already know that all dogs are mammals and that all humans live on land.
But due to the nature of deductive reasoning, you need those broad truths to draw conclusions from. A more narrow truth won’t give you much to work with.
Example 1: All humans are mortal. Susan is a human. Susan is mortal.
This is a classic example of deductive reasoning. It starts with a statement that is entirely true – you can’t poke holes in it or argue against it. (Maybe in a few decades you can, but not today!) The next statement is also true, and ties into the first statement. The conclusion brings both statements together to create a statement that we have now proven is true.
Example 2: Marketing
In everyday life, we don’t always use deductive reasoning using the strict rules of traditional logic. Marketers, for example, may use deductive reasoning to make decisions about how they want to advertise their products toward certain groups of customers.
They may use information from focus groups or surveys to create a profile of their products. Let’s say a company that makes cleaning products wants to target single women, in their late 20s, who are upper-middle-class. They collect information about the demographic and learn that upper-middle-class single women in their late 20s find more value products that have natural ingredients and are “green.”
Premise 1 is that upper-middle-class women in their 20s find more value in products that have natural ingredients and are “green.”
Premise 2 is that the company’s target audience is upper-middle class women in their 20s.
The marketers draw a conclusion that if they brand their products as “green” and highlight their natural ingredients, their target audience will find more value in their products.
Again, this doesn’t exactly fit the rules of “top-down logic.” Not every upper-middle-class woman particularly cares what is in their cleaning products. And not every upper-middle-class woman is in the company’s target audience. But this is often how we use deductive reasoning to draw conclusions. These conclusions can still be very helpful, even if the conclusions aren’t 100% true.
Example 3: Deductive Reasoning in Math
Deductive reasoning is introduced in math classes to help students understand equations and create proofs. When math teachers discuss deductive reasoning, they usually talk about syllogisms. Syllogisms are a form of deductive reasoning that help people discover a truth.
Here’s an example.
The sum of any triangle’s three angles is 180 degrees.
You are given a triangle to work with.
You can conclude that the sum of the triangle’s three angles is 180 degrees.
This conclusion will help you move forward when working with the triangle and discovering the length of each side or the measurement of each angle.
Example 4: Deductive Reasoning in Science
Both deduction and induction are used to prove hypotheses and support the scientific method. Deduction requires us to look at how closely a premises and the conclusion are related. If the premises are backed by evidence and experiment, then the conclusion is more likely to be true.
In the scientific method, scientists form a hypothesis. They then conduct a series of experiments to see whether that hypothesis is true. With each experiment, they prove the strength of the premises and support their conclusion about whether or not their hypothesis is correct.
Without deductive reasoning, scientists may come to untrue conclusions or accept things that are likely as things that are true.
Deductive vs inductive reasoning
In the beginning of this video, I mentioned that child psychologist Jean Piaget theorized that children developed the skills of deductive reasoning around 11 or 12 years old. From then on, it’s not exactly something that we think about.
So we’re more likely to draw conclusions about things in the opposite direction. We use inductive reasoning to make sense of the world around us. We take a single experience or a few experiences from the past to make a conclusion about what might happen in the immediate future or indefinitely.
Inductive reasoning is more prevalent in our everyday lives because it just requires a personal experience or a handful of facts to work. Getting down to the “truth,” especially if you are a philosopher or someone who is especially skilled in logic, is not always an easy thing to do. Plus, using deductive reasoning doesn’t usually give us any incentive or confidence to take action. It just helps us build the world.
But I’ll talk more about inductive reasoning in my next video. I’ll break down what inductive reasoning is, the different types of inductive reasoning that we use in everyday life, and the problems that come with inductive reasoning.
Have you been listening? Let’s test your knowledge with a quick, three-question quiz on deductive reasoning.
First question:
Is deductive reasoning considered “top-down” or “bottom-up” logic?
“Top-down logic.” It starts with broad truths and makes its way down to a more narrow conclusion. “Bottom-up logic” is called induction.
Second question:
What can interfere with deduction?
A: False premises
B: Lack of entailment
C: Narrow truth
D: All of the above
All of the above! In order to arrive at the truth, you will need to provide true premises, that logically lead to the conclusion. This means starting with a very broad truth and making your way down.
Last question: does this “count” as deductive reasoning?
Premise 1: All pigeons are birds.
Premise 2: John is a pigeon.
Conclusion: John is a bird.
Yes, it counts! All of the premises are true and contribute to the final conclusion, which is also true.
Related posts:
- Inductive Reasoning (Definition + Examples)
- Perceptual Reasoning (Definition + Examples)
- Formal Operational Stage
- Robbers Cave Experiment
- Ivan Pavlov (Biography + Experiments)

Practical Psychology began as a collection of study material for psychology students in 2016, created by a student in the field. It has since evolved into an online blog and YouTube channel providing mental health advice, tools, and academic support to individuals from all backgrounds. With over 2 million YouTube subscribers, over 500 articles, and an annual reach of almost 12 million students, it has become one of the most popular sources of psychological information.
Leave a Comment Cancel reply

PracticalPie.com is a participant in the Amazon Associates Program. As an Amazon Associate we earn from qualifying purchases.
Follow Us On:
Psychology Resources
Developmental
Personality
Relationships
Psychologists
Serial Killers
Psychology Tests
Personality Quiz
Memory Test
Depression test
Type A/B Personality Test
© PracticalPsychology. All rights reserved
Privacy Policy | Terms of Use
Inductive VS Deductive Reasoning – The Meaning of Induction and Deduction, with Argument Examples
If you're conducting research on a topic, you'll use various strategies and methods to gather information and come to a conclusion.
Two of those methods are inductive and deductive reasoning.
So what's the difference between inductive and deductive reasoning, when should you use each method, and is one better than the other?
We'll answer those questions and give you some examples of both types of reasoning in this article.
What is Inductive Reasoning?
The method behind inductive reasoning.
When you're using inductive reasoning to conduct research, you're basing your conclusions off your observations. You gather information - from talking to people, reading old newspapers, observing people, animals, or objects in their natural habitat, and so on.
Inductive reasoning helps you take these observations and form them into a theory. So you're starting with some more specific information (what you've seen/heard) and you're using it to form a more general theory about the way things are.
What does the inductive reasoning process look like?
You can think of this process as a reverse funnel – starting with more specifics and getting broader as you reach your conclusions (theory).
Some people like to think of it as a "bottom up" approach (meaning you're starting at the bottom with the info and are going up to the top where the theory forms).
Here's an example of an inductive argument:
Observation (premise): My Welsh Corgis were incredibly stubborn and independent (specific observation of behavior). Observation (premise): My neighbor's Corgis are the same way (another specific observation of behavior). Theory: All Welsh Corgis are incredibly stubborn and independent (general statement about the behavior of Corgis).
As you can see, I'm basing my theory on my observations of the behavior of a number of Corgis. Since I only have a small amount of data, my conclusion or theory will be quite weak.
If I was able to observe the behavior of 1000 Corgis (omg that would be amazing), my conclusion would be stronger – but still not certain. Because what if 10 of them were extremely well-behaved and obedient? Or what if the 1001st Corgi was?
So, as you can see, I can make a general statement about Corgis being stubborn, but I can't say that ALL of them are.
What can you conclude with inductive reasoning?
As I just discussed, one of the main things to know about inductive reasoning is that any conclusions you make from inductive research will not be 100% certain or confirmed.
Let's talk about the language we use to describe inductive arguments and conclusions. You can have a strong argument (if your premise(s) are true, meaning your conclusion is probably true). And that argument becomes cogent if the conclusion ends up being true.
Still, even if the premises of your argument are true, and that means that your conclusion is probably true, or likely true, or true much of the time – it's not certain.
And – weirdly enough – your conclusion can still be false even if all your premises are true (my Corgis were stubborn, my neighbor's corgis were stubborn, perhaps a friend's Corgis and the Queen of England's Corgis were stubborn...but that doesn't guarantee that all Corgis are stubborn).
How to make your inductive arguments stronger
If you want to make sure your inductive arguments are as strong as possible, there are a couple things you can do.
First of all, make sure you have a large data set to work with. The larger your sample size, the stronger (and more certain/conclusive) your results will be. Again, thousands of Corgis are better than four (I mean, always, amiright?).
Second, make sure you're taking a random and representative sample of the population you're studying. So, for example, don't just study Corgi puppies (cute as they may be). Or show Corgis (theoretically they're better trained). You'd want to make sure you looked at Corgis from all walks of life and of all ages.
If you want to dig deeper into inductive reasoning, look into the three different types – generalization, analogy, and causal inference. You can also look into the two main methods of inductive reasoning, enumerative and eliminative. But those things are a bit out of the scope of this beginner's guide. :)
What is Deductive Reasoning?
The method behind deductive reasoning.
In order to use deductive reasoning, you have to have a theory to begin with. So inductive reasoning usually comes before deductive in your research process.
Once you have a theory, you'll want to test it to see if it's valid and your conclusions are sound. You do this by performing experiments and testing your theory, narrowing down your ideas as the results come in. You perform these tests until only valid conclusions remain.
What does the deductive reasoning process look like?
You can think of this as a proper funnel – you start with the broad open top end of the funnel and get more specific and narrower as you conduct your deductive research.
Some people like to think of this as a "top down" approach (meaning you're starting at the top with your theory, and are working your way down to the bottom/specifics). I think it helps to think of this as " reductive " reasoning – you're reducing your theories and hypotheses down into certain conclusions.
Here's an example of a deductive argument:
We'll use a classic example of deductive reasoning here – because I used to study Greek Archaeology, history, and language:
Theory: All men are mortal Premise: Socrates is a man Conclusion: Therefore, Socrates is mortal
As you can see here, we start off with a general theory – that all men are mortal. (This is assuming you don't believe in elves, fairies, and other beings...)
Then we make an observation (develop a premise) about a particular example of our data set (Socrates). That is, we say that he is a man, which we can establish as a fact.
Finally, because Socrates is a man, and based on our theory, we conclude that Socrates is therefore mortal (since all men are mortal, and he's a man).
You'll notice that deductive reasoning relies less on information that could be biased or uncertain. It uses facts to prove the theory you're trying to prove. If any of your facts lead to false premises, then the conclusion is invalid. And you start the process over.
What can you conclude with deductive reasoning?
Deductive reasoning gives you a certain and conclusive answer to your original question or theory. A deductive argument is only valid if the premises are true. And the arguments are sound when the conclusion, following those valid arguments, is true.
To me, this sounds a bit more like the scientific method. You have a theory, test that theory, and then confirm it with conclusive/valid results.
To boil it all down, in deductive reasoning:
"If all premises are true, the terms are clear , and the rules of deductive logic are followed, then the conclusion reached is necessarily true ." ( Source )
So Does Sherlock Holmes Use Inductive or Deductive Reasoning?
Sherlock Holmes is famous for using his deductive reasoning to solve crimes. But really, he mostly uses inductive reasoning. Now that we've gone through what inductive and deductive reasoning are, we can see why this is the case.
Let's say Sherlock Holmes is called in to work a case where a woman was found dead in her bed, under the covers, and appeared to be sleeping peacefully. There are no footprints in the carpet, no obvious forced entry, and no immediately apparent signs of struggle, injury, and so on.
Sherlock observes all this as he looks in, and then enters the room. He walks around the crime scene making observations and taking notes. He might talk to anyone who lives with her, her neighbors, or others who might have information that could help him out.
Then, once he has all the info he needs, he'll come to a conclusion about how the woman died.
That pretty clearly sounds like an inductive reasoning process to me.
Now you might say - what if Sherlock found the "smoking gun" so to speak? Perhaps this makes his arguments and process seem more deductive.
But still, remember how he gets to his conclusions: starting with observations and evidence, processing that evidence to come up with a hypothesis, and then forming a theory (however strong/true-seeming) about what happened.
How to Use Inductive and Deductive Reasoning Together
As you might be able to tell, researchers rarely just use one of these methods in isolation. So it's not that deductive reasoning is better than inductive reasoning, or vice versa – they work best when used in tandem.
Often times, research will begin inductively. The researcher will make their observations, take notes, and come up with a theory that they want to test.
Then, they'll come up with ways to definitively test that theory. They'll perform their tests, sort through the results, and deductively come to a sure conclusion.
So if you ever hear someone say "I deduce that x happened", they better make sure they're working from facts and not just observations. :)
TL;DR: Inductive vs Deductive Reasoning – What are the Main Differences?
Inductive reasoning:.
- Based on observations, conversations, stuff you've read
- Starts with information/evidence and works towards a broader theory
- Arguments can be strong and cogent, but never valid or sound (that is, certain)
- Premises can all be true, but conclusion doesn't have to be true
Deductive reasoning:
- Based on testing a theory, narrowing down the results, and ending with a conclusion
- Starts with a broader theory and works towards certain conclusion
- Arguments can be valid/invalid or sound/unsound, because they're based on facts
- If premises are true, conclusion has to be true
And here's a cool and helpful chart if you're a visual learner:
That's about it!
Now, if you need to conduct some research, you should have a better idea of where to start – and where to go from there.
Just remember that induction is all about observing, hypothesizing, and forming a theory. Deducing is all about taking that (or any) theory, boiling it down, and testing until a certain conclusion(s) is all that remains.
Happy reasoning!
Former archaeologist, current editor and podcaster, life-long world traveler and learner.
If you read this far, tweet to the author to show them you care. Tweet a thanks
Learn to code for free. freeCodeCamp's open source curriculum has helped more than 40,000 people get jobs as developers. Get started
- Search Search Please fill out this field.
What Is a Deduction?
Understanding deductions.
- Standard vs. Itemized Deductions
Deductions vs. Credits
Special considerations, the bottom line.
Deductions & Credits
Tax Deductions
Deduction Definition and Standard Deductions for 2022
Investopedia contributors come from a range of backgrounds, and over 24 years there have been thousands of expert writers and editors who have contributed.
:max_bytes(150000):strip_icc():format(webp)/Group1805-3b9f749674f0434184ef75020339bd35.jpg)
Lea Uradu, J.D. is a Maryland State Registered Tax Preparer, State Certified Notary Public, Certified VITA Tax Preparer, IRS Annual Filing Season Program Participant, and Tax Writer.
:max_bytes(150000):strip_icc():format(webp)/lea_uradu_updated_hs_121119-LeaUradu-7c54e49efe814a048e41278182ba02f8.jpg)
A deduction is an expense that can be subtracted from a taxpayer's gross income in order to reduce the amount of income that is subject to taxation.
For example, if you earn $50,000 in a year and make a $1,000 donation to charity during that year, you are eligible to claim a deduction for that donation, reducing your taxable income to $49,000. The Internal Revenue Service (IRS) often refers to a deduction as an allowable deduction.
Key Takeaways
- A deduction is an expense that can be subtracted from taxable income to reduce the amount owed.
- Most taxpayers who take the standard deduction only need to file Form 1040.
- Taxpayers who itemize deductions must use Schedule A Form 1040 to list all of their allowable deductions.
- The standard tax deductions have increased steadily since the passage of the Tax Cuts and Jobs Act in 2017.
Taxpayers in the United States have the choice of claiming the standard deduction or itemizing their deductions . Claiming the standard deduction is easier and requires less paperwork and record-keeping. The Internal Revenue Service (IRS) has revamped Form 1040 , which most taxpayers now use, and retired the old 1040A and 1040EZ forms.
Taxpayers who itemize deductions must use Schedule A Form 1040 , an attachment to the standard 1040 form, and are required to fill in a list of their allowable deductions and keep receipts to prove them if they are audited . This longer form is used by filers who have substantial deductions that add up to more than the standard deduction.
An itemized deduction is an expense subtracted from adjusted gross income (AGI) , which reduces taxable income and, therefore, the amount of taxes owed. Common itemized deductions include interest on a mortgage loan, unreimbursed healthcare costs, charitable contributions , and state and local taxes. Please consult a tax professional to determine whether a standard deduction or itemizing works for your financial situation.
Standard Tax Deductions
Since the passage of the Tax Cuts and Jobs Act of 2017 (TCJA) , the standard deduction has increased over the years to help keep pace with rising prices—called inflation .
Below are the standard deductions for tax years 2022 and 2023, depending on tax filing status . These are set to expire in 2025.
2022 Standard Deductions
- For singles and married people filing separately: $12,950
- For married couples filing jointly: $25,900
- For heads of households: $19,400
2023 Standard Deductions
- For singles and married people filing separately: $13,850
- For married couples filing jointly: $27,700
- For heads of households: $20,800
The current standard deductions are a significant upgrade from levels before the Tax Cuts and Jobs Act was passed. For example, in the 2017 tax year, the standard deduction was $6,350 for single filers and $12,700 for married people filing jointly.
If you opt to claim the standard deduction, there are still some itemized deductions you can claim on your income tax return , including eligible student loan interest and tuition and fees .
A deduction is different from a tax credit , which is subtracted from the amount of taxes owed, not from your reported income.
There are both refundable and non-refundable credits. Non-refundable credits cannot trigger a tax refund, but refundable credits can.
For example, imagine that after reporting your income and claiming your deductions, you owe $500 in income tax; however, you are eligible for a $600 credit. If the credit is non-refundable, your tax bill is erased, but you do not receive any extra money. If the credit is refundable, you receive a $100 tax refund .
Some businesses qualify for business tax credits , which offset or reduce a company’s taxes owed to the federal government. Business tax credits are designed to encourage a particular behavior that benefits the overall economy, such as upgrading a building or factory and investing in research. While tax deductions reduce taxable income, business tax credits reduce the taxes owed.
Business owners have a much more involved process during tax time since they're taxed on business profits, not business proceeds or revenue . That means documenting their costs of doing business to subtract them from the gross proceeds, revealing the taxable profits. The process is the same for the smallest businesses to the largest corporations, although the corporations at least have accounting departments to take care of the paperwork.
Businesses are required to report all of their gross income and then deduct business expenses from it. The difference between the two numbers is the business's net taxable income. Thus, business expenses work in a way that is similar to deductions.
Although the process of tracking expenses can be burdensome, the total amount of these expenses can help reduce a company's taxable income substantially, thus, lowering the taxes owed.
What Are Tax Deduction Examples?
Examples of common tax deductions include mortgage interest, contributions towards retirement plans, student loan interest, charitable contributions, certain health expenses, gambling losses, and HSA contributions.
Are Tax Deductions Good?
Yes, tax deductions are good because they lower your income and, therefore, the amount of taxes you owe. For example, if you had to pay 10% in taxes on your income and your income was $1,000, you would owe $100 in taxes; however, if you had a tax deduction of $200, that would lower your income to $800, and you would now owe $80 in taxes.
What Is the Tax Deduction for 2022?
The standard tax deduction for single filers for tax year 2022 is $12,950 and is $13,850 in 2023. This is the same for married individuals filing separately. For those married and filing jointly, the deduction for tax year 2022 is $25,900 and is $27,700 in 2023. For heads of households, it is $19,400 for tax year 2022 and is $20,800 in 2023.
A deduction is an expense that a taxpayer can use to reduce their gross income, thereby reducing the overall taxes they pay. The IRS allows for a variety of deductions that individuals can use to reduce their gross income.
Taxpayers are allowed to itemize their deductions or to take the standard deduction, which is much larger now than in the past thanks to the TCJA of 2017. It is best to consult a tax professional or financial advisor to see which method of deductions has the greatest benefit for your individual tax situation.
Internal Revenue Service. " About Schedule A (Form 1040), Itemized Deductions ."
Internal Revenue Service. " Topic No. 551 Standard Deduction ."
Internal Revenue Service. " Here Are Five Facts About the New Form 1040 ."
Internal Revenue Service. " Credits and Deductions for Individuals ."
Internal Revenue Service. “ IRS Provides Tax Inflation Adjustments for Tax Year 2022 .”
Internal Revenue Service. “ IRS Provides Tax Inflation Adjustments for Tax Year 2023 .”
Internal Revenue Service. " Taxpayers Can Choose to Itemize or Take Standard Deduction for Tax Year 2017 ."
Internal Revenue Service. " Tax Benefits for Education: Information Center ."
- Terms of Service
- Editorial Policy
- Privacy Policy
- Do Not Sell My Personal Information
By clicking “Accept All Cookies”, you agree to the storing of cookies on your device to enhance site navigation, analyze site usage, and assist in our marketing efforts.
- Azerbaijani
What is the difference between Conclusion and Deduction ?Feel free to just provide example sentences.
- Report copyright infringement

- English (US)
a conclusion is a final understanding and reasoning behind the action or problem. once you find the conclusion, the reason behind, it you can fix the problem. deduction is a process used to find a conclusion and answer. for example scientists use different ways to deduct other possible reasons behind something before they can find the correct conclusion.

- What is the difference between conclusion and inference ? answer An inference is your guess of what is going to happen before it happens and a conclusion is what actually happens. An inference example wou...
- What is the difference between Inference and Conclusion ? answer Conclusion is the ending to something. Inference means to make an educated guess.
- What is the difference between Conclusion and Consequence ? answer Conclusion - the end of something Consequence - something that happens as a result of something else
- What is the difference between conclusion and implication ? answer Conclusion means the end. It is commonly used as a transition to the closing sentence of a paper or speech Example: “In conclusion, we need ...
- What is the difference between Conclusion and Inference ? answer An inference is a generalization based on some implication. A conclusion is reached at the end of a thought process.
- What is the difference between I'm down for it and I'm up for it ?
- What is the difference between man and men ?
- What is the difference between rape and molest ?
- What is the difference between keep awake and stay awake ?
- What is the difference between signature and printed name ?
- What is the difference between relish and look forward to ?
- What is the difference between travel and trip ?
- What is the difference between naive and innocent ?
- What is the difference between What and That ?
- What is the difference between The odds of me dying today is zero and The odds of dying for me to...
- What is the difference between lent credence to and confirm ?
- What is the difference between misjudge and underestimate ?
- What is the difference between stoll and saunter ?
- How do you say this in Korean? "I still don't understand this world very well.. ㅋㅋㅋㅋ I have a lot...
- How do you say this in Korean? Soy
The Language Level symbol shows a user's proficiency in the languages they're interested in. Setting your Language Level helps other users provide you with answers that aren't too complex or too simple.
Has difficulty understanding even short answers in this language.
Can ask simple questions and can understand simple answers.
Can ask all types of general questions and can understand longer answers.
Can understand long, complex answers.
Show your appreciation in a way that likes and stamps can't.
By sending a gift to someone, they will be more likely to answer your questions again!

If you post a question after sending a gift to someone, your question will be displayed in a special section on that person’s feed.

Solve your problems more easily with the app!
- Find the answer you're looking for from 45 million answers logged!
- Enjoy the auto-translate feature when searching for answers!
- It’s FREE!!

- What is the difference ...
- definitions
“Inductive” vs. “Deductive”: How To Reason Out Their Differences
- What Does Inductive Mean?
- What Does Deductive Mean?
- Inductive Reasoning Vs. Deductive Reasoning
Inductive and deductive are commonly used in the context of logic, reasoning, and science. Scientists use both inductive and deductive reasoning as part of the scientific method . Fictional detectives like Sherlock Holmes are famously associated with methods of deduction (though that’s often not what Holmes actually uses—more on that later). Some writing courses involve inductive and deductive essays.
But what’s the difference between inductive and deductive ? Broadly speaking, the difference involves whether the reasoning moves from the general to the specific or from the specific to the general. In this article, we’ll define each word in simple terms, provide several examples, and even quiz you on whether you can spot the difference.
⚡ Quick summary
Inductive reasoning (also called induction ) involves forming general theories from specific observations. Observing something happen repeatedly and concluding that it will happen again in the same way is an example of inductive reasoning. Deductive reasoning (also called deduction ) involves forming specific conclusions from general premises, as in: everyone in this class is an English major; Jesse is in this class; therefore, Jesse is an English major.
What does inductive mean?
Inductive is used to describe reasoning that involves using specific observations, such as observed patterns, to make a general conclusion. This method is sometimes called induction . Induction starts with a set of premises , based mainly on experience or experimental evidence. It uses those premises to generalize a conclusion .
For example, let’s say you go to a cafe every day for a month, and every day, the same person comes at exactly 11 am and orders a cappuccino. The specific observation is that this person has come to the cafe at the same time and ordered the same thing every day during the period observed. A general conclusion drawn from these premises could be that this person always comes to the cafe at the same time and orders the same thing.
While inductive reasoning can be useful, it’s prone to being flawed. That’s because conclusions drawn using induction go beyond the information contained in the premises. An inductive argument may be highly probable , but even if all the observations are accurate, it can lead to incorrect conclusions.
Follow up this discussion with a look at concurrent vs. consecutive .
In our basic example, there are a number of reasons why it may not be true that the person always comes at the same time and orders the same thing.
Additional observations of the same event happening in the same way increase the probability that the event will happen again in the same way, but you can never be completely certain that it will always continue to happen in the same way.
That’s why a theory reached via inductive reasoning should always be tested to see if it is correct or makes sense.
What else does inductive mean?
Inductive can also be used as a synonym for introductory . It’s also used in a more specific way to describe the scientific processes of electromagnetic and electrostatic induction —or things that function based on them.
What does deductive mean?
Deductive reasoning (also called deduction ) involves starting from a set of general premises and then drawing a specific conclusion that contains no more information than the premises themselves. Deductive reasoning is sometimes called deduction (note that deduction has other meanings in the contexts of mathematics and accounting).
Here’s an example of deductive reasoning: chickens are birds; all birds lay eggs; therefore, chickens lay eggs. Another way to think of it: if something is true of a general class (birds), then it is true of the members of the class (chickens).
Deductive reasoning can go wrong, of course, when you start with incorrect premises. For example, look where this first incorrect statement leads us: all animals that lay eggs are birds; snakes lay eggs; therefore, snakes are birds.
The scientific method can be described as deductive . You first formulate a hypothesis —an educated guess based on general premises (sometimes formed by inductive methods). Then you test the hypothesis with an experiment . Based on the results of the experiment, you can make a specific conclusion as to the accuracy of your hypothesis.
You may have deduced there are related terms to this topic. Start with a look at interpolation vs. extrapolation .
Deductive reasoning is popularly associated with detectives and solving mysteries. Most famously, Sherlock Holmes claimed to be among the world’s foremost practitioners of deduction , using it to solve how crimes had been committed (or impress people by guessing where they had been earlier in the day).
However, despite this association, reasoning that’s referred to as deduction in many stories is actually more like induction or a form of reasoning known as abduction , in which probable but uncertain conclusions are drawn based on known information.
Sherlock’s (and Arthur Conan Doyle ’s) use of the word deduction can instead be interpreted as a way (albeit imprecise) of referring to systematic reasoning in general.
What is the difference between inductive vs. deductive reasoning?
Inductive reasoning involves starting from specific premises and forming a general conclusion, while deductive reasoning involves using general premises to form a specific conclusion.
Conclusions reached via deductive reasoning cannot be incorrect if the premises are true. That’s because the conclusion doesn’t contain information that’s not in the premises. Unlike deductive reasoning, though, a conclusion reached via inductive reasoning goes beyond the information contained within the premises—it’s a generalization , and generalizations aren’t always accurate.
The best way to understand the difference between inductive and deductive reasoning is probably through examples.
Go Behind The Words!
- Email This field is for validation purposes and should be left unchanged.
Examples of inductive and deductive reasoning
Examples of inductive reasoning.
Premise: All known fish species in this genus have yellow fins. Conclusion: Any newly discovered species in the genus is likely to have yellow fins.
Premises: This volcano has erupted about every 500 years for the last 1 million years. It last erupted 499 years ago. Conclusion: It will erupt again soon.
Examples of deductive reasoning
Premises: All plants with rainbow berries are poisonous. This plant has rainbow berries. Conclusion: This plant is poisonous.
Premises: I am lactose intolerant. Lactose intolerant people get sick when they consume dairy. This milkshake contains dairy. Conclusion: I will get sick if I drink this milkshake.
Reason your way to the best score by taking our quiz on "inductive" vs. "deductive" reasoning!
Science & Technology

Mixed-up Meanings
Word Origins
Pop Culture
Trending Words
Homepage » Logic » Arguments Index » Deductive and Inductive Arguments

Deductive and Inductive Arguments
Abstract : A deductive argument's premises provide conclusive evidence for the truth of its conclusion. An inductive argument's premises provide probable evidence for the truth of its conclusion. The difference between deductive and inductive arguments does not specifically depend on the specificity or generality of the composite statements. Both kinds of arguments are characterized and distinguished with examples and exercises.
- Deductive Arguments Defined
- Some Types of Deductive Arguments
- Inductive Arguments Defined
- Some Types of Inductive Arguments
- Specificity and Generality of Statements Do Not Always Distinguish Deduction From Induction
How to Distinguish Inductive Arguments from Deductive Arguments:
- Additional Examples Distinguishing Deduction and Induction
The Difference between Deduction and Induction:
“When an argument is such that the truth of the premises guarantees the truth of the conclusion, we shall say that it is deductively valid. When an argument is not deductively valid but nevertheless the premises provide good evidence for the conclusion, the argument is said to be inductively strong.” [2]
Deductive Arguments Defined:
All men are mortal. Socrates is a man. Therefore, Socrates is mortal.
All B is [in] C . All A is [in] B . ∴ All A is [in] C .
- If the premises are true and they necessitate the truth of the conclusion, then the argument is said to be deductively valid and sound . In such a case, it is impossible for the conclusion to be logically inconsistent with the premises.

Some Examples of Types of Deductive Arguments:
”Peter is John's brother, so John must be Peter's brother.” The argument is deductive since it relies on the lexical definition of “brother.” (Note this trivial deductive argument has no general statements.)
“Mystery is delightful, but unscientific, (2) since it depends upon ignorance.“ [4]
c. “Grant that the phenomena of intelligence conform to laws; grant that the evolution of intelligence in a child also conforms to laws; and it follows inevitably that education cannot be rightly guided without a knowledge of those laws.” [5]

No druggist is a chemist. That's because all apothecaries are chemists.
All [apothecaries] are [chemists].
→ {No [apothecaries] are [druggists]}
No [druggists] are [chemists].
E.g. , “Since a shell weighing 64 lbs leaves a gun with a velocity of 3,000 feet per second, and arrives at a target with a striking velocity of 500 feet per second, 11,250 BTU of heat resistance is generated.” [6]
- Logical inferences : arguments which can be described by symbolic notation in order to simplify the relationships among the structures (rather being evaluated from the meaning of the statements themselves). E.g. , “If you w ork hard, then you will s ucceed, and if you s ucceed, then you will be h appy; therefore, if you w ork hard, you will be h appy.” Given w → s → h , it follows w → h (These types of inferences follow from the truths of logic. The logic rule used here is called a hypothetical syllogism.)
Inductive Arguments Defined:
- Inductive arguments can range in probability from very low to very high, but always less than 100%. The probability of the conclusion drawn from an inductive argument is only an estimate and usually not known exactly. [7] (Note that the mathematical calculations in statistical reasoning are deductive even though the conclusions themselves are only probable. In other words, in statistics, the probability expressed in the conclusion follows from the premises with mathematical necessity.)
- Often (but not always!) induction is the sort of inference which attempts to reach a conclusion concerning all the members of a class or group on the basis of the observations of only some of them. So to put it another way, the conclusion of a very strong inductive argument with true premises is improbably false.
“I've seen many persons with creased earlobes who have had heart attacks, so I conclude that (all) persons who have creased earlobes are prone to have heart attacks.” [8]
Some Examples of Types of Inductive Arguments:

Modal Verbs and Probability Indicators:
Extrapolation:.
“A systematic evaluation of genotoxic responses will allow us to determine how genotoxic effects in rodents extrapolate to similar effects in humans. Research has already indicated that human cells may be more capable than rodent cells of repairing at least some DNA lesions, implying that human cells may be less sensitive to genotoxic agents.” [10]
“Since past experience indicates that irrigation is necessary for sustained production, the cost of a commercial grove with irrigation facilities would probably be at least $200.00 per acre higher than the official estimate.” [11]
One bird species with one color-form in the same population has been shown to be relatively stable over time, so all bird species with one color-form in that same population will remain relatively stable over time, as well.
“According to a Jenkins Group survey, 42% of college graduates will never read another book. Since most people read bestsellers printed in the past 10 years, it follows that virtually no one is reading the classics.” [13]
”[The reason] as to why productivity has slumped since 2004 is a simple one. That year coincided with the creation of Facebook .” [14]
“I share … [a] disrespect for religious certitude, which is a simulacrum of faith; but suggest that scientific certitude is barely less lethal. Just as we do not judge the value of science by nuclear weapons, pollution and junk food, we should not judge religion by its abuses.” [15]
Specificity and Generality of Statements Do Not Always Distinguish Deductive Arguments from Inductive Arguments:
All organisms have chromosomes. [ This fruit fly is an organism.] ∴ This fruit fly has chromosomes.
A red-eyed fruit fly has large chromosomes. A white-eyed fruit fly has large chromosomes. A Hawaiian fruit fly has large chromosomes. ∴ All fruit flies have large chromosomes.
Only Plato and Aristotle were great Greek philosophers. Plato and Aristotle lived in Athens. ∴ All the great Greek philosophers lived in Athens.
Each senator was present at today's session. ∴ All senators were present at today's session.
Entities E 1 , E 2 , and E 3 all have property p . Entities E 1 , E 2 , and E 3 are the only members of class M . ∴ All members of class M have property p .
All the great Greek philosophers wrote treatises on science. All philosophers named Aristotle wrote treatises on science. ∴ Aristotle was a great Greek philosopher. [17]
- The whale is a mammal. [as in an encyclopedia entry]
- All novelists of Waverly named Sir Walter Scott are historical writers. [a definite description]
- All present kings of France are bald. [a non-existent entity]
- All ideal gases are perfectly elastic. [a theoretical entity or nonobservable entity]
Begging the Question:
George is a man. George is 100 years old. George has arthritis. ∴ George will not run a four-minute mile. [19]
- For example adding the information that that George has a sprained ankle, a broken leg, and a heart condition makes it even less likely that George can run a 4 minute mile.
- However, when we add the premise that George is paraplegic, then the argument is transformed into a deductive argument because now the conclusion follows with certainty by the meanings of the words used in the statements.
Two performers in the Kronos Quartet play violin, one plays viola and another plays cello. ∴ The Kronos Quartet is composed of performers who all play stringed instruments.
- In valid deductive arguments, if the premises are true, then the truth of the conclusion follows with certainty.
“If we hate a person, we hate something in him that is part of ourselves. What isn't part of ourselves doesn't disturb us.” [20]
What isn't part of ourselves doesn't disturb us. ∴ If we hate a person, we hate something in him that is part of ourselves.
All things disturbing us are things part of ourselves. ∴ Our hating a person is hating something in him which is part of ourselves.
All [ things disturbing us ] are [things part of ourselves]. → {[Our hating a person] is [a thing that disturbs us ].} ∴ [Our hating a person] is [hating a thing part of ourselves].
Additional Examples Distinguishing Deduction and Induction:
- All throughout history people repeat the same mistakes, so we can conclude that similar mistakes will be made in the future. Answer Inductive Argument —The conclusion does not follow with absolute certainty. The reasoning assumes that the future will be in some sense like the past.
- The whale is a mammal, so all killer whales are mammals. Answer Deductive Argument — With the implicit premise that killer whales are whales, the conclusion follows with absolute certainty. In this example, the reasoning does proceed from general to less general, but the first general statement can be misleading to some persons.
- All killer whales are mammals, so the whale is a mammal. Answer Inductive Argument — As the argument stands, the conclusion is only probable. Notice that the reasoning is from part to whole even though the argument “appears” to be reasoned from general to specific. Even if it is assumed that all persons know whales are necessarily mammals, the reasoning in this argument is that the reason whales are mammals is due to one of its subclasses (killer whales) being mammals. This reason, considered by itself, is insufficient to prove the truth of the conclusion.
- “Because of our preoccupation with the present moment and the latest discovery, we do not read the great books of the past. Because we do not do this sort of reading, and do not think it is important, we do not bother about trying to learn to read difficult books. As a result, we do not learn to read well at all.” [21] Answer Inductive Argument and/or Explanation — Depending on the context of the passage, it is most likely to be an explanation as to why many persons do not read well rather than an argument proving why we do not read well. If it is evaluated as an argument, then it would be inductive, since it is possible for someone who has already learned to read well to be preoccupied at the present moment and that is why that person does not now read great or difficult works.
All persons who only look upon friends for profit are people who do not seek friendship without some ulterior motive. ∴ They seek only to profit from friends and don't look solely for friendship-in-itself ( i.e. , a friendship without an ulterior motive.) “Africans are notoriously religious, and each people has it own religious system with a set of beliefs and practices. Religion permeates into all the departments of life so fully, that it is not easy or possible always to isolate it. A study of these religious systems, is, therefore, ultimately a study of the peoples themselves in all the complexities of other traditional and modern life.” [23] Answer Inductive Argument — The argument is a strong inductive argument since a premise indicates it is not always easy or possible to study each people apart from their religion, suggesting that in some cases studying some people without considering their religion might be possible. So the conclusion does not follow with absolute certainty. It might be surprising to note that had the conclusion substituted the phrase “almost always” for “ultimately”, the argument would have been deductive. That is, the argument would be comparable to the following simplification: Most African people's religious beliefs are integrated into their lives. A study of African religions involves studying how most African peoples live.
- “Political change is a process that consolidates privilege, which further entrenches the oligarchic order. Thus the every widening economic gap — inevitably, a cultural and political gap—between the rich and the poor.” [24] Answer Deductive Argument — The conclusion is claimed to follow with complete certainty as indicated by the adverb “inevitably” in the conclusion of the argument.
- “[A]nxiety sufferers … may respond to their feelings by leaving nothing to chance. At work, they appear polished and prepared when giving a presentation because they consider every question that could be posed by colleagues beforehand and memorize possible answer in the days leading up to a meeting.” [25] Answer Inductive Argument — Since the first sentence, the conclusion of this argument, includes the modal verb “may” the conclusion is not claimed to follow with certainty. Hence, this argument is inductive since the conclusion is claimed to follow with some degree of probability.
- ”The exegete needs to possess not only scholarly training but also the continual deepening of his own meditative experience, so that he will not introduce mistaken interpretations into the text. Thus he is required not only to read the subject deeply enough to penetrate its key themes, but also to meditate in order to have the necessary mental purity and ‘wisdom eye&rsqquo; to carry out the work.” [26] Answer Deductive Argument —The meaning of the premise phrasing “scholarly training” and “deepening meditative experience” respectively imply the meaning of the conclusion phrasings ”reading deeply” and “meditate in order to have the necessary mental purity …” So the conclusion follows from the premise by the meaning of the words used.
- “One thing upon which Africana scholars and intellectuals largely agree is that the criteria used to define what is and what is not philosophy in the world today are unfairly biased by and for ‘philosophy’ as presently construed by Western culture. There may have to be some common ground if the work ‘philosophy’ is to continue to have cross-cultural significance. But Africa, in particular, has not received just consideration in that regard … In so many respects, it seems, Africa's cultures have not benefited from the kinds of exhaustive and empathetic scholarship that are being lavished upon other parts of the world. The oral literature of the African continent, therefore, has not even begun to receive the attention it merits.” [27] Answer Inductive Argument — Since most Africana scholars agree Western culture's definition of philosophy is biased, this assumed bias is reported to be the reason the Western definition has little common ground with the oral literature of Africa and the reason African philosophy is underrepresented in the world. This argument is inductive as its claim is based on a consensus rather than universal accord. Note that this argument is not an ad populum fallacy (an inappropriate appeal to the people) nor an ad verecundiam fallacy (an appeal to an irrrelevant authority) since Africana scholars and intellectuals are proper authorities as to the nature of African thought.

Deduction and Induction Notes
1. Richard Whately pointed out in 1831 that induction can be stated as a syllogism with a suppressed universal major premise which is substantially “what belongs to the individual or individuals we have examined, belongs to the whole class under which they come.” [Richard Whately, Elements of Logic (London: B. Fellowes, 1831), 230.] This influential text led many early logicians ( e.g. , John Stuart Mill) to think mistakenly that inductive logic can be somehow transformed into demonstrative reasoning. Following, George Henrik von Wright's A Treatise on Induction and Probability (1951 Abingdon, Oxon: Routledge, 2003. doi: 10.4324/9781315823157 ), logicians have abandoned this program [ C.f. , 29-30].
There is some controversy in the recent informal logic movement as to whether conductive, abductive, analogical, plausible, and other arguments can be classified as either inductive or deductive. Conductive, abductive and analogical arguments in this course are interpreted and reconstructed as inductive arguments.
A conductive argument is a complex argument which provides premises which separately provide evidence for a conclusion — each is independently relevant to the conclusion. Conductive arguments can also provide evidence for and against a conclusion (as in evaluations or decision).
Abductive argument is a process of selecting hypotheses which best explain a state of affairs very much like inference to the best explanation.
An analogical argument specifies that events or entities alike in several respects are probably alike in other respects as well. See e.g. Yun Xie, “ Conductive Argument as a Mode of Strategic Maneuvering ,” Informal Logic 37 no. 1 (January, 2017), 2-22. doi: 10.22329/il.v37i1.4696 And Bruce N. Waller, “ Classifying and Analyzing Analogies ” Informal Logic 21 no. 3 (Fall 2001), 199-218. 10.22329/il.v21i3.2246 ↩
2. Bryan Skyrms, Choice and Chance: An Introduction to Inductive Logic (Dickenson, 1975), 6-7.
Some logicians argue that all arguments are exclusively either deductive or inductive, and there are no other kinds. Also, they claim deductive arguments can only be evaluated by deductive standards and inductive arguments can only be evaluated by inductive standards. [ E.g. , George Bowles, “The Deductive/Inductive Distinction,” Informal Logic 16 no. 3 (Fall, 1994), 160. doi: 10.22329/il.v16i3.2455 ]
Stephen Barker argues:
“Our definition of deduction must refer to what the speaker is claiming, if it is to allow us to distinguish between invalid deductions and nondeductions.”
[S.F. Barker, “Must Every Inference be Either Deductive or Inductive?,” in Philosophy in America ed. Max Black (1964 London: Routledge, 2013), 62.]
On the one hand, for monotonic reasoning, Barker's definition makes the tail wag the dog since on this view the distinction between the two kinds of arguments depends upon the arbitrary psychological factor of what type of argument someone declares it to be rather than the nature or character of the argument itself. On Barker's view (and many current textbook views), the speaker's claim determines whether an argument is deductive or inductive regardless of the structure of the argument itself.
Barker explains the distinction from a dialogical point of view:
“Suppose someone argues, ‘All vegetarians are teetotallers, and he's a teetotaller, so I think he's a vegetarian.’ Is this inference a definitely illegitimate deduction, or is it an induction which may possibly be logically legitimate? We cannot decide without considering whether the speaker is claiming that his conclusion is strictly guaranteed by the premises (in which case, the inference is a fallacious deduction) or whether he is merely claiming that the premises supply real reason for believing the conclusion (in which case, the inference is an induction which in an appropriate context might be legitimate).” [Barker, 66.]
On Barker's view, an invalid deduction cannot be considered a weak induction since, for him, deduction and induction are exclusive forms of argumentation. This is a popular view, but we do not follow this view in these notes. Trudy Govier points out:
“If arguers' intentions are to provide the basis for a distinction between deductive and inductive arguments which will be anything like the traditional one, those arguers will have to formulate their intentions with a knowledge of the difference between logical and empirical connection, and the distinction between considerations of truth and those of validity.”
[Trudy Govier, “ More on Deductive and Inductive Arguments ,” Informal Logic (formerly Informal Logic Newsletter ) 2 no. 3 (March, 1979), 8. doi: 10.22329/il.v2i3.2824 ]
This point is obvious for monotonic reasoning where arguments are evaluated independently of claims (1) by the person who espouses them or when (2) arguments are evaluated in terms of the principle of charity . Even for dialogical reasoning, a speaker's intention should not determine the distinction between inductive and inductive arguments, for few speakers are informed of the epistemological differences to begin with. ↩
3. “Intentional account” named by Robert Wachbrit, “ A Note on the Difference Between Deduction and Induction ,” Philosophy & Rhetoric 29 no. 2 (1996), 168. doi: 10.2307/40237896 (doi link not activated 2022.06.28) ↩
4. Bertrand Russell, The Analysis of Mind (London: George Allen & Unwin, 1921), 40. ↩
5. Herbert Spencer, Education: Intellectual, Moral and Physical (New York: D. Appleton, 1860), 45-46. ↩
6. O.B. Goldman, “ Heat Engineering ,” The International Steam Engineer 37 no. 2(February 1920), 96. ↩
7. Arguments in statistics and probability theory are mathematical idealizations and are considered deductive inferences since their probable conclusions are logically entailed by their probable premises by means of a “rule-based definitions.”
Consequently, even though the premises and conclusion of these arguments are only probable, the probabilistic conclusion necessarily follows from the truth of the probabilistic premises. The inference itself is claimed to be certain given the truth of the premises.
In a valid deductive argument the conclusion must be true, if the premises are true. The proper description of the truth value of the conclusion of a valid statistical argument is that the statistical result is true, if the premises are true. The truth of the probability value established in the conclusion is certain given the truth of the data provided in the premises. ↩
8. This inductive argument is suggested by this study: Aris P. Agouridis, Moses S. Elisaf, Devaki R. Nair, and Dimitri P. Mikhailidis, “ Ear Lobe Crease: A Marker of Coronary Artery Disease? ” Archives of Medical Science 11 no. 6 (December 10, 2015) 1145-1155. doi: 10.5114/aoms.2015.56340> ↩
9. Friedrich Schlegel, Lectures on the History of Literature: Ancient and Modern trans. Henry G. Bohn (London: George Bell & Sons, 1880), 34. ↩
10. R. Schoeny and W. Farland, “ Determination of Relative Rodent-Human Interspecies Sensitivities to Chemical Carcinogens/Mutagens, ” Research to Improve Health Risk Assessments (Washington, D.C.: U.S. Environmental Protection Agency, 1990), Appendix D, 44. ↩
11. Foreign Agriculture Circular (Washington D.C.: U.S. Department of Agriculture, 5 no. 64 (November, 1964), 4. ↩
12. This type of induction describes the most common variety: it's often called “induction by incomplete enumeration.” ↩
13. John Wesley, “ 10 Ways to Improve Your Mind by Reading the Classics ,” Pick the Brain: Grow Yourself (June 20, 2007). ↩
14. Adapted from Nikko Schaff, “Letters: Let the Inventors Speak,” Economist 460 no. 8820 (January 26, 2013), 16. ↩
16. Historically, from the time of Aristotle, the distinction between deduction and induction, more or less, has been described as:
“[I]nduction is a progression from singulars to universals … and induction is more calculated to persuade, is clearer, and according to sense more known, and common to many things.” [Aristotle, Top. I.xii 105a12-13;16-19 (trans. Owen)
“Induction, then, is that operation of the mind, by which we infer that what we know to be true in a particular case or cases, will be true in all cases which resemble the former in certain assignable respects. In other words, Induction is the process by which we conclude that what is true of certain individuals of a class is true of the whole class, or that what is true at certain times will be true in similar circumstances at all times.” [John Stuart Mill, A System of Logic 2 vols.(London: Longmans, Green, Reader, and Dyer,) I:333.]
“[D]eduction consists in passing from more general to less general truths; induction is the contrary process from less to more general truths.” [W. Stanley Jevons, The Principles of Science 2nd ed. rev. (1887 London: Macmillan, 1913), 11.]
This view remains a popular view and does distinguish many arguments correctly. However, since this characterization is not true in all instances of these arguments, this distinction is no longer considered correct in the discipline of logic.
William Whewell was perhaps the earliest philosopher to register a correction to the view that induction can be defined as a process of reasoning from specific statements to a generalization. Throughout his writings he explains that induction requires more than simply generalizing from an enumeration of facts. He suggests as early as 1831 that the facts must be brought together by the recognition of a new generality of the relationship among the facts by applying that general relation to each of the facts. See. esp. William Whewell, The Mechanical Euclid (Cambridge: J. and J.J. Deighton, 1837), 173-175; The Philosophy of the Inductive Sciences , vol. 2 (London: J.W. Parker and Sons, 1840), 214; On the Philosophy of Discovery (London: John W. Parker and Son, 1860), 254. ↩
17. Notice that if this argument were to be taken as a syllogism (which will be studied later in the course), it would be considered an invalid deductive argument. A valid deductive argument has its conclusion follow with necessity; when the conclusion does not logically follow as in the “great Greek philosophers” example, there still is some small bit of evidence for the truth of the conclusion, so the argument could be evaluated as an extremely weak inductive argument.
No matter what class names ( i.e. no matter what subjects and predicates) are substituted into the form or grammatical structure of this argument (assuming the statements themselves are not tautological in some sense), it could never be a valid deductive argument — even when all the statements in it happen to be true. ↩
18. P.F. Strawson distinguishes the particular and the general in this manner:
“[W]hen we refer to general things, we abstract from their actual distribution and limits, if they have any, as we cannot do when we refer to particulars. Hence, with general things, meaning suffices to determine reference. And with this is connected the tendency, on the whole dominant, to ascribe superior reality to particular things. Meaning is not enough, in their case, to determine the reference of their designations; the extra, contextual element is essential. … So general things may have instances, while particular things may not.”
P.F. Strawson, “ Particular and General ,” Proceedings of the Aristotelian Society New Series 54 no. 1 (1953-1954), 260. doi: 10.1093/aristotelian/54.1.233 Also by JStor (free access by registration). ↩
19. Bryan Skyrms, Choice and Chance: An Introduction to Inductive Logic (Dickenson, 1975), 7. ↩
20. Adapted from Hermann Hesse, Demian (Berlin: S. Fischer, 1925), 157. ↩
21. Mortimer J. Adler, How to Read a Book (New York: Simon and Schuster: 1940), 89. ↩
22. Marcus Tullius Cicero, Old Age in Letters of Marcus Tullius Cicero with his Treatises on Friendship and Old Age and Letters of Gaius Plinius Caecilius Secundus , trans. E.E. Shuckburgh and William Melmoth, Harvard Classics, vol. 9 (P.F. Collier & Son, 1909), 35. ↩
23. John S. Mbiti, African Religions & Philosophy (Oxford: Heinemann, 1969), 1. ↩
24. Ferdinand E. Marcos, The Democratic Revolution in the Philippines (Englewood Cliffs, NJ: Prentice-Hall, 1974), 93. Also, Ferdinand E. Marcos, Toward the New Society: Essays on Aspects of Philippine Development (Philippines: National Media, 1974), 7. ↩
25. Francine Russo, “The Personality Trait ‘Intolerance of Uncertainty’ Causes Anguish During COVID,” Scientific American Mind 33 no. 3 (May-June 2022), 14. Also, here: Francine Russo, “ The Personality Trait ‘Intolerance of Uncertainty’ … ” Scientific American (accessed June 25, 2022). ↩
26. Charles Muller, “ A Korean Contribution to the Zen Canon; The Oga Hae Seorui (Commentaries of Five Masters on the Diamond Sūtra) ,” in Zen Classics: Formative Text in the History of Zen Buddhism eds. Steven Heine and Dale S. Wright (Oxford: Oxford University Press, 2006), 54. ↩
27. Barry Hallen, Short History of African Philosophy 2nd.ed. (Bloomington: Indiana University, 2009), 21. ↩
Readings on Induction and Deduction
S.F. Barker, “Must Every Inference be Either Deductive or Inductive?,” in Philosophy in America ed. Max Black (1964 London: Routledge, 2013), 62. doi: 10.4324/9781315830636
George Bowles, “ The Deductive/Inductive Distinction ,” Informal Logic 16, no. 3 (Fall 1994), 159-184. doi: 10.22329/il.v16i3.2455
Trudy Govier, “ More on Deductive and Inductive Arguments ,” Informal Logic (formerly Informal Logic Newsletter ) 2 no. 3 (March, 1979), 7-8. doi: 10.22329/il.v2i3.2824
David Hitchcock, “ Deduction, Induction and Conduction ,” 3 no. 2 Informal Logic (formerly Informal Logic Newsletter ) (January, 1980), 7-15. doi: 10.22329/il.v3i2.2786
IEP Staff, “ Deduction and Induction ,” The Internet Encyclopedia of Philosophy
P.F. Strawson, “ Particular and General ,” Proceedings of the Aristotelian Society New Series 54 no. 1 (1953-1954), 233-260. Also by JStor (free access by registration). doi: 10.1093/aristotelian/54.1.233
Robert Wachbrit, “ A Note on the Difference Between Deduction and Induction ,” Philosophy & Rhetoric 29 no. 2 (1996), 168-178. doi: 10.2307/40237896 (doi link not activated 2022.06.25) JStor (free with registration)

Stack Exchange Network
Stack Exchange network consists of 181 Q&A communities including Stack Overflow , the largest, most trusted online community for developers to learn, share their knowledge, and build their careers.
Philosophy Stack Exchange is a question and answer site for those interested in the study of the fundamental nature of knowledge, reality, and existence. It only takes a minute to sign up.
Q&A for work
Connect and share knowledge within a single location that is structured and easy to search.
What is the difference between inference and deduction?
This question contains my question, but tries to accomplish too much at once; I would like a clear answer to the distinction between inference and deduction.
What is the difference between coming to a conclusion via inference and coming to a conclusion via deduction?
The way I understand it, we deduce conclusions by using inferences. Inferences are statements in the form "if X then Y" and when it turns out previous statements which we assume or have otherwise proven to be true match the "X" part, we call it deduction. Since we used that inference, we say that we obtained our conclusion via inference. (This seems to suggest that while they are different terms, whenever you obtain a conclusion via deduction, you also obtain that conclusion via inference, and vice versa.)
However, other sources claim that deduction must come from originally observed or assumed facts, and that after you deduce one conclusion, you can no longer use that conclusion to "deduce" more; it then becomes "inference".
Is there any widely agreed upon difference between "deduction" and "inference"? If so, what is it? If not, in what ways might the terms differ?
- See Inference : "Inferences are steps in reasoning, moving from premises to logical consequences. Charles Sanders Peirce divided inference into three kinds: deduction, induction , and abduction . Deduction is inference deriving logical conclusions from premises known or assumed to be true, with the laws of valid inference being studied in logic. Induction is inference from particular premises to a universal conclusion. Abduction is inference to the best explanation." – Mauro ALLEGRANZA Nov 30, 2018 at 8:35
- Deduction : "Deductive reasoning, also deductive logic, logical deduction is the process of reasoning from one or more statements ( premises ) to reach a logically certain conclusion ." – Mauro ALLEGRANZA Nov 30, 2018 at 8:35
- Deduction is a form of inference. – user20253 Nov 30, 2018 at 12:45
4 Answers 4
The term 'deduction' is often used rather loosely in ordinary English. Conan Doyle infamously used it to describe Sherlock Holmes' reasoning, whereas today we would say that what Holmes did was abductive reasoning, which is generally taken to mean reasoning to the best explanation. In logic, we only use 'deduction' to refer to reasoning where there is no possibility of the conclusion being false if the premises are true. It is frequently used, even more narrowly, only in cases where the reasoning relies on formal rules of implication, rather than semantic or model theoretic considerations.
'Inference' is a more general term and refers to any reasoning by which a conclusion is reached from premises. As such, it encompasses both deductive and non-deductive kinds of reasoning. If I see a friend who has been absent for two weeks and notice he has a suntan, I might well infer that he has just returned from holiday. This is not a certain inference, since there are other possible explanations, but it is the most likely. This would be an example of abductive reasoning. If I notice that every morning the sun rises, I might infer that it is likely to do so again tomorrow. Again, this is not certain, but it might be characterised as a plausible inductive inference.
To make matters slightly more confusing, 'inference' is sometimes used for the individual steps within an argument, and logicians traditionally use the term 'rules of inference' for the formal rules, such as modus ponens, that characterise deductive logic. Gilbert Harman, among others, has long argued that this usage is misleading and we should be careful to distinguish between logic and reasoning. He advocates using the term 'rules of implication' for these formal rules.
In any case, deduction and inference have nothing to do with whether your premises are direct observations, assumptions, reported facts, or were themselves inferred from other things. It does not matter where your premises come from.
- 1 To add to the confusion, there is the Deduction Theorem, which is a particular inference involving the discharge of an assumption. – Graham Kemp Nov 30, 2018 at 8:07
Inference is more abstract: The law says that dog over 20 lbs cannot board a plane. All adult German Shephards weigh over 20 lbs. Therefore, adult German Shephards are not allowed on planes.
A deduction would be more specific to a particular instance: Therefore my dog Mimsy is not allowed on my flight to Florida.
The difference between limbs and roots. Deduction is applying the implication.
Deduction and non-psychological logical relationships
I take the brief answer to be that deduction holds between propositions or statements :
If p then q p
This relationship of deductive validity, where the conclusion cannot be false if the premises are true, is non-psychological and holds regardless of anyone thinking it. Inference in contrast is a psychological process of reasoning and is totally dependent on thinking.
Inference and the psychological process of reasoning
'To infer is to change to or take up a position which seems to the thinker to account for or explain the presented data' (Alan White, 'Inference', The Philosophical Quarterly (1950-), Vol. 21, No. 85 (Oct., 1971), pp. 289-302: 292.)
Thus from the fact that my silver has been stolen and only the butler, who has a long history of criminal convictions for theft, and my angelic five-year-old niece, could have stolen it, I infer (I take up the position) that the butler stole the silver. This is my inference to the best explanation. It is (a) psychological and (b) open to error. By contrast deductively validity is non-psychological - a matter of purely logical relationships between propositions or statements - and my inference can be wrong given the presented data whereas in a deductively valid argument the conclusion cannot be false - wrong - given the premises.
Inference can be deductive reasoning (I might infer : 'If p then q; q; therefore p') but can be inductive, abductive or as in the example inference to the best explanation. It is not limited to deductive reasoning. Equally deductive validity is a logical relationship between propositions and statements which holds good whether anyone has reasoned it out or not; it is psychology-free.
This is only a partial answer. The most it attempts is to illustrate how to approach this question: I would like a clear answer to the distinction between inference and deduction.
Answers may differ depending on the logicians one is quoting. One can expect all of these answers to be clear, that is, internally consistent from any particular logician, but not that all logicians will agree on any one definition.
Here is how the authors of forallx use inference : (page 8)
So: we are interested in whether or not a conclusion follows from some premises. Don’t, though, say that the premises infer the conclusion. Entailment is a relation between premises and conclusions; inference is something we do. (So if you want to mention inference when the conclusion follows from the premises, you could say that one may infer the conclusion from the premises.)
For these authors there are subtle differences between entailment and inference.
They use deduction to describe "proof-theoretic" systems, such as "natural deduction", in contrast with semantic arguments using truth tables or interpretations: (page vi)
But entailment is not the only important notion. We will also consider the relationship of being consistent, i.e., of not being mutually contradictory. These notions can be defined semantically, using precise definitions of entailment based on interpretations of the language—or proof-theoretically, using formal systems of deduction.
One thing to note from this is reaching a usable definition of a term may require multiple concepts to keep track of, such as, entailment, consistency, mutually contradictory, semantic, interpretations, and formal system. A full understanding of inference and deduction may require understanding other terms as well.
To see how things might be done differently, Quine uses the two words in the following note: (page 88)
Frege was perhaps the first to distinguish clearly between axioms and the rules of inference whereby theorems are generated from the axioms. Once this distinction is drawn, a recursive characterization of the class of theorems is virtually at hand. But the highly explicit way of presenting formal deductive systems which is customary nowadays dates back only to Hilbert (1922) or Post (1921).
The important thing to observe besides any differences with the previous use of the words is that these terms not only have a definition but they also have a history. One way to acknowledge that history is to associate any definition with whomever is the source of that definition.
So, for a clear answer to the distinction between inference and deduction one needs to further specify which logician's definition of these terms one is interested in.
Because of the differences between logicians I don't have an answer to the final question: Is there any widely agreed upon difference between "deduction" and "inference"? If so, what is it? If not, in what ways might the terms differ?
I suggest, however, given the above, that one doubt any answer one might receive to such questions. Any answer to the differences between these terms should also be associated with the logician providing the description of those differences, because the chief way the terms differ is due to their different sources, that is, the different logicians providing those definitions.
P. D. Magnus, Tim Button with additions by J. Robert Loftis remixed and revised by Aaron Thomas-Bolduc, Richard Zach, forallx Calgary Remix: An Introduction to Formal Logic, Winter 2018. http://forallx.openlogicproject.org/
W. V. O. Quine, (1981) Mathematical Logic, Harvard

You must log in to answer this question.
Not the answer you're looking for browse other questions tagged logic ..
- Featured on Meta
- We are graduating the updated button styling for vote arrows
- AI/ML Tool examples part 3 - Title-Drafting Assistant
- We are updating our Code of Conduct and we would like your feedback
Hot Network Questions
- Was Ford the only person who became a US president without winning an election?
- Why do I get a list from an integral over a region when my integrand is scalar?
- Assign unique slugs to set of people. Slugs are as short as possible. (New elements only added to ensure uniqueness.)
- Yield curve bootstrapping not producing expected cash flow start date
- Is colour alone enough to convey warning?
- For a finite set A of positive reals, prove that the set A + A - A contains at least as many positive as negative elements
- Better way to specifying arguments for multivariable function
- Photo contrast in early morning
- Where does the colon (:) belong in the Apex SOQL query - to the operator or the variable?
- Positive and negative voltage readings of chemical cells
- How does social inequality differ from wealth inequality? (In a society with capitalism esque ideals)
- How would one protect one self from a suit while skiing?
- Check if a string is a date
- Ignorance and 5 hindrances
- What can cause two deep "stepped" grooves in a rotor?
- I will grab a taxi back. vs. I will grab back a taxi
- Does this joke format exist in English, and if so, who is it about?
- How to find transaction's original author in Bitcoin Core?
- Modeling uncertainty from known physics
- How does GPL differ from CC BY-NC-SA?
- Is an autopilot failure an emergency?
- What kind of heating would occur during a suborbital re-entry?
- Roots of this sextic
- Should replacement GCFI receptacle match the one being replaced or the breaker rating?
Your privacy
By clicking “Accept all cookies”, you agree Stack Exchange can store cookies on your device and disclose information in accordance with our Cookie Policy .
Have a language expert improve your writing
Run a free plagiarism check in 10 minutes, generate accurate citations for free.
- Knowledge Base
- Methodology
- Inductive vs. Deductive Research Approach | Steps & Examples
Inductive vs. Deductive Research Approach | Steps & Examples
Published on April 18, 2019 by Raimo Streefkerk . Revised on March 31, 2023.
The main difference between inductive and deductive reasoning is that inductive reasoning aims at developing a theory while deductive reasoning aims at testing an existing theory .
In other words, inductive reasoning moves from specific observations to broad generalizations . Deductive reasoning works the other way around.
Both approaches are used in various types of research , and it’s not uncommon to combine them in your work.

Table of contents
Inductive research approach, deductive research approach, combining inductive and deductive research, frequently asked questions about inductive vs deductive reasoning.
When there is little to no existing literature on a topic, it is common to perform inductive research , because there is no theory to test. The inductive approach consists of three stages:
- A low-cost airline flight is delayed
- Dogs A and B have fleas
- Elephants depend on water to exist
- Another 20 flights from low-cost airlines are delayed
- All observed dogs have fleas
- All observed animals depend on water to exist
- Low cost airlines always have delays
- All dogs have fleas
- All biological life depends on water to exist
Limitations of an inductive approach
A conclusion drawn on the basis of an inductive method can never be fully proven. However, it can be invalidated.
When conducting deductive research , you always start with a theory. This is usually the result of inductive research. Reasoning deductively means testing these theories. Remember that if there is no theory yet, you cannot conduct deductive research.
The deductive research approach consists of four stages:
- If passengers fly with a low cost airline, then they will always experience delays
- All pet dogs in my apartment building have fleas
- All land mammals depend on water to exist
- Collect flight data of low-cost airlines
- Test all dogs in the building for fleas
- Study all land mammal species to see if they depend on water
- 5 out of 100 flights of low-cost airlines are not delayed
- 10 out of 20 dogs didn’t have fleas
- All land mammal species depend on water
- 5 out of 100 flights of low-cost airlines are not delayed = reject hypothesis
- 10 out of 20 dogs didn’t have fleas = reject hypothesis
- All land mammal species depend on water = support hypothesis
Limitations of a deductive approach
The conclusions of deductive reasoning can only be true if all the premises set in the inductive study are true and the terms are clear.
- All dogs have fleas (premise)
- Benno is a dog (premise)
- Benno has fleas (conclusion)
Here's why students love Scribbr's proofreading services
Discover proofreading & editing
Many scientists conducting a larger research project begin with an inductive study. This helps them develop a relevant research topic and construct a strong working theory. The inductive study is followed up with deductive research to confirm or invalidate the conclusion. This can help you formulate a more structured project, and better mitigate the risk of research bias creeping into your work.
Remember that both inductive and deductive approaches are at risk for research biases, particularly confirmation bias and cognitive bias , so it’s important to be aware while you conduct your research.
Inductive reasoning is a bottom-up approach, while deductive reasoning is top-down.
Inductive reasoning takes you from the specific to the general, while in deductive reasoning, you make inferences by going from general premises to specific conclusions.
Inductive reasoning is a method of drawing conclusions by going from the specific to the general. It’s usually contrasted with deductive reasoning, where you proceed from general information to specific conclusions.
Inductive reasoning is also called inductive logic or bottom-up reasoning.
Deductive reasoning is a logical approach where you progress from general ideas to specific conclusions. It’s often contrasted with inductive reasoning , where you start with specific observations and form general conclusions.
Deductive reasoning is also called deductive logic.
Exploratory research aims to explore the main aspects of an under-researched problem, while explanatory research aims to explain the causes and consequences of a well-defined problem.
Explanatory research is used to investigate how or why a phenomenon occurs. Therefore, this type of research is often one of the first stages in the research process , serving as a jumping-off point for future research.
Exploratory research is often used when the issue you’re studying is new or when the data collection process is challenging for some reason.
You can use exploratory research if you have a general idea or a specific question that you want to study but there is no preexisting knowledge or paradigm with which to study it.
A research project is an academic, scientific, or professional undertaking to answer a research question . Research projects can take many forms, such as qualitative or quantitative , descriptive , longitudinal , experimental , or correlational . What kind of research approach you choose will depend on your topic.
Cite this Scribbr article
If you want to cite this source, you can copy and paste the citation or click the “Cite this Scribbr article” button to automatically add the citation to our free Citation Generator.
Streefkerk, R. (2023, March 31). Inductive vs. Deductive Research Approach | Steps & Examples. Scribbr. Retrieved May 24, 2023, from https://www.scribbr.com/methodology/inductive-deductive-reasoning/
Is this article helpful?
Raimo Streefkerk
Other students also liked, qualitative vs. quantitative research | differences, examples & methods, explanatory research | definition, guide, & examples, exploratory research | definition, guide, & examples, what is your plagiarism score.
- 5.4 Types of Inferences
- Introduction
- 1.1 What Is Philosophy?
- 1.2 How Do Philosophers Arrive at Truth?
- 1.3 Socrates as a Paradigmatic Historical Philosopher
- 1.4 An Overview of Contemporary Philosophy
- Review Questions
- Further Reading
- 2.1 The Brain Is an Inference Machine
- 2.2 Overcoming Cognitive Biases and Engaging in Critical Reflection
- 2.3 Developing Good Habits of Mind
- 2.4 Gathering Information, Evaluating Sources, and Understanding Evidence
- 2.5 Reading Philosophy
- 2.6 Writing Philosophy Papers
- 3.1 Indigenous Philosophy
- 3.2 Classical Indian Philosophy
- 3.3 Classical Chinese Philosophy
- 4.1 Historiography and the History of Philosophy
- 4.2 Classical Philosophy
- 4.3 Jewish, Christian, and Islamic Philosophy
- 5.1 Philosophical Methods for Discovering Truth
- 5.2 Logical Statements
- 5.3 Arguments
- 5.5 Informal Fallacies
- 6.1 Substance
- 6.2 Self and Identity
- 6.3 Cosmology and the Existence of God
- 6.4 Free Will
- 7.1 What Epistemology Studies
- 7.2 Knowledge
- 7.3 Justification
- 7.4 Skepticism
- 7.5 Applied Epistemology
- 8.1 The Fact-Value Distinction
- 8.2 Basic Questions about Values
- 8.3 Metaethics
- 8.4 Well-Being
- 8.5 Aesthetics
- 9.1 Requirements of a Normative Moral Theory
- 9.2 Consequentialism
- 9.3 Deontology
- 9.4 Virtue Ethics
- 9.6 Feminist Theories of Ethics
- 10.1 The Challenge of Bioethics
- 10.2 Environmental Ethics
- 10.3 Business Ethics and Emerging Technology
- 11.1 Historical Perspectives on Government
- 11.2 Forms of Government
- 11.3 Political Legitimacy and Duty
- 11.4 Political Ideologies
- 12.1 Enlightenment Social Theory
- 12.2 The Marxist Solution
- 12.3 Continental Philosophy’s Challenge to Enlightenment Theories
- 12.4 The Frankfurt School
- 12.5 Postmodernism
Learning Objectives
By the end of this section, you will be able to:
- Define deductive, inductive, and abductive inferences.
- Classify inferences as deductive, inductive, or abductive.
- Explain different explanatory virtues used in abductive reasoning.
Inferences can be deductive, inductive, or abductive. Deductive inferences are the strongest because they can guarantee the truth of their conclusions. Inductive inferences are the most widely used, but they do not guarantee the truth and instead deliver conclusions that are probably true. Abductive inferences also deal in probability.
Deductive Reasoning
Deductive inferences, which are inferences arrived at through deduction (deductive reasoning), can guarantee truth because they focus on the structure of arguments. Here is an example:
- Either you can go to the movies tonight, or you can go to the party tomorrow.
- You cannot go to the movies tonight.
- So, you can go to the party tomorrow.
This argument is good, and you probably knew it was good even without thinking too much about it. The argument uses “or,” which means that at least one of the two statements joined by the “or” must be true. If you find out that one of the two statements joined by “or” is false, you know that the other statement is true by using deduction. Notice that this inference works no matter what the statements are. Take a look at the structure of this form of reasoning:
- X or Y is true.
- X is not true.
- Therefore, Y is true.
By replacing the statements with variables, we get to the form of the initial argument above. No matter what statements you replace X and Y with, if those statements are true, then the conclusion must be true as well. This common argument form is called a disjunctive syllogism.
Valid Deductive Inferences
A good deductive inference is called a valid inference , meaning its structure guarantees the truth of its conclusion given the truth of the premises. Pay attention to this definition. The definition does not say that valid arguments have true conclusions. Validity is a property of the logical forms of arguments, and remember that logic and truth are distinct. The definition states that valid arguments have a form such that if the premises are true, then the conclusion must be true. You can test a deductive inference’s validity by testing whether the premises lead to the conclusion. If it is impossible for the conclusion to be false when the premises are assumed to be true, then the argument is valid.
Deductive reasoning can use a number of valid argument structures:
Disjunctive Syllogism :
- Therefore X.
Modus Ponens :
- If X, then Y.
- Therefore Y.
Modus Tollens :
- Therefore, not X.
You saw the first form, disjunctive syllogism, in the previous example. The second form, modus ponens, uses a conditional, and if you think about necessary and sufficient conditions already discussed, then the validity of this inference becomes apparent. The conditional in premise 1 expresses that X is sufficient for Y. So if X is true, then Y must be true. And premise 2 states that X is true. So the conclusion (the truth of Y) necessarily follows. You can also use your knowledge of necessary and sufficient conditions to understand the last form, modus tollens. Remember, in a conditional, the consequent is the necessary condition. So Y is necessary for X. But premise 2 states that Y is not true. Because Y must be the case if X is the case, and we are told that Y is false, then we know that X is also false. These three examples are only a few of the numerous possible valid inferences.
Invalid Deductive Inferences
A bad deductive inference is called an invalid inference . In invalid inferences, their structure does not guarantee the truth of the conclusion—that is to say, even if the premises are true, the conclusion may be false. This does not mean that the conclusion must be false, but that we simply cannot know whether the conclusion is true or false. Here is an example of an invalid inference:
- If it snows more than three inches, the schools are mandated to close.
- The schools closed.
- Therefore, it snowed more than three inches.
If the premises of this argument are true (and we assume they are), it may or may not have snowed more than three inches. Schools close for many reasons besides snow. Perhaps the school district experienced a power outage or a hurricane warning was issued for the area. Again, you can use your knowledge of necessary and sufficient conditions to understand why this form is invalid. Premise 2 claims that the necessary condition is the case. But the truth of the necessary condition does not guarantee that the sufficient condition is true. The conditional states that the closing of schools is guaranteed when it has snowed more than 3 inches, not that snow of more than 3 inches is guaranteed if the schools are closed.
Invalid deductive inferences can also take general forms. Here are two common invalid inference forms:
Affirming the Consequent:
- Therefore, X.
Denying the Antecedent:
- Therefore, not Y.
You saw the first form, affirming the consequent, in the previous example concerning school closures. The fallacy is so called because the truth of the consequent (the necessary condition) is affirmed to infer the truth of the antecedent statement. The second form, denying the antecedent, occurs when the truth of the antecedent statement is denied to infer that the consequent is false. Your knowledge of sufficiency will help you understand why this inference is invalid. The truth of the antecedent (the sufficient condition) is only enough to know the truth of the consequent. But there may be more than one way for the consequent to be true, which means that the falsity of the sufficient condition does not guarantee that the consequent is false. Going back to an earlier example, that a creature is not a dog does not let you infer that it is not a mammal, even though being a dog is sufficient for being a mammal. Watch the video below for further examples of conditional reasoning. See if you can figure out which incorrect selection is structurally identical to affirming the consequent or denying the antecedent.
The Wason Selection Task
Testing deductive inferences.
Earlier it was explained that logical analysis involves assuming the premises of an argument are true and then determining whether the conclusion logically follows, given the truth of those premises. For deductive arguments, if you can come up with a scenario where the premises are true but the conclusion is false, you have proven that the argument is invalid. An instance of a deductive argument where the premises are all true but the conclusion false is called a counterexample . As with counterexamples to statements, counterexamples to arguments are simply instances that run counter to the argument. Counterexamples to statements show that the statement is false, while counterexamples to deductive arguments show that the argument is invalid. Complete the exercise below to get a better understanding of coming up with counterexamples to prove invalidity.
Think Like a Philosopher
Using the sample arguments given, come up with a counterexample to prove that the argument is invalid. A counterexample is a scenario in which the premises are true but the conclusion is false. Solutions are provided below.
Argument 1:
- If an animal is a dog, then it is a mammal.
- Charlie is not a dog.
- Therefore, Charlie is not a mammal.
Argument 2:
- All desserts are sweet foods.
- Some sweet foods are low fat.
- So all desserts are low fat.
Argument 3:
- If Jad doesn’t finish his homework on time, he won’t go to the party.
- Jad doesn’t go to the party.
- Jad didn’t finish his homework on time.
When you have completed your work on the three arguments, check your answers against the solutions below.
Solution 1: Invalid. If you imagine that Charlie is a cat (or other animal that is not a dog but is a mammal), then both the premises are true, while the conclusion is false. Charlie is not a dog, but Charlie is a mammal.
Solution 2: Invalid. Buttercream cake is a counterexample. Buttercream cake is a dessert and is sweet, which shows that not all desserts are low fat.
Solution3: Invalid. Assuming the first two premises are true, you can still imagine that Jad is too tired after finishing his homework and decides not to go to the party, thus making the conclusion false.
Inductive Inferences
When we reason inductively, we gather evidence using our experience of the world and draw general conclusions based on that experience. Inductive reasoning (induction) is also the process by which we use general beliefs we have about the world to create beliefs about our particular experiences or about what to expect in the future. Someone can use their past experiences of eating beets and absolutely hating them to conclude that they do not like beets of any kind, cooked in any manner. They can then use this conclusion to avoid ordering a beet salad at a restaurant because they have good reason to believe they will not like it. Because of the nature of experience and inductive inference, this method can never guarantee the truth of our beliefs. At best, inductive inference generates only probable true conclusions because it goes beyond the information contained in the premises. In the example, past experience with beets is concrete information, but the person goes beyond that information when making the general claim that they will dislike all beets (even those varieties they’ve never tasted and even methods of preparing beets they’ve never tried).
Consider a belief as certain as “the sun will rise tomorrow.” The Scottish philosopher David Hume famously argued against the certainty of this belief nearly three centuries ago ([1748, 1777] 2011, IV, i). Yes, the sun has risen every morning of recorded history (in truth, we have witnessed what appears to be the sun rising, which is a result of the earth spinning on its axis and creating the phenomenon of night and day). We have the science to explain why the sun will continue to rise (because the earth’s rotation is a stable phenomenon). Based on the current science, we can reasonably conclude that the sun will rise tomorrow morning. But is this proposition certain ? To answer this question, you have to think like a philosopher, which involves thinking critically about alternative possibilities. Say the earth gets hit by a massive asteroid that destroys it, or the sun explodes into a supernova that encompasses the inner planets and incinerates them. These events are extremely unlikely to occur, although no contradiction arises in imagining that they could take place. We believe the sun will rise tomorrow, and we have good reason for this belief, but the sun’s rising is still only probable (even if it is nearly certain).
While inductive inferences are not always a sure thing, they can still be quite reliable. In fact, a good deal of what we think we know is known through induction. Moreover, while deductive reasoning can guarantee the truth of conclusions if the premises are true, many times the premises themselves of deductive arguments are inductively known. In studying philosophy, we need to get used to the possibility that our inductively derived beliefs could be wrong.
There are several types of inductive inferences, but for the sake of brevity, this section will cover the three most common types: reasoning from specific instances to generalities, reasoning from generalities to specific instances, and reasoning from the past to the future.
Reasoning from Specific Instances to Generalities
Perhaps I experience several instances of some phenomenon, and I notice that all instances share a similar feature. For example, I have noticed that every year, around the second week of March, the red-winged blackbirds return from wherever they’ve wintering. So I can conclude that generally the red-winged blackbirds return to the area where I live (and observe them) in the second week of March. All my evidence is gathered from particular instances, but my conclusion is a general one. Here is the pattern:
Instance 1 , Instance 2 , Instance 3 . . . Instance n --> Generalization
And because each instance serves as a reason in support of the generalization, the instances are premises in the argument form of this type of inductive inference:
Specific to General Inductive Argument Form:
- General Conclusion
Reasoning from Generalities to Specific Instances
Induction can work in the opposite direction as well: reasoning from accepted generalizations to specific instances. This feature of induction relies on the fact that we are learners and that we learn from past experiences and from one another. Much of what we learn is captured in generalizations. You have probably accepted many generalizations from your parents, teachers, and peers. You probably believe that a red “STOP” sign on the road means that when you are driving and see this sign, you must bring your car to a full stop. You also probably believe that water freezes at 32° Fahrenheit and that smoking cigarettes is bad for you. When you use accepted generalizations to predict or explain things about the world, you are using induction. For example, when you see that the nighttime low is predicted to be 30°F, you may surmise that the water in your birdbath will be frozen when you get up in the morning.
Some thought processes use more than one type of inductive inference. Take the following example:
Every cat I have ever petted doesn’t tolerate its tail being pulled. So this cat probably will not tolerate having its tail pulled.
Notice that this reasoner has gone through a series of instances to make an inference about one additional instance. In doing so, the reasoner implicitly assumed a generalization along the way. The reasoner’s implicit generalization is that no cat likes its tail being pulled. They then use that generalization to determine that they shouldn’t pull the tail of the cat in front of them now. A reasoner can use several instances in their experience as premises to draw a general conclusion and then use that generalization as a premise to draw a conclusion about a specific new instance.
Inductive reasoning finds its way into everyday expressions, such as “Where there is smoke, there is fire.” When people see smoke, they intuitively come to believe that there is fire. This is the result of inductive reasoning. Consider your own thought process as you examine Figure 5.5 .
Reasoning from Past to Future
We often use inductive reasoning to predict what will happen in the future. Based on our ample experience of the past, we have a basis for prediction. Reasoning from the past to the future is similar to reasoning from specific instances to generalities. We have experience of events across time, we notice patterns concerning the occurrence of those events at particular times, and then we reason that the event will happen again in the future. For example:
I see my neighbor walking her dog every morning. So my neighbor will probably walk her dog this morning.
Could the person reasoning this way be wrong? Yes—the neighbor could be sick, or the dog could be at the vet. But depending upon the regularity of the morning dog walks and on the number of instances (say the neighbor has walked the dog every morning for the past year), the inference could be strong in spite of the fact that it is possible for it to be wrong.
Strong Inductive Inferences
The strength of inductive inferences depends upon the reliability of premises given as evidence and their relation to the conclusions drawn. A strong inductive inference is one where, if the evidence offered is true, then the conclusion is probably true. A weak inductive inference is one where, if the evidence offered is true, the conclusion is not probably true. But just how strong an inference needs to be to be considered good is context dependent. The word “probably” is vague. If something is more probable than not, then it needs at least a 51 percent chance of happening. However, in most instances, we would expect to have a much higher probability bar to consider an inference to be strong. As an example of this context dependence, compare the probability accepted as strong in gambling to the much higher probability of accuracy we expect in determining guilt in a court of law.
Figure 5.6 illustrates three forms of reasoning are used in the scientific method. Induction is used to glean patterns and generalizations, from which hypotheses are made. Hypotheses are tested, and if they remain unfalsified, induction is used again to assume support for the hypothesis.
Abductive Reasoning
Abductive reasoning is similar to inductive reasoning in that both forms of inference are probabilistic. However, they differ in the relationship of the premises to the conclusion. In inductive argumentation, the evidence in the premises is used to justify the conclusion. In abductive reasoning, the conclusion is meant to explain the evidence offered in the premises. In induction the premises explain the conclusion, but in abduction the conclusion explains the premises.
Inference to the Best Explanation
Because abduction reasons from evidence to the most likely explanation for that evidence, it is often called “inference to the best explanation.” We start with a set of data and attempt to come up with some unifying hypothesis that can best explain the existence of those data. Given this structure, the evidence to be explained is usually accepted as true by all parties involved. The focus is not the truth of the evidence, but rather what the evidence means.
Although you may not be aware, you regularly use this form of reasoning. Let us say your car won’t start, and the engine won’t even turn over. Furthermore, you notice that the radio and display lights are not on, even when the key is in and turned to the ON position. Given this evidence, you conclude that the best explanation is that there is a problem with the battery (either it is not connected or is dead). Or perhaps you made pumpkin bread in the morning, but it is not on the counter where you left it when you get home. There are crumbs on the floor, and the bag it was in is also on the floor, torn to shreds. You own a dog who was inside all day. The dog in question is on the couch, head hanging low, ears back, avoiding eye contact. Given the evidence, you conclude that the best explanation for the missing bread is that the dog ate it.
Detectives and forensic investigators use abduction to come up with the best explanation for how a crime was committed and by whom. This form of reasoning is also indispensable to scientists who use observations (evidence) along with accepted hypotheses to create new hypotheses for testing. You may also recognize abduction as a form of reasoning used in medical diagnoses. A doctor considers all your symptoms and any further evidence gathered from preliminarily tests and reasons to the best possible conclusion (a diagnosis) for your illness.
Explanatory Virtues
Good abductive inferences share certain features. Explanatory virtues are aspects of an explanation that generally make it strong. There are many explanatory virtues, but we will focus on four. A good hypothesis should be explanatory, simple , and conservative and must have depth .
To say that a hypothesis must be explanatory simply means that it must explain all the available evidence. The word “explanatory” for our purposes is being used in a narrower sense than used in everyday language. Take the pumpkin bread example: a person might reason that perhaps their roommate ate the loaf of pumpkin bread. However, such an explanation would not explain why the crumbs and bag were on the floor, nor the guilty posture of the dog. People do not normally eat an entire loaf of pumpkin bread, and if they do, they don’t eviscerate the bag while doing so, and even if they did, they’d probably hide the evidence. Thus, the explanation that your roommate ate the bread isn’t as explanatory as the one that pinpoints your dog as the culprit.
But what if you reason that a different dog got into the house and ate the bread, then got out again, and your dog looks guilty because he did nothing to stop the intruder? This explanation seems to explain the missing bread, but it is not as good as the simpler explanation that your dog is the perpetrator. A good explanation is often simple . You may have heard of Occam’s razor , formulated by William of Ockham (1287–1347), which says that the simplest explanation is the best explanation. Ockham said that “entities should not be multiplied beyond necessity” (Spade & Panaccio 2019). By “entities,” Ockham meant concepts or mechanisms or moving parts.
Examples of explanations that lack simplicity abound. For example, conspiracy theories present the very opposite of simplicity since such explanations are by their very nature complex. Conspiracy theories must posit plots, underhanded dealings, cover-ups (to explain the existence of alternative evidence), and maniacal people to explain phenomena and to further explain away the simpler explanation for those phenomena. Conspiracy theories are never simple, but that is not the only reason they are suspect. Conspiracy theories also generally lack the virtues of being conservative and having depth .
A conservative explanation maintains or conserves much of what we already believe. Conservativeness in science is when a theory or hypothesis fits with other established scientific theories and explanations. For example, a theory that accounts for some physical phenomenon but also does not violate Newton’s first law of motion is an example of a conservative theory. On the other hand, consider the conspiracy theory that we never landed on the moon. Someone might posit that the televised Apollo 11 space landing was filmed in a secret studio somewhere. But the reality of the first televised moon landing is not the only belief we must get rid of to maintain the theory. Five more manned moon landings occurred. Furthermore, the reality of the moon landings fits into beliefs about technological advancement over the next five decades. Many of the technologies developed were later adopted by the military and private sector (NASA, n.d.). Moreover, the Apollo missions are a key factor in understanding the space race of the Cold War era. Accepting the conspiracy theory requires rejecting a wide range of beliefs, and so the theory is not conservative.
A conspiracy theorist may offer alternative explanations to account for the tension between their explanation and established beliefs. However, for each explanation the conspiracist offers, more questions are raised. And a good explanation should not raise more questions than it answers. This characteristic is the virtue of depth . A deep explanation avoids unexplained explainers, or an explanation that itself is in need of explanation. For example, the theorist might claim that John Glenn and the other astronauts were brainwashed to explain the astronauts’ firsthand accounts. But this claim raises a question about how brainwashing works. Furthermore, what about the accounts of the thousands of other personnel who worked on the project? Were they all brainwashed? And if so, how? The conspiracy theorist’s explanation raises more questions than it answers.
Extraordinary Claims Require Extraordinary Evidence
Is it possible that our established beliefs (or scientific theories) could be wrong? Why give precedence to an explanation because it upholds our beliefs? Scientific thought would never have advanced if we deferred to conservative explanations all the time. In fact, the explanatory virtues are not laws but rules of thumb, none of which are supreme or necessary. Sometimes the correct explanation is more complicated, and sometimes the correct explanation will require that we give up long-held beliefs. Novel and revolutionary explanations can be strong if they have evidence to back them up. In the sciences, this approach is expressed in the following principle: Extraordinary claims will require extraordinary evidence. In other words, a novel claim that disrupts accepted knowledge will need more evidence to make it credible than a claim that already aligns with accepted knowledge.
Table 5.2 summarizes the three types of inferences just discussed.
As an Amazon Associate we earn from qualifying purchases.
Want to cite, share, or modify this book? This book uses the Creative Commons Attribution License and you must attribute OpenStax.
Access for free at https://openstax.org/books/introduction-philosophy/pages/1-introduction
- Authors: Nathan Smith
- Publisher/website: OpenStax
- Book title: Introduction to Philosophy
- Publication date: Jun 15, 2022
- Location: Houston, Texas
- Book URL: https://openstax.org/books/introduction-philosophy/pages/1-introduction
- Section URL: https://openstax.org/books/introduction-philosophy/pages/5-4-types-of-inferences
© Jun 21, 2022 OpenStax. Textbook content produced by OpenStax is licensed under a Creative Commons Attribution License . The OpenStax name, OpenStax logo, OpenStax book covers, OpenStax CNX name, and OpenStax CNX logo are not subject to the Creative Commons license and may not be reproduced without the prior and express written consent of Rice University.

- Table of Contents
- Random Entry
- Chronological
- Editorial Information
- About the SEP
- Editorial Board
- How to Cite the SEP
- Special Characters
- Advanced Tools
- Support the SEP
- PDFs for SEP Friends
- Make a Donation
- SEPIA for Libraries
- Entry Contents
Bibliography
Academic tools.
- Friends PDF Preview
- Author and Citation Info
- Back to Top
Inductive Logic
An inductive logic is a logic of evidential support. In a deductive logic, the premises of a valid deductive argument logically entail the conclusion, where logical entailment means that every logically possible state of affairs that makes the premises true must make the conclusion true as well. Thus, the premises of a valid deductive argument provide total support for the conclusion. An inductive logic extends this idea to weaker arguments. In a good inductive argument, the truth of the premises provides some degree of support for the truth of the conclusion, where this degree-of-support might be measured via some numerical scale. By analogy with the notion of deductive entailment, the notion of inductive degree-of-support might mean something like this: among the logically possible states of affairs that make the premises true, the conclusion must be true in (at least) proportion r of them—where r is some numerical measure of the support strength.
If a logic of good inductive arguments is to be of any real value, the measure of support it articulates should be up to the task. Presumably, the logic should at least satisfy the following condition:
Criterion of Adequacy (CoA) : The logic should make it likely (as a matter of logic) that as evidence accumulates, the total body of true evidence claims will eventually come to indicate, via the logic’s measure of support , that false hypotheses are probably false and that true hypotheses are probably true.
The CoA stated here may strike some readers as surprisingly strong. Given a specific logic of evidential support, how might it be shown to satisfy such a condition? Section 4 will show precisely how this condition is satisfied by the logic of evidential support articulated in Sections 1 through 3 of this article.
This article will focus on the kind of the approach to inductive logic most widely studied by epistemologists and logicians in recent years. This approach employs conditional probability functions to represent measures of the degree to which evidence statements support hypotheses. Presumably, hypotheses should be empirically evaluated based on what they say (or imply) about the likelihood that evidence claims will be true. A straightforward theorem of probability theory, called Bayes’ Theorem, articulates the way in which what hypotheses say about the likelihoods of evidence claims influences the degree to which hypotheses are supported by those evidence claims. Thus, this approach to the logic of evidential support is often called a Bayesian Inductive Logic or a Bayesian Confirmation Theory . This article will first provide a detailed explication of a Bayesian approach to inductive logic. It will then examine the extent to which this logic may pass muster as an adequate logic of evidential support for hypotheses. In particular, we will see how such a logic may be shown to satisfy the Criterion of Adequacy stated above.
Sections 1 through 3 present all of the main ideas underlying the (Bayesian) probabilistic logic of evidential support. These three sections should suffice to provide an adequate understanding of the subject. Section 5 extends this account to cases where the implications of hypotheses about evidence claims (called likelihoods ) are vague or imprecise. After reading Sections 1 through 3, the reader may safely skip directly to Section 5, bypassing the rather technical account in Section 4 of how how the CoA is satisfied.
Section 4 is for the more advanced reader who wants an understanding of how this logic may bring about convergence to the true hypothesis as evidence accumulates. This result shows that the Criterion of Adequacy is indeed satisfied—that as evidence accumulates, false hypotheses will very probably come to have evidential support values (as measured by their posterior probabilities ) that approach 0; and as this happens, a true hypothesis may very probably acquire evidential support values (as measured by its posterior probability ) that approaches 1.
1. Inductive Arguments
2.1 the historical origins of probabilistic logic, 2.2 probabilistic logic: axioms and characteristics, 2.3 two conceptions of inductive probability, 3.1 likelihoods, 3.2 posterior probabilities and prior probabilities, 3.3 bayes’ theorem, 3.4 on prior probabilities and representations of vague and diverse plausibility assessments, 4.1 the space of possible outcomes of experiments and observations, 4.2 probabilistic independence, 4.3 likelihood ratio convergence when falsifying outcomes are possible, 4.4 likelihood ratio convergence when no falsifying outcomes are possible, 5. when the likelihoods are vague or diverse, list of supplements, other internet resources, related entries.
Let us begin by considering some common kinds of examples of inductive arguments. Consider the following two arguments:
Example 1. Every raven in a random sample of 3200 ravens is black. This strongly supports the following conclusion: All ravens are black.
Example 2. 62 percent of voters in a random sample of 400 registered voters (polled on February 20, 2004) said that they favor John Kerry over George W. Bush for President in the 2004 Presidential election. This supports with a probability of at least .95 the following conclusion: Between 57 percent and 67 percent of all registered voters favor Kerry over Bush for President (at or around the time the poll was taken).
This kind of argument is often called an induction by enumeration . It is closely related to the technique of statistical estimation. We may represent the logical form of such arguments semi-formally as follows:
Premise: In random sample S consisting of n members of population B , the proportion of members that have attribute A is r .
Therefore, with degree of support p ,
Conclusion: The proportion of all members of B that have attribute A is between \(r-q\) and \(r+q\) (i.e., lies within margin of error q of r ).
Let’s lay out this argument more formally. The premise breaks down into three separate statements: [ 1 ]
Any inductive logic that treats such arguments should address two challenges. (1) It should tell us which enumerative inductive arguments should count as good inductive arguments. In particular, it should tell us how to determine the appropriate degree p to which such premises inductively support the conclusion, for a given margin of error q . (2) It should demonstrably satisfy the CoA . That is, it should be provable (as a metatheorem) that if a conclusion expressing the approximate proportion for an attribute in a population is true, then it is very likely that sufficiently numerous random samples of the population will provide true premises for good inductive arguments that confer degrees of support p approaching 1 for that true conclusion—where, on pain of triviality, these sufficiently numerous samples are only a tiny fraction of a large population. The supplement on Enumerative Inductions: Bayesian Estimation and Convergence , shows precisely how a a Bayesian account of enumerative induction may meet these two challenges.
Enumerative induction is, however, rather limited in scope. This form of induction is only applicable to the support of claims involving simple universal conditionals (i.e., claims of form ‘All B s are A s’) and claims about the proportion of an attribute in a population (i.e., claims of form ‘the frequency of A s among the B s is r ’). But, many important empirical hypotheses are not reducible to this simple form, and the evidence for these hypotheses is not composed of an enumeration of such instances. Consider, for example, the Newtonian Theory of Mechanics:
All objects remain at rest or in uniform motion unless acted upon by some external force. An object’s acceleration (i.e., the rate at which its motion changes from rest or from uniform motion) is in the same direction as the force exerted on it; and the rate at which the object accelerates due to a force is equal to the magnitude of the force divided by the object’s mass. If an object exerts a force on another object, the second object exerts an equal amount of force on the first object, but in the opposite direction to the force exerted by the first object.
The evidence for (and against) this theory is not gotten by examining a randomly selected subset of objects and the forces acting upon them. Rather, the theory is tested by calculating what this theory says (or implies) about observable phenomena in a wide variety of specific situations—e.g., ranging from simple collisions between small bodies to the trajectories of planets and comets—and then seeing whether those phenomena occur in the way that the theory says they will. This approach to testing hypotheses and theories is ubiquitous, and should be captured by an adequate inductive logic.
More generally, for a wide range of cases where inductive reasoning is important, enumerative induction is inadequate. Rather, the kind of evidential reasoning that judges the likely truth of hypotheses on the basis of what they say (or imply) about the evidence is more appropriate. Consider the kinds of inferences jury members are supposed to make, based on the evidence presented at a murder trial. The inference to probable guilt or innocence is based on a patchwork of evidence of various kinds. It almost never involves consideration of a randomly selected sequences of past situations when people like the accused committed similar murders. Or, consider how a doctor diagnoses her patient on the basis of his symptoms. Although the frequency of occurrence of various diseases when similar symptoms have been present may play a role, this is clearly not the whole story. Diagnosticians commonly employ a form of hypothesis evaluation —e.g., would the hypothesis that the patient has a brain tumor account for his symptoms?; or are these symptoms more likely the result of a minor stroke?; or may some other hypothesis better account for the patient’s symptoms? Thus, a fully adequate account of inductive logic should explicate the logic of hypothesis evaluation , through which a hypothesis or theory may be tested on the basis of what it says (or "predicts") about observable phenomena. In Section 3 we will see how a kind of probabilistic inductive logic called "Bayesian Inference" or "Bayesian Confirmation Theory" captures such reasoning. The full logical structure of such arguments will be spelled out in that section.
2. Inductive Logic and Inductive Probabilities
Perhaps the oldest and best understood way of representing partial belief, uncertain inference, and inductive support is in terms of probability and the equivalent notion odds . Mathematicians have studied probability for over 350 years, but the concept is certainly much older. In recent times a number of other, related representations of partial belief and uncertain inference have emerged. Some of these approaches have found useful application in computer based artificial intelligence systems that perform inductive inferences in expert domains such as medical diagnosis. Nevertheless, probabilistic representations have predominated in such application domains. So, in this article we will focus exclusively on probabilistic representations of inductive support. A brief comparative description of some of the most prominent alternative representations of uncertainty and support-strength can be found in the supplement Some Prominent Approaches to the Representation of Uncertain Inference .
The mathematical study of probability originated with Blaise Pascal and Pierre de Fermat in the mid-17 th century. From that time through the early 19 th century, as the mathematical theory continued to develop, probability theory was primarily applied to the assessment of risk in games of chance and to drawing simple statistical inferences about characteristics of large populations—e.g., to compute appropriate life insurance premiums based on mortality rates. In the early 19 th century Pierre de Laplace made further theoretical advances and showed how to apply probabilistic reasoning to a much wider range of scientific and practical problems. Since that time probability has become an indispensable tool in the sciences, business, and many other areas of modern life.
Throughout the development of probability theory various researchers appear to have thought of it as a kind of logic. But the first extended treatment of probability as an explicit part of logic was George Boole’s The Laws of Thought (1854). John Venn followed two decades later with an alternative empirical frequentist account of probability in The Logic of Chance (1876). Not long after that the whole discipline of logic was transformed by new developments in deductive logic.
In the late 19 th and early 20 th century Frege, followed by Russell and Whitehead, showed how deductive logic may be represented in the kind of rigorous formal system we now call quantified predicate logic . For the first time logicians had a fully formal deductive logic powerful enough to represent all valid deductive arguments that arise in mathematics and the sciences. In this logic the validity of deductive arguments depends only on the logical structure of the sentences involved. This development in deductive logic spurred some logicians to attempt to apply a similar approach to inductive reasoning. The idea was to extend the deductive entailment relation to a notion of probabilistic entailment for cases where premises provide less than conclusive support for conclusions. These partial entailments are expressed in terms of conditional probabilities , probabilities of the form \(P[C \pmid B] = r\) (read “the probability of C given B is r ”), where P is a probability function, C is a conclusion sentence, B is a conjunction of premise sentences, and r is the probabilistic degree of support that premises B provide for conclusion C . Attempts to develop such a logic vary somewhat with regard to the ways in which they attempt to emulate the paradigm of formal deductive logic.
Some inductive logicians have tried to follow the deductive paradigm by attempting to specify inductive support probabilities solely in terms of the syntactic structures of premise and conclusion sentences. In deductive logic the syntactic structure of the sentences involved completely determines whether premises logically entail a conclusion. So these inductive logicians have attempted to follow suit. In such a system each sentence confers a syntactically specified degree of support on each of the other sentences of the language. Thus, the inductive probabilities in such a system are logical in the sense that they depend on syntactic structure alone. This kind of conception was articulated to some extent by John Maynard Keynes in his Treatise on Probability (1921). Rudolf Carnap pursued this idea with greater rigor in his Logical Foundations of Probability (1950) and in several subsequent works (e.g., Carnap 1952). (For details of Carnap’s approach see the section on logical probability in the entry on interpretations of the probability calculus , in this Encyclopedia .)
In the inductive logics of Keynes and Carnap, Bayes’ theorem, a straightforward theorem of probability theory, plays a central role in expressing how evidence comes to bear on hypotheses. Bayes’ theorem expresses how the probability of a hypothesis h on the evidence e , \(P[h \pmid e]\), depends on the probability that e should occur if h is true, \(P[e \pmid h]\), and on the probability of hypothesis h prior to taking the evidence into account, \(P[h]\) (called the prior probability of h ). (Later we’ll examine Bayes’ theorem in detail.) So, such approaches might well be called Bayesian logicist inductive logics. Other prominent Bayesian logicist attempts to develop a probabilistic inductive logic include the works of Jeffreys (1939), Jaynes (1968), and Rosenkrantz (1981).
It is now widely held that the core idea of this syntactic approach to Bayesian logicism is fatally flawed—that syntactic logical structure cannot be the sole determiner of the degree to which premises inductively support conclusions. A crucial facet of the problem faced by syntactic Bayesian logicism involves how the logic is supposed to apply in scientific contexts where the conclusion sentence is some scientific hypothesis or theory, and the premises are evidence claims. The difficulty is that in any probabilistic logic that satisfies the usual axioms for probabilities, the inductive support for a hypothesis must depend in part on its prior probability . This prior probability represents (arguably) how plausible the hypothesis is taken to be on the basis of considerations other than the observational and experimental evidence (e.g., perhaps due to various plausibility arguments). A syntactic Bayesian logicist must tell us how to assign values to these pre-evidential prior probabilities of hypotheses in a way that relies only on the syntactic logical structure of the hypothesis, perhaps based on some measure of syntactic simplicity. There are severe problems with getting this idea to work. Various kinds of examples seem to show that such an approach must assign intuitively quite unreasonable prior probabilities to hypotheses in specific cases (see the footnote cited near the end of Section 3.2 for details). Furthermore, for this idea to apply to the evidential support of real scientific theories, scientists would have to formalize theories in a way that makes their relevant syntactic structures apparent, and then evaluate theories solely on that syntactic basis (together with their syntactic relationships to evidence statements). Are we to evaluate alternative theories of gravitation, and alternative quantum theories, this way? This seems an extremely dubious approach to the evaluation of real scientific hypotheses and theories. Thus, it seems that logical structure alone may not suffice for the inductive evaluation of scientific hypotheses. (This issue will be treated in more detail in Section 3 , after we first see how probabilistic logics employ Bayes’ theorem to represent the evidential support for hypotheses as a function of prior probabilities together with evidential likelihoods .)
At about the time that the syntactic Bayesian logicist idea was developing, an alternative conception of probabilistic inductive reasoning was also emerging. This approach is now generally referred to as the Bayesian subjectivist or personalist approach to inductive reasoning (see, e.g., Ramsey 1926; De Finetti 1937; Savage 1954; Edwards, Lindman, & Savage 1963; Jeffrey 1983, 1992; Howson & Urbach 1993; Joyce 1999). This approach treats inductive probability as a measure of an agent’s degree-of-belief that a hypothesis is true, given the truth of the evidence. This approach was originally developed as part of a larger normative theory of belief and action known as Bayesian decision theory . The principal idea is that the strength of an agent’s desires for various possible outcomes should combine with her belief-strengths regarding claims about the world to produce optimally rational decisions. Bayesian subjectivists provide a logic of decision that captures this idea, and they attempt to justify this logic by showing that in principle it leads to optimal decisions about which of various risky alternatives should be pursued. On the Bayesian subjectivist or personalist account of inductive probability, inductive probability functions represent the subjective (or personal) belief-strengths of ideally rational agents, the kind of belief strengths that figure into rational decision making. (See the section on subjective probability in the entry on interpretations of the probability calculus , in this Encyclopedia .)
Elements of a logicist conception of inductive logic live on today as part of the general approach called Bayesian inductive logic . However, among philosophers and statisticians the term ‘Bayesian’ is now most closely associated with the subjectivist or personalist account of belief and decision. And the term ‘Bayesian inductive logic’ has come to carry the connotation of a logic that involves purely subjective probabilities. This usage is misleading since, for inductive logics, the Bayesian/non-Bayesian distinction should really turn on whether the logic gives Bayes’ theorem a prominent role, or the approach largely eschews the use of Bayes’ theorem in inductive inferences, as do the classical approaches to statistical inference developed by R. A. Fisher (1922) and by Neyman & Pearson (1967)). Indeed, any inductive logic that employs the same probability functions to represent both the probabilities of evidence claims due to hypotheses and the probabilities of hypotheses due to those evidence claims must be a Bayesian inductive logic in this broader sense; because Bayes’ theorem follows directly from the axioms that each probability function must satisfy, and Bayes’ theorem expresses a necessary connection between the probabilities of evidence claims due to hypotheses and the probabilities of hypotheses due to those evidence claims .
In this article the probabilistic inductive logic we will examine is a Bayesian inductive logic in this broader sense. This logic will not presuppose the subjectivist Bayesian theory of belief and decision, and will avoid the objectionable features of the syntactic version of Bayesian logicism. We will see that there are good reasons to distinguish inductive probabilities from degree-of-belief probabilities and from purely syntactic logical probabilities . So, the probabilistic logic articulated in this article will be presented in a way that depends on neither of these conceptions of what the probability functions are . However, this version of the logic will be general enough that it may be fitted to a Bayesian subjectivist or Bayesian syntactic-logicist program, if one desires to do that.
All logics derive from the meanings of terms in sentences. What we now recognize as formal deductive logic rests on the meanings (i.e., the truth-functional properties) of the standard logical terms. These logical terms, and the symbols we will employ to represent them, are as follows:
- ‘not’, ‘\({\nsim}\)’;
- ‘and’, ‘\(\cdot\)’;
- ‘inclusive or’, ‘\(\vee\)’;
- truth-functional ‘if-then’, ‘\(\supset\)’;
- ‘if and only if’, ‘\(\equiv\)’;
- ‘all’, ‘\(\forall\)’, and
- ‘some’, ‘\(\exists\)’;
- the identity relation, ‘=’.
The meanings of all other terms, the non-logical terms such as names and predicate and relational expressions, are permitted to “float free”. That is, the logical validity of deductive arguments depends neither on the meanings of the name and predicate and relation terms, nor on the truth-values of sentences containing them. It merely supposes that these non-logical terms are meaningful, and that sentences containing them have truth-values. Deductive logic then tells us that the logical structures of some sentences—i.e., the syntactic arrangements of their logical terms—preclude them from being jointly true of any possible state of affairs. This is the notion of logical inconsistency . The notion of logical entailment is inter-definable with it. A collection of premise sentences logically entails a conclusion sentence just when the negation of the conclusion is logically inconsistent with those premises.
An inductive logic must, it seems, deviate from the paradigm provided by deductive logic in several significant ways. For one thing, logical entailment is an absolute, all-or-nothing relationship between sentences, whereas inductive support comes in degrees-of-strength. For another, although the notion of inductive support is analogous to the deductive notion of logical entailment , and is arguably an extension of it, there seems to be no inductive logic extension of the notion of logical inconsistency —at least none that is inter-definable with inductive support in the way that logical inconsistency is inter-definable with logical entailment . Indeed, it turns out that when the unconditional probability of \((B\cdot{\nsim}A)\) is very nearly 0 (i.e., when \((B\cdot{\nsim}A)\) is “nearly inconsistent”), the degree to which B inductively supports A , \(P[A \pmid B]\), may range anywhere between 0 and 1.
Another notable difference is that when B logically entails A , adding a premise C cannot undermine the logical entailment—i.e., \((C\cdot B)\) must logically entail A as well. This property of logical entailment is called monotonicity . But inductive support is nonmonotonic . In general, depending on what \(A, B\), and C mean, adding a premise C to B may substantially raise the degree of support for A , or may substantially lower it, or may leave it completely unchanged—i.e., \(P[A \pmid (C\cdot B)]\) may have a value much larger than \(P[A \pmid B]\), or may have a much smaller value, or it may have the same, or nearly the same value as \(P[A \pmid B]\).
In a formal treatment of probabilistic inductive logic, inductive support is represented by conditional probability functions defined on sentences of a formal language L . These conditional probability functions are constrained by certain rules or axioms that are sensitive to the meanings of the logical terms (i.e., ‘not’, ‘and’, ‘or’, etc., the quantifiers ‘all’ and ‘some’, and the identity relation). The axioms apply without regard for what the other terms of the language may mean. In essence the axioms specify a family of possible support functions , \(\{P_{\beta}, P_{\gamma}, \ldots ,P_{\delta}, \ldots \}\) for a given language L . Although each support function satisfies these same axioms, the further issue of which among them provides an appropriate measure of inductive support is not settled by the axioms alone. That may depend on additional factors, such as the meanings of the non-logical terms (i.e., the names and predicate expressions) of the language.
A good way to specify the axioms of the logic of inductive support functions is as follows. These axioms are apparently weaker than the usual axioms for conditional probabilities. For instance, the usual axioms assume that conditional probability values are restricted to real numbers between 0 and 1. The following axioms do not assume this, but only that support functions assign some real numbers as values for support strengths. However, it turns out that the following axioms suffice to derive all the usual axioms for conditional probabilities (including the usual restriction to values between 0 and 1). We draw on these weaker axioms only to forestall some concerns about whether the support function axioms may assume too much, or may be overly restrictive.
Let L be a language for predicate logic with identity, and let ‘\(\vDash\)’ be the standard logical entailment relation—i.e., the expression ‘\(B \vDash A\)’ says “ B logically entails A ” and the expression ‘\(\vDash A\)’ says “ A is a tautology”. A support function is a function \(P_{\alpha}\) from pairs of sentences of L to real numbers that satisfies the following axioms:
- (1) \(P_{\alpha}[E \pmid F] \ne P_{\alpha}[G \pmid H]\) for at least some sentences \(E, F, G\), and H .
For all sentence \(A, B, C\), and D :
- (2) If \(B \vDash A\), then \(P_{\alpha}[A \pmid B] \ge P_{\alpha}[C \pmid D]\);
- (3) \(P_{\alpha}[A \pmid (B \cdot C)] = P_{\alpha}[A \pmid (C \cdot B)]\);
- (4) If \(C \vDash{\nsim}(B \cdot A)\), then either \[P_{\alpha}[(A \vee B) \pmid C] = P_{\alpha}[A \pmid C] + P_{\alpha}[B \pmid C]\] or else \[P_{\alpha}[E \pmid C] = P_{\alpha}[C \pmid C]\] for every sentence E ;
- (5) \(P_{\alpha}[(A \cdot B) \pmid C] = P_{\alpha}[A \pmid (B \cdot C)] \times P_{\alpha}[B \pmid C]\).
This axiomatization takes conditional probability as basic, as seems appropriate for evidential support functions . (These functions agree with the more usual unconditional probability functions when the latter are defined—just let \(P_{\alpha}[A] = P_{\alpha}[A \pmid (D \vee{\nsim}D)]\). However, these axioms permit conditional probabilities \(P_{\alpha}[A \pmid C]\) to remain defined even when condition statement C has probability 0—i.e., even when \(P_{\alpha}[C \pmid (D\vee{\nsim}D)] = 0\).)
Notice that conditional probability functions apply only to pairs of sentences, a conclusion sentence and a premise sentence. So, in probabilistic inductive logic we represent finite collections of premises by conjoining them into a single sentence. Rather than say,
A is supported to degree r by the set of premises \(\{B_1\), \(B_2\), \(B_3\),…, \(B_n\}\),
we instead say that
A is supported to degree r by the conjunctive premise \((((B_1\cdot B_2)\cdot B_3)\cdot \ldots \cdot B_n)\),
and write this as
The above axioms are quite weak. For instance, they do not say that logically equivalent sentences are supported by all other sentences to the same degree; rather, that result is derivable from these axioms (see result 6 below). Nor do these axioms say that logically equivalent sentences support all other sentences to the same degree; rather, that result is also derivable (see result 8 below). Indeed, from these axioms all of the usual theorems of probability theory may be derived. The following results are particularly useful in probabilistic logic. Their derivations from these axioms are provided in note 2. [ 2 ]
- If \(B \vDash A\), then \(P_{\alpha}[A \pmid B] = 1\).
- If \(C \vDash{\nsim}(B\cdot A)\), then either \[P_{\alpha}[(A \vee B) \pmid C] = P_{\alpha}[A \pmid C] + P_{\alpha}[B \pmid C]\] or else \(P_{\alpha}[E \pmid C] = 1\) for every sentence E .
- \(P_{\alpha}[{\nsim}A \pmid B] = 1 - P_{\alpha}[A \pmid B]\) or else \(P_{\alpha}[C \pmid B] = 1\) for every sentence C .
- \(1 \ge P_{\alpha}[A \pmid B] \ge 0\).
- If \(B \vDash A\), then \(P_{\alpha}[A \pmid C] \ge P_{\alpha}[B \pmid C]\).
- If \(B \vDash A\) and \(A \vDash B\), then \(P_{\alpha}[A \pmid C] = P_{\alpha}[B \pmid C]\).
- If \(C \vDash B\), then \(P_{\alpha}[(A\cdot B) \pmid C] = P_{\alpha}[(B\cdot A) \pmid C] = P_{\alpha}[A \pmid C]\).
- If \(C \vDash B\) and \(B \vDash C\), then \(P_{\alpha}[A \pmid B] = P_{\alpha}[A \pmid C]\).
- \(P_{\alpha}[B \pmid C] \gt 0\), then \[P_{\alpha}[A \pmid (B\cdot C)] = P_{\alpha}[B \pmid (A\cdot C)] \times \frac{P_{\alpha}[A \pmid C]}{P_{\alpha}[B \pmid C]}\] (this is a simple form of Bayes’ theorem).
- \(P_{\alpha}[(A\vee B) \pmid C] = P_{\alpha}[A \pmid C] + P_{\alpha}[B \pmid C] - P_{\alpha}[(A\cdot B) \pmid C]\).
- If \(\{B_1 , \ldots ,B_n\}\) is any finite set of sentences such that for each pair \(B_i\) and \(B_j, C \vDash{\nsim}(B_{i}\cdot B_{j})\) (i.e., the members of the set are mutually exclusive, given C ), then either \(P_{\alpha}[D \pmid C] = 1\) for every sentence D , or \[ P_{\alpha}[((B_1\vee B_2)\vee \ldots \vee B_n) \pmid C] = \sum ^{n}_{i=1} P_{\alpha}[B_i \pmid C]. \]
- If \(\{B_1 , \ldots ,B_n , \ldots \}\) is any countably infinite set of sentences such that for each pair \(B_i\) and \(B_j, C \vDash{\nsim}(B_{i}\cdot B_{j})\), then either \(P_{\alpha}[D \pmid C] = 1\) for every sentence D , or [ 3 ] \[ \lim_n P_{\alpha}[((B_1\vee B_2)\vee \ldots \vee B_n) \pmid C] = \sum^{\infty}_{i=1} P_{\alpha}[B_i \pmid C]. \]
Let us now briefly consider each axiom to see how plausible it is as a constraint on a quantitative measure of inductive support, and how it extends the notion of deductive entailment. First notice that each degree-of-support function \(P_{\alpha}\) on L measures support strength with some real number values, but the axioms don’t explicitly restrict these values to lie between 0 and 1. It turns out that the all support values must lie between 0 and 1, but this follows from the axioms, rather than being assumed by them. The scaling of inductive support via the real numbers is surely a reasonable way to go.
Axiom 1 is a non-triviality requirement. It says that the support values cannot be the same for all sentence pairs. This axiom merely rules out the trivial support function that assigns the same amount of support to each sentence by every sentence. One might replace this axiom with the following rule:
But this alternative rule turns out to be derivable from axiom 1 together with the other axioms.
Axiom 2 asserts that when B logically entail A , the support of A by B is as strong as support can possibly be. This comports with the idea that an inductive support function is a generalization of the deductive entailment relation, where the premises of deductive entailments provide the strongest possible support for their conclusions.
Axiom 3 merely says that \((B \cdot C)\) supports sentences to precisely the same degree that \((C \cdot B)\) supports them. This is an especially weak axiom. But taken together with the other axioms, it suffices to entail that logically equivalent sentences support all sentences to precisely the same degree.
Axiom 4 says that inductive support adds up in a plausible way. When C logically entails the incompatibility of A and B , i.e., when no possible state of affairs can make both A and B true together, the degrees of support that C provides to each of them individually must sum to the support it provides to their disjunction. The only exception is in those cases where C acts like a logical contradiction and supports all sentences to the maximum possible degree (in deductive logic a logical contradiction logically entails every sentence).
To understand what axiom 5 says, think of a support function \(P_{\alpha}\) as describing a measure on possible states of affairs. Read each degree-of-support expression of form ‘\(P_{\alpha}[D \pmid E] = r\)’ to say that the proportion of states of affairs in which D is true among those states of affairs where E is true is r . Read this way, axiom 5 then says the following. Suppose B is true in proportion q of all the states of affairs where C is true, and suppose A is true in fraction r of those states where B and C are true together. Then A and B should be true together in what proportion of all the states where C is true? In fraction r (the \((A\cdot B)\) part) of proportion q (the B portion) of all those states where C is true.
The degree to which a sentence B supports a sentence A may well depend on what these sentences mean. In particular it will usually depend on the meanings we associate with the non-logical terms (those terms other than the logical terms not , and , or , etc., the quantifiers , and identity ), that is, on the meanings of the names, and the predicate and relation terms of the language. For example, we should want
given the usual meanings of ‘bachelor’ and ‘married’, since “all bachelors are unmarried” is analytically true—i.e. no empirical evidence is required to establish this connection. (In the formal language for predicate logic, if we associate the meaning “is married” with predicate term ‘ M ’, the meaning “is a bachelor” with the predicate term ‘ B ’, and take the name term ‘ g ’ to refer to George, then we should want \(P_{\alpha}[{\nsim}Mg \pmid Bg] = 1\), since \(\forall x (Bx \supset{\nsim}Mx)\) is analytically true on this meaning assignment to the non-logical terms.) So, let’s associate with each individual support function \(P_{\alpha}\) a specific assignment of meanings ( primary intensions ) to all the non-logical terms of the language. (However, evidential support functions should not presuppose meaning assignments in the sense of so-called secondary intensions —e.g., those associated with rigid designators across possible states of affairs. For, we should not want a confirmation function \(P_{\alpha}\) to make
since we presumably want the inductive logic to draw on explicit empirical evidence to support the claim that water is made of H 2 O. Thus, the meanings of terms we associate with a support function should only be their primary intensions, not their secondary intensions.)
In the context of inductive logic it makes good sense to supplement the above axioms with two additional axioms. Here is the first of them:
- (6) If A is an axiom of set theory or any other piece of pure mathematics employed by the sciences, or if A is analytically true (i.e., if the truth of A depends only on the meanings of the words it contains, where the specific meanings for names and predicates are those associated with the particular support function \(P_{\alpha})\), then, for all sentences C , \(P_{\alpha}[A \pmid C] = P_{\alpha}[C \pmid C]\) (i.e., \(P_{\alpha}[A \pmid C] = 1)\).
Here is how axiom 6 applies to the above example, yielding \(P_{\alpha}[{\nsim}Mg \pmid Bg] = 1\) when the meaning assignments to non-logical terms associated with support function \(P_{\alpha}\) makes \(\forall x(Bx \supset{\nsim}Mx)\) analytically true. From axiom 6 (followed by results 7, 5, and 4) we have
thus, \(P_{\alpha}[{\nsim}Mg \pmid Bg] = 1\). The idea behind axiom 6 is that inductive logic is about evidential support for contingent claims. Nothing can count as empirical evidence for or against non-contingent truths. In particular, analytic truths should be maximally supported by all premises C .
One important respect in which inductive logic should follow the deductive paradigm is that the logic should not presuppose the truth of contingent statements. If a statement C is contingent, then some other statements should be able to count as evidence against C . Otherwise, a support function \(P_{\alpha}\) will take C and all of its logical consequences to be supported to degree 1 by all possible evidence claims. This is no way for an inductive logic to behave. The whole idea of inductive logic is to provide a measure of the extent to which premise statements indicate the likely truth-values of contingent conclusion statements. This idea won’t work properly if the truth-values of some contingent statements are presupposed by assigning them support value 1 on every possible premise. Such probability assignments would make the inductive logic enthymematic by hiding significant premises in inductive support relationships. It would be analogous to permitting deductive arguments to count as valid in cases where the explicitly stated premises are insufficient to logically entail the conclusion, but where the validity of the argument is permitted to depend on additional unstated premises. This is not how a rigorous approach to deductive logic should work, and it should not be a common practice in a rigorous approach to inductive logic.
Nevertheless, it is common practice for probabilistic logicians to sweep provisionally accepted contingent claims under the rug by assigning them probability 1 (regardless of the fact that no explicit evidence for them is provided). This practice saves the trouble of repeatedly writing a given contingent sentence B as a premise, since \(P_{\gamma}[A \pmid B\cdot C]\) will equal \(P_{\gamma}[A \pmid C]\) whenever \(P_{\gamma}[B \pmid C] = 1\). Although this convention is useful, such probability functions should be considered mere abbreviations for proper, logically explicit, non-enthymematic, inductive support relations. Thus, properly speaking, an inductive support function \(P_{\alpha}\) should not assign probability 1 to a sentence on every possible premise unless that sentence is either (i) logically true, or (ii) an axiom of set theory or some other piece of pure mathematics employed by the sciences, or (iii) unless according to the interpretation of the language that \(P_{\alpha}\) presupposes, the sentence is analytic (and so outside the realm of evidential support). Thus, we adopt the following version of the so-called “axiom of regularity”.
- (7) If, for all C , \(P_{\alpha}[A \pmid C] = P_{\alpha}[C \pmid C]\) (i.e., \(P_{\alpha}[A \pmid C] = 1\)), then A must be a logical truth or an axiom of set theory or some other piece of pure mathematics employed by the sciences, or A must be analytically true (according to the meanings of the terms of L associated with support function \(P_{\alpha})\).
Axioms 6 and 7 taken together say that a support function \(P_{\alpha}\) counts as non-contingently true, and so not subject to empirical support, just those sentences that are assigned probability 1 by every premise.
Some Bayesian logicists have proposed that an inductive logic might be made to depend solely on the logical form of sentences, as is the case for deductive logic. The idea is, effectively, to supplement axioms 1–7 with additional axioms that depend only on the logical structures of sentences, and to introduce enough such axioms to reduce the number of possible support functions to a single uniquely best support function. It is now widely agreed that this project cannot be carried out in a plausible way. Perhaps support functions should obey some rules in addition to axioms 1–7. But it is doubtful that any plausible collection of additional rules can suffice to determine a single, uniquely qualified support function. Later, in Section 3 , we will briefly return to this issue, after we develop a more detailed account of how inductive probabilities capture the relationship between hypotheses and evidence.
Axioms 1–7 for conditional probability functions merely place formal constraints on what may properly count as a degree of support function . Each function \(P_{\alpha}\) that satisfies these axioms may be viewed as a possible way of applying the notion of inductive support to a language L that respects the meanings of the logical terms, much as each possible truth-value assignment for a language represents a possible way of assigning truth-values to its sentences in a way that respects the meanings of the logical terms. The issue of which of the possible truth-value assignments to a language represents the actual truth or falsehood of its sentences depends on more than this. It depends on the meanings of the non-logical terms and on the state of the actual world. Similarly, the degree to which some sentences actually support others in a fully meaningful language must rely on something more than the mere satisfaction of the axioms for support functions. It must, at least, rely on what the sentences of the language mean, and perhaps on much more besides. But, what more? Perhaps a better understanding of what inductive probability is may provide some help by filling out our conception of what inductive support is about. Let’s pause to discuss two prominent views—two interpretations of the notion of inductive probability.
One kind of non-syntactic logicist reading of inductive probability takes each support function \(P_{\alpha}\) to be a measure on possible states of affairs. The idea is that, given a fully meaningful language (associated with support function \(P_{\alpha}\)) ‘\(P_{\alpha}[A \pmid B] = r\)’ says that among those states of affairs in which B is true, A is true in proportion r of them. There will not generally be a single privileged way to define such a measure on possible states of affairs. Rather, each of a number of functions \(P_{\alpha}\), \(P_{\beta}\), \(P_{\gamma}\),…, etc., that satisfy the constraints imposed by axioms 1–7 may represent a viable measure of the inferential import of the propositions expressed by sentences of the language. This idea needs more fleshing out, of course. The next section will provide some indication of how that might go.
Subjectivist Bayesians offer an alternative reading of the support functions. First, they usually take unconditional probability as basic, and take conditional probabilities as defined in terms of unconditional probabilities: the conditional probability ‘\(P_{\alpha}[A \pmid B]\)’ is defined as a ratio of unconditional probabilities:
Subjectivist Bayesians take each unconditional probability function \(P_{\alpha}\) to represent the belief-strengths or confidence-strengths of an ideally rational agent, \(\alpha\). On this understanding ‘\(P_{\alpha}[A] =r\)’ says, “the strength of \(\alpha\)’s belief (or confidence) that A is truth is r ”. Subjectivist Bayesians usually tie such belief strengths to how much money (or how many units of utility ) the agent would be willing to bet on A turning out to be true. Roughly, the idea is this. Suppose that an ideally rational agent \(\alpha\) would be willing to accept a wager that would yield (no less than) $ u if A turns out to be true and would lose him $1 if A turns out to be false. Then, under reasonable assumptions about the agent’s desire money, it can be shown that the agent’s belief strength that A is true should be
And it can further be shown that any function \(P_{\alpha}\) that expresses such betting-related belief-strengths on all statements in agent \(\alpha\)’s language must satisfy axioms for unconditional probabilities analogous to axioms 1–5. [ 4 ] Moreover, it can be shown that any function \(P_{\beta}\) that satisfies these axioms is a possible rational belief function for some ideally rational agent \(\beta\). These relationships between belief-strengths and the desirability of outcomes (e.g., gaining money or goods on bets) are at the core of subjectivist Bayesian decision theory . Subjectivist Bayesians usually take inductive probability to just be this notion of probabilistic belief-strength .
Undoubtedly real agents do believe some claims more strongly than others. And, arguably, the belief strengths of real agents can be measured on a probabilistic scale between 0 and 1, at least approximately. And clearly the inductive support of a hypothesis by evidence should influence the strength of an agent’s belief in the truth of that hypothesis—that’s the point of engaging in inductive reasoning, isn’t it? However, there is good reason for caution about viewing inductive support functions as Bayesian belief-strength functions, as we’ll see a bit later. So, perhaps an agent’s support function is not simply identical to his belief function, and perhaps the relationship between inductive support and belief-strength is somewhat more complicated.
In any case, some account of what support functions are supposed to represent is clearly needed. The belief function account and the logicist account (in terms of measures on possible states of affairs) are two attempts to provide this account. But let us put this interpretative issue aside for now. One may be able to get a better handle on what inductive support functions really are after one sees how the inductive logic that draws on them is supposed to work.
3. The Application of Inductive Probabilities to the Evaluation of Scientific Hypotheses
One of the most important applications of an inductive logic is its treatment of the evidential evaluation of scientific hypotheses. The logic should capture the structure of evidential support for all sorts of scientific hypotheses, ranging from simple diagnostic claims (e.g., “the patient is infected by the HIV”) to complex scientific theories about the fundamental nature of the world, such as quantum mechanics or the theory of relativity. This section will show how evidential support functions (a.k.a. Bayesian confirmation functions) represent the evidential evaluation of scientific hypotheses and theories. This logic is essentially comparative. The evaluation of a hypothesis depends on how strongly evidence supports it over alternative hypotheses.
Consider some collection of mutually incompatible, alternative hypotheses (or theories) about a common subject matter, \(\{h_1, h_2 , \ldots \}\). The collection of alternatives may be very simple, e.g., {“the patient has HIV”, “the patient is free of HIV”}. Or, when the physician is trying to determine which among a range of diseases is causing the patient’s symptoms, the collection of alternatives may consist of a long list of possible disease hypotheses. For the cosmologist, the collection of alternatives may consist of several distinct gravitational theories, or several empirically distinct variants of the “same” theory. Whenever two variants of a hypothesis (or theory) differ in empirical import, they count as distinct hypotheses. (This should not be confused with the converse positivistic assertion that theories with the same empirical content are really the same theory. Inductive logic doesn’t necessarily endorse that view.)
The collection of competing hypotheses (or theories) to be evaluated by the logic may be finite in number, or may be countably infinite. No realistic language contains more than a countable number of expressions; so it suffices for a logic to apply to countably infinite number of sentences. From a purely logical perspective the collection of competing alternatives may consist of every rival hypothesis (or theory) about a given subject matter that can be expressed within a given language — e.g., all possible theories of the origin and evolution of the universe expressible in English and contemporary mathematics. In practice, alternative hypotheses (or theories) will often be constructed and evidentially evaluated over a long period of time. The logic of evidential support works in much the same way regardless of whether all alternative hypotheses are considered together, or only a few alternative hypotheses are available at a time.
Evidence for scientific hypotheses consists of the results of specific experiments or observations. For a given experiment or observation, let ‘\(c\)’ represent a description of the relevant conditions under which it is performed, and let ‘\(e\)’ represent a description of the result of the experiment or observation, the evidential outcome of conditions \(c\).
The logical connection between scientific hypotheses and the evidence often requires the mediation of background information and auxiliary hypotheses. Let ‘\(b\)’ represent whatever background and auxiliary hypotheses are required to connect each hypothesis \(h_i\) among the competing hypotheses \(\{h_1, h_2 , \ldots \}\) to the evidence. Although the claims expressed by the auxiliary hypotheses within \(b\) may themselves be subject to empirical evaluation, they should be the kinds of claims that are not at issue in the evaluation of the alternative hypothesis in the collection \(\{h_1, h_2 , \ldots \}\). Rather, each of the alternative hypotheses under consideration draws on the same background and auxiliaries to logically connect to the evidential events. (If competing hypotheses \(h_i\) and \(h_j\) draw on distinct auxiliary hypotheses \(a_i\) and \(a_j\), respectively, in making logical contact with evidential claims, then the following treatment should be applied to the respective conjunctive hypotheses, \((h_{i}\cdot a_{i})\) and \((h_{j}\cdot a_{j})\), since these alternative conjunctive hypotheses will constitute the empirically distinct alternatives at issue.)
In cases where a hypothesis is deductively related to an outcome \(e\) of an observational or experimental condition \(c\) (via background and auxiliaries \(b\)), we will have either \(h_i\cdot b\cdot c \vDash e\) or \(h_i\cdot b\cdot c \vDash{\nsim}e\) . For example, \(h_i\) might be the Newtonian Theory of Gravitation. A test of the theory might involve a condition statement \(c\) that describes the results of some earlier measurements of Jupiter’s position, and that describes the means by which the next position measurement will be made; the outcome description \(e\) states the result of this additional position measurement; and the background information (and auxiliary hypotheses) \(b\) might state some already well confirmed theory about the workings and accuracy of the devices used to make the position measurements. Then, from \(h_i\cdot b\cdot c\) we may calculate the specific outcome \(e\) we expect to find; thus, the following logical entailment holds: \(h_i\cdot b\cdot c \vDash e\) . Then, provided that the experimental and observational conditions stated by \(c\) are in fact true, if the evidential outcome described by \(e\) actually occurs, the resulting conjoint evidential claim \((c\cdot e)\) may be considered good evidence for \(h_i\), given \(b\). (This method of theory evaluation is called the hypothetical-deductive approach to evidential support.) On the other hand, when from \(h_i\cdot b\cdot c\) we calculate some outcome incompatible with the observed evidential outcome \(e\), then the following logical entailment holds: \(h_i\cdot b\cdot c \vDash{\nsim}e\). In that case, from deductive logic alone we must also have that \(b\cdot c\cdot e \vDash{\nsim}h_i\) ; thus, \(h_i\) is said to be falsified by \(b\cdot c\cdot e\). The Bayesian account of evidential support we will be describing below extends this deductivist approach to include cases where the hypothesis \(h_i\) (and its alternatives) may not be deductive related to the evidence, but may instead imply that the evidential outcome is likely or unlikely to some specific degree r . That is, the Bayesian approach applies to cases where we may have neither \(h_i\cdot b\cdot c \vDash e\) nor \(h_i\cdot b\cdot c \vDash{\nsim}e\), but may instead only have \(P[e \pmid h_i\cdot b\cdot c] = r\), where r is some “entailment strength” between 0 and 1.
Before going on to describing the logic of evidential support in more detail, perhaps a few more words are in order about the background knowledge and auxiliary hypotheses, represented here by ‘\(b\)’. Duhem (1906) and Quine (1953) are generally credited with alerting inductive logicians to the importance of auxiliary hypotheses in connecting scientific hypotheses and theories to empirical evidence. (See the entry on Pierre Duhem .) They point out that scientific hypotheses often make little contact with evidence claims on their own. Rather, in most cases scientific hypotheses make testable predictions only relative to background information and auxiliary hypotheses that tie them to the evidence. (Some specific examples of such auxiliary hypotheses will be provided in the next subsection.) Typically auxiliaries are highly confirmed hypotheses from other scientific domains. They often describe the operating characteristics of various devices (e.g., measuring instruments) used to make observations or conduct experiments. Their credibility is usually not at issue in the testing of hypothesis \(h_i\) against its competitors, because \(h_i\) and its alternatives usually rely on the same auxiliary hypotheses to tie them to the evidence. But even when an auxiliary hypothesis is already well-confirmed, we cannot simply assume that it is unproblematic, or just known to be true . Rather, the evidential support or refutation of a hypothesis \(h_i\) is relative to whatever auxiliaries and background information (in \(b\)) is being supposed in the confirmational context. In other contexts the auxiliary hypotheses used to test \(h_i\) may themselves be among a collection of alternative hypotheses that are subject to evidential support or refutation. Furthermore, to the extent that competing hypotheses employ different auxiliary hypotheses in accounting for evidence, the evidence only tests each such hypothesis in conjunction with its distinct auxiliaries against alternative hypotheses packaged with their distinct auxiliaries, as described earlier. Thus, what counts as a hypothesis to be tested , \(h_i\), and what counts as auxiliary hypotheses and background information, \(b\), may depend on the epistemic context—on what class of alternative hypotheses are being tested by a collection of experiments or observations, and on what claims are presupposed in that context. No statement is intrinsically a test hypothesis , or intrinsically an auxiliary hypothesis or background condition . Rather, these categories are roles statements may play in a particular epistemic context.
In a probabilistic inductive logic the degree to which the evidence \((c\cdot e)\) supports a hypothesis \(h_i\) relative to background and auxiliaries \(b\) is represented by the posterior probability of \(h_i\), \(P_{\alpha}[h_i \pmid b\cdot c\cdot e]\), according to an evidential support function \(P_{\alpha}\). It turns out that the posterior probability of a hypothesis depends on just two kinds of factors: (1) its prior probability , \(P_{\alpha}[h_i \pmid b]\), together with the prior probabilities of its competitors, \(P_{\alpha}[h_j \pmid b]\), \(P_{\alpha}[h_k \pmid b]\), etc.; and (2) the likelihood of evidential outcomes \(e\) according to \(h_i\) in conjunction with with \(b\) and \(c\), \(P[e \pmid h_i\cdot b\cdot c]\), together with the likelihoods of these same evidential outcomes according to competing hypotheses, \(P[e \pmid h_j\cdot b\cdot c]\), \(P[e \pmid h_k\cdot b\cdot c]\), etc. We will now examine each of these factors in some detail. Following that we will see precisely how the values of posterior probabilities depend on the values of likelihoods and prior probabilities.
In probabilistic inductive logic the likelihoods carry the empirical import of hypotheses. A likelihood is a support function probability of form \(P[e \pmid h_i\cdot b\cdot c]\). It expresses how likely it is that outcome \(e\) will occur according to hypothesis \(h_i\) together with the background and auxiliaries \(b\) and the experimental (or observational) conditions \(c\). [ 5 ] If a hypothesis together with auxiliaries and experimental/observation conditions deductively entails an evidence claim, the axioms of probability make the corresponding likelihood objective in the sense that every support function must agree on its values: \(P[e \pmid h_i\cdot b\cdot c] = 1\) if \(h_i\cdot b\cdot c \vDash e\); \(P[e \pmid h_i\cdot b\cdot c] = 0\) if \(h_i\cdot b\cdot c \vDash{\nsim}e\). However, in many cases a hypothesis \(h_i\) will not be deductively related to the evidence, but will only imply it probabilistically. There are several ways this might happen: (1) hypothesis \(h_i\) may itself be an explicitly probabilistic or statistical hypothesis; (2) an auxiliary statistical hypothesis, as part of the background b , may connect hypothesis \(h_i\) to the evidence; (3) the connection between the hypothesis and the evidence may be somewhat loose or imprecise, not mediated by explicit statistical claims, but nevertheless objective enough for the purposes of evidential evaluation. Let’s briefly consider examples of the first two kinds. We’ll treat case (3) in Section 5 , which addresses the the issue of vague and imprecise likelihoods.
The hypotheses being tested may themselves be statistical in nature. One of the simplest examples of statistical hypotheses and their role in likelihoods are hypotheses about the chance characteristic of coin-tossing. Let \(h_{[r]}\) be a hypothesis that says a specific coin has a propensity (or objective chance ) r for coming up heads on normal tosses, let \(b\) say that such tosses are probabilistically independent of one another. Let \(c\) state that the coin is tossed n times in the normal way; and let \(e\) say that on these tosses the coin comes up heads m times. In cases like this the value of the likelihood of the outcome \(e\) on hypothesis \(h_{[r]}\) for condition \(c\) is given by the well-known binomial formula:
There are, of course, more complex cases of likelihoods involving statistical hypotheses. Consider, for example, the hypothesis that plutonium 233 nuclei have a half-life of 20 minutes—i.e., that the propensity (or objective chance ) for a Pu-233 nucleus to decay within a 20 minute period is 1/2. The full statistical model for the lifetime of such a system says that the propensity (or objective chance ) for that system to remain intact (i.e., to not decay) within any time period x is governed by the formula \(1/2^{x/\tau}\), where \(\tau\) is the half-life of such a system. Let \(h\) be a hypothesis that says that this statistical model applies to Pu-233 nuclei with \(\tau = 20\) minutes; let \(c\) say that some specific Pu-233 nucleus is intact within a decay detector (of some specific kind) at an initial time \(t_0\); let \(e\) say that no decay of this same Pu-233 nucleus is detected by the later time \(t\); and let \(b\) say that the detector is completely accurate (it always registers a real decay, and it never registers false-positive detections). Then, the associated likelihood of \(e\) given \(h\) and \(c\) is this: \(P[e \pmid h\cdot b\cdot c] = 1/2^{(t - t_0)/\tau}\), where the value of \(\tau\) is 20 minutes.
An auxiliary statistical hypothesis, as part of the background \(b\), may be required to connect hypothesis \(h_i\) to the evidence. For example, a blood test for HIV has a known false-positive rate and a known true-positive rate. Suppose the false-positive rate is .05—i.e., the test tends to incorrectly show the blood sample to be positive for HIV in 5% of all cases where HIV is not present . And suppose that the true-positive rate is .99—i.e., the test tends to correctly show the blood sample to be positive for HIV in 99% of all cases where HIV really is present . When a particular patient’s blood is tested, the hypotheses under consideration are this patient is infected with HIV , \(h\), and this patient is not infected with HIV , \({\nsim}h\). In this context the known test characteristics function as background information, b . The experimental condition \(c\) merely states that this particular patient was subjected to this specific kind of blood test for HIV, which was processed by the lab using proper procedures. Let us suppose that the outcome \(e\) states that the result is a positive test result for HIV. The relevant likelihoods then, are \(P[e \pmid h\cdot b\cdot c] = .99\) and \(P[e \pmid {\nsim}h\cdot b\cdot c]\) = .05.
In this example the values of the likelihoods are entirely due to the statistical characteristics of the accuracy of the test, which is carried by the background/auxiliary information \(b\). The hypothesis \(h\) being tested by the evidence is not itself statistical.
This kind of situation may, of course, arise for much more complex hypotheses. The alternative hypotheses of interest may be deterministic physical theories, say Newtonian Gravitation Theory and some specific alternatives. Some of the experiments that test this theory relay on somewhat imprecise measurements that have known statistical error characteristics, which are expressed as part of the background or auxiliary hypotheses, \(b\). For example, the auxiliary \(b\) may describe the error characteristics of a device that measures the torque imparted to a quartz fiber, where the measured torque is used to assess the strength of the gravitational force between test masses. In that case \(b\) may say that for this kind of device the measurement errors are normally distributed about whatever value a given gravitational theory predicts, with some specified standard deviation that is characteristic of the device. This results in specific values \(r_i\) for the likelihoods, \(P[e \pmid h_i\cdot b\cdot c] = r_i\), for each of the various gravitational theories, \(h_i\), being tested.
Likelihoods that arise from explicit statistical claims—either within the hypotheses being tested, or from explicit statistical background claims that tie the hypotheses to the evidence—are often called direct inference likelihoods . Such likelihoods should be completely objective. So, all evidential support functions should agree on their values, just as all support functions agree on likelihoods when evidence is logically entailed. Direct inference likelihoods are logical in an extended, non-deductive sense. Indeed, some logicians have attempted to spell out the logic of direct inferences in terms of the logical form of the sentences involved. [ 6 ] But regardless of whether that project succeeds, it seems reasonable to take likelihoods of this sort to have highly objective or intersubjectively agreed values.
Not all likelihoods of interest in confirmational contexts are warranted deductively or by explicitly stated statistical claims. In such cases the likelihoods may have vague, imprecise values, but values that are determinate enough to still underwrite an objective evaluation of hypotheses on the evidence. In Section 5 we’ll consider such cases, where no underlying statistical theory is involved, but where likelihoods are determinate enough to play their standard role in the evidential evaluation of scientific hypotheses. However, the proper treatment of such cases will be more easily understood after we have first seen how the logic works when likelihoods are precisely known (such as cases where the likelihood values are endorsed by explicit statistical hypotheses and/or explicit statistical auxiliaries). In any case, the likelihoods that relate hypotheses to evidence claims in many scientific contexts will have such objective values. So, although a variety of different support functions \(P_{\alpha}\), \(P_{\beta}\),…, \(P_{\gamma}\), etc., may be needed to represent the differing “inductive proclivities” of the various members of a scientific community, for now we will consider cases where all evidential support functions agree on the values of the likelihoods. For, the likelihoods represent the empirical content of a scientific hypothesis, what the hypothesis (together with experimental conditions, \(c\), and background and auxiliaries \(b\)) says or probabilistically implies about the evidence. Thus, the empirical objectivity of a science relies on a high degree of objectivity or intersubjective agreement among scientists on the numerical values of likelihoods.
To see the point more vividly, imagine what a science would be like if scientists disagreed widely about the values of likelihoods. Each practitioner interprets a theory to say quite different things about how likely it is that various possible evidence statements will turn out to be true. Whereas scientist \(\alpha\) takes theory \(h_1\) to probabilistically imply that event \(e\) is highly likely, his colleague \(\beta\) understands the empirical import of \(h_1\) to say that \(e\) is very unlikely. And, conversely, \(\alpha\) takes competing theory \(h_2\) to probabilistically imply that \(e\) is very unlikely, whereas \(\beta\) reads \(h_2\) to say that \(e\) is extremely likely. So, for \(\alpha\) the evidential outcome \(e\) supplies strong support for \(h_1\) over \(h_2\), because
But his colleague \(\beta\) takes outcome \(e\) to show just the opposite, that \(h_2\) is strongly supported over \(h_1\), because
If this kind of situation were to occur often, or for significant evidence claims in a scientific domain, it would make a shambles of the empirical objectivity of that science. It would completely undermine the empirical testability of such hypotheses and theories within that scientific domain. Under these circumstances, although each scientist employs the same sentences to express a given theory \(h_i\), each understands the empirical import of these sentences so differently that \(h_i\) as understood by \(\alpha\) is an empirically different theory than \(h_i\) as understood by \(\beta\). (Indeed, arguably, \(\alpha\) must take at least one of the two sentences, \(h_1\) or \(h_2\), to express a different proposition than does \(\beta\).) Thus, the empirical objectivity of the sciences requires that experts should be in close agreement about the values of the likelihoods. [ 7 ]
For now we will suppose that the likelihoods have objective or intersubjectively agreed values, common to all agents in a scientific community. We mark this agreement by dropping the subscript ‘\(\alpha\)’, ‘\(\beta\)’, etc., from expressions that represent likelihoods, since all support functions under consideration are supposed to agree on the values for likelihoods. One might worry that this supposition is overly strong. There are legitimate scientific contexts where, although scientists should have enough of a common understanding of the empirical import of hypotheses to assign quite similar values to likelihoods, precise agreement on their numerical values may be unrealistic. This point is right in some important kinds of cases. So later, in Section 5, we will see how to relax the supposition that precise likelihood values are available, and see how the logic works in such cases. But for now the main ideas underlying probabilistic inductive logic will be more easily explained if we focus on those contexts were objective or intersubjectively agreed likelihoods are available. Later we will see that much the same logic continues to apply in contexts where the values of likelihoods may be somewhat vague, or where members of the scientific community disagree to some extent about their values.
An adequate treatment of the likelihoods calls for the introduction of one additional notational device. Scientific hypotheses are generally tested by a sequence of experiments or observations conducted over a period of time. To explicitly represent the accumulation of evidence, let the series of sentences \(c_1\), \(c_2\), …, \(c_n\), describe the conditions under which a sequence of experiments or observations are conducted. And let the corresponding outcomes of these observations be represented by sentences \(e_1\), \(e_2\), …, \(e_n\). We will abbreviate the conjunction of the first n descriptions of experimental or observational conditions by ‘\(c^n\)’, and abbreviate the conjunction of descriptions of their outcomes by ‘\(e^n\)’. Then, for a stream of n observations or experiments and their outcomes, the likelihoods take form \(P[e^n \pmid h_{i}\cdot b\cdot c^{n}] = r\), for appropriate values of \(r\). In many cases the likelihood of the evidence stream will be equal to the product of the likelihoods of the individual outcomes:
When this equality holds, the individual bits of evidence are said to be probabilistically independent on the hypothesis (together with auxiliaries) . In the following account of the logic of evidential support, such probabilistic independence will not be assumed, except in those places where it is explicitly invoked.
The probabilistic logic of evidential support represents the net support of a hypothesis by the posterior probability of the hypothesis , \(P_{\alpha}[h_i \pmid b\cdot c^{n}\cdot e^{n}]\). The posterior probability represents the net support for the hypothesis that results from the evidence, \(c^n \cdot e^n\), together with whatever plausibility considerations are taken to be relevant to the assessment of \(h_i\). Whereas the likelihoods are the means through which evidence contributes to the posterior probability of a hypothesis, all other relevant plausibility consideration are represented by a separate factor, called the prior probability of the hypothesis : \(P_{\alpha}[h_i \pmid b]\). The prior probability represents the weight of any important considerations not captured by the evidential likelihoods. Any relevant considerations that go beyond the evidence itself may be explicitly stated within expression \(b\) (in addition to whatever auxiliary hypotheses \(b\) may contain in support of the likelihoods). Thus, the prior probability of \(h_i\) may depend explicitly on the content of \(b\). It turns out that posterior probabilities depend only on the values of evidential likelihoods together with the values of prior probabilities.
As an illustration of the role of prior probabilities , consider the HIV test example described in the previous section. What the physician and the patient want to know is the value of the posterior probability, \(P_{\alpha}[h \pmid b\cdot c\cdot e]\), that the patient has HIV, \(h\), given the evidence of the positive test, \(c\cdot e\), and given the error rates of the test, described within \(b\). The value of this posterior probability depends on the likelihood (due to the error rates) of this patient obtaining a true-positive result, \(P[e \pmid h\cdot b\cdot c] = .99\), and of obtaining a false-positive result, \(P[e \pmid {\nsim}h\cdot b\cdot c] = .05\). In addition, the value of the of the posterior probability depends on how plausible it is that the patient has HIV prior to taking the test results into account, \(P_{\alpha}[h \pmid b]\). In the context of medical diagnosis, this prior probability is usually assessed on the basis of the base rate for HIV in the patient’s risk group (i.e., whether the patient is an IV drug user, has unprotected sex with multiple partners, etc.). On a rigorous approach to the logic, such information and its risk-relevance should be explicitly stated within the background information \(b\). To see the importance of this information, consider the following numerical results (which may be calculated using the formula called Bayes’ Theorem, presented in the next section). If the base rate for the patient’s risk group is relatively high, say \(P_{\alpha}[h \pmid b] = .10\), then the positive test result yields a posterior probability value for his having HIV of \(P_{\alpha}[h \pmid b\cdot c\cdot e] = .69\). However, if the patient is in a very low risk group, say \(P_{\alpha}[h \pmid b] = .001\), then a positive test result only raises the posterior probability of his having an HIV infection to \(P_{\alpha}[h \pmid b\cdot c\cdot e] = .02\). This posterior probability is much higher than the prior probability of .001, but should not worry the patient too much. This positive test result may well be due to the comparatively high false-positive rate for the test, rather than to the presence of HIV. This sort of test, with a false-positive rate as large as .05, is best used as a screening test; a positive result warrants conducting a second, more rigorous, less error-prone test.
More generally, in the evidential evaluation of scientific hypotheses and theories, prior probabilities represent assessments of non-evidential plausibility weightings among hypotheses. However, because the strengths of such plausibility assessments may vary among members of a scientific community, critics often brand such assessments as merely subjective , and take their role in Bayesian inference to be highly problematic. Bayesian inductivists counter that plausibility assessments play an important, legitimate role in the sciences, especially when evidence cannot suffice to distinguish among some alternative hypotheses. And, they argue, the epithet “merely subjective” is unwarranted. Such plausibility assessments are often backed by extensive arguments that may draw on forceful conceptual considerations.
Scientists often bring plausibility arguments to bear in assessing competing views. Although such arguments are seldom decisive, they may bring the scientific community into widely shared agreement, especially with regard to the implausibility of some logically possible alternatives. This seems to be the primary epistemic role of thought experiments. Consider, for example, the kinds of plausibility arguments that have been brought to bear on the various interpretations of quantum theory (e.g., those related to the measurement problem). These arguments go to the heart of conceptual issues that were central to the original development of the theory. Many of these issues were first raised by those scientists who made the greatest contributions to the development of quantum theory, in their attempts to get a conceptual hold on the theory and its implications.
Given any body of evidence, it is fairly easy to cook up a host of logically possible alternative hypotheses that make the evidence as probable as desired. In particular, it is easy to cook up hypotheses that logically entail any given body evidence, providing likelihood values equal to 1 for all the available evidence. Although most of these cooked up hypotheses will be laughably implausible, evidential likelihoods cannot rule them out. But, the only factors other than likelihoods that figure into the values of posterior probabilities for hypotheses are the values of their prior probabilities; so only prior probability assessments provide a place for the Bayesian logic to bring important plausibility considerations to bear. Thus, the Bayesian logic can only give implausible hypotheses their due via prior probability assessments.
It turns out that the mathematical structure of Bayesian inference makes prior probabilities especially well-suited to represent plausibility assessments among competing hypotheses. For, in the fully fleshed out account of evidential support for hypotheses (spelled out below), it will turn out that only ratios of prior probabilities for competing hypotheses, \(P_{\alpha}[h_j \pmid b] / P_{\alpha}[h_i \pmid b]\), together with ratios of likelihoods, \(P_{\alpha}[e \pmid h_j\cdot b\cdot c] / P_{\alpha}[e \pmid h_2\cdot b\cdot c]\), play essential roles. The ratio of prior probabilities is well-suited to represent how much more (or less) plausible hypothesis \(h_j\) is than competing hypothesis \(h_i\). Furthermore, the plausibility arguments on which such this comparative assessment is based may be explicitly stated within \(b\). So, given that an inductive logic needs to incorporate well-considered plausibility assessments (e.g. in order to lay low wildly implausible alternative hypotheses), the comparative assessment of Bayesian prior probabilities seems well-suited to do the job.
Thus, although prior probabilities may be subjective in the sense that agents may disagree on the relative strengths of plausibility arguments, the priors used in scientific contexts need not represent mere subjective whims . Rather, the comparative strengths of the priors for hypotheses should be supported by arguments about how much more plausible one hypothesis is than another. The important role of plausibility assessments is captured by such received bits of scientific wisdom as the well-known scientific aphorism, extraordinary claims require extraordinary evidence . That is, it takes especially strong evidence, in the form of extremely high values for (ratios of) likelihoods, to overcome the extremely low pre-evidential plausibility values possessed by some hypotheses. In the next section we’ll see precisely how this idea works, and we’ll return to it again in Section 3.4 .
When sufficiently strong evidence becomes available, it turns out that the contributions of prior plausibility assessments to the values of posterior probabilities may be substantially “washed out”, overridden by the evidence. That is, provided the prior probability of a true hypothesis isn’t assessed to be too close to zero, the influence of the values of the prior probabilities will very probably fade away as evidence accumulates. In Section 4 we’ll see precisely how this kind of Bayesian convergence to the true hypothesis works. Thus, it turns out that prior plausibility assessments play their most important role when the distinguishing evidence represented by the likelihoods remains weak.
One more point before moving on to the logic of Bayes’ Theorem. Some Bayesian logicists have maintained that posterior probabilities of hypotheses should be determined by syntactic logical form alone. The idea is that the likelihoods might reasonably be specified in terms of syntactic logical form; so if syntactic form might be made to determine the values of prior probabilities as well, then inductive logic would be fully “formal” in the same way that deductive logic is “formal”. Keynes and Carnap tried to implement this idea through syntactic versions of the principle of indifference—the idea that syntactically similar hypotheses should be assigned the same prior probability values. Carnap showed how to carry out this project in detail, but only for extremely simple formal languages. Most logicians now take the project to have failed because of a fatal flaw with the whole idea that reasonable prior probabilities can be made to depend on logical form alone. Semantic content should matter. Goodmanian grue-predicates provide one way to illustrate this point. [ 8 ] Furthermore, as suggested earlier, for this idea to apply to the evidential support of real scientific theories, scientists would have to assess the prior probabilities of each alternative theory based only on its syntactic structure. That seems an unreasonable way to proceed. Are we to evaluate the prior probabilities of alternative theories of gravitation, or for alternative quantum theories, by exploring only their syntactic structures, with absolutely no regard for their content—with no regard for what they say about the world? This seems an extremely dubious approach to the evaluation of real scientific theories. Logical structure alone cannot, and should not suffice for determining reasonable prior probability values for real scientific theories. Moreover, real scientific hypotheses and theories are inevitably subject to plausibility considerations based on what they say about the world. Prior probabilities are well-suited to represent the comparative weight of plausibility considerations for alternative hypotheses. But no reasonable assessment of comparative plausibility can derive solely from the logical form of hypotheses.
We will return to a discussion of prior probabilities a bit later. Let’s now see how Bayesian logic combines likelihoods with prior probabilities to yield posterior probabilities for hypotheses.
Any probabilistic inductive logic that draws on the usual rules of probability theory to represent how evidence supports hypotheses must be a Bayesian inductive logic in the broad sense. For, Bayes’ Theorem follows directly from the usual axioms of probability theory. Its importance derives from the relationship it expresses between hypotheses and evidence. It shows how evidence, via the likelihoods, combines with prior probabilities to produce posterior probabilities for hypotheses. We now examine several forms of Bayes’ Theorem, each derivable from axioms 1–5 .
The simplest version of Bayes’ Theorem as it applies to evidence for a hypothesis goes like this:
Bayes’ Theorem: Simple Form
This equation expresses the posterior probability of hypothesis \(h_i\) due to evidence \(e\), \(P_{\alpha}[h_i \pmid e]\), in terms of the likelihood of the evidence on that hypothesis, \(P_{\alpha}[e \pmid h_i]\), the prior probability of the hypothesis , \(P_{\alpha}[h_i]\), and the simple probability of the evidence , \(P_{\alpha}[e]\). The factor \(P_{\alpha}[e]\) is often called the expectedness of the evidence . Written this way, the theorem suppresses the experimental (or observational) conditions, \(c\), and all background information and auxiliary hypotheses, \(b\). As discussed earlier, both of these terms play an important role in logically connecting the hypothesis at issue, \(h_i\), to the evidence \(e\). In scientific contexts the objectivity of the likelihoods, \(P_{\alpha}[e \pmid h_i\cdot b \cdot c]\), almost always depends on such terms. So, although the suppression of experimental (or observational) conditions and auxiliary hypotheses is a common practice in accounts of Bayesian inference, the treatment below, and throughout the remainder of this article will make the role of these terms explicit.
The subscript \(\alpha\) on the evidential support function \(P_{\alpha}\) is there to remind us that more than one such function exists. A host of distinct probability functions satisfy axioms 1–5 , so each of them satisfies Bayes’ Theorem. Some of these probability functions may provide a better fit with our intuitive conception of how the evidential support for hypotheses should work. Nevertheless, there are bound to be reasonable differences among Bayesian agents regarding to the initial plausibility of a hypothesis \(h_i\). This diversity in initial plausibility assessments is represented by diverse values for prior probabilities for the hypothesis: \(P_{\alpha}[h_i]\), \(P_{\beta}[h_i]\), \(P_{\gamma}[h_i]\), etc. This usually results in diverse values for posterior probabilities for hypotheses: \(P_{\alpha}[h_i \pmid e]\), \(P_{\beta}[h_i \pmid e]\), \(P_{\gamma}[h_i \pmid e]\), etc. So it is important to keep the diversity among evidential support functions in mind.
Here is how the Simple Form of Bayes’ Theorem looks when terms for the experimental (or observational) conditions, \(c\), and the background information and auxiliary hypotheses \(b\) are made explicit:
Bayes’ Theorem: Simple Form with explicit Experimental Conditions, Background Information and Auxiliary Hypotheses
This version of the theorem determines the posterior probability of the hypothesis, \(P_{\alpha}[h_i \pmid b\cdot c\cdot e]\), from the value of the likelihood of the evidence according to that hypothesis (taken together with background and auxiliaries and the experimental conditions), \(P[e \pmid h_i\cdot b\cdot c]\), the value of the prior probability of the hypothesis (on background and auxiliaries), \(P_{\alpha}[h_i \pmid b]\), and the value of the expectedness of the evidence (on background and auxiliaries and the experimental conditions), \(P_{\alpha}[e \pmid b\cdot c]\). Notice that in the factor for the likelihood, \(P[e \pmid h_i\cdot b\cdot c]\), the subscript \(\alpha\) has been dropped. This marks the fact that in scientific contexts the likelihood of an evidential outcome \(e\) on the hypothesis together with explicit background and auxiliary hypotheses and the description of the experimental conditions, \(h_i\cdot b\cdot c\), is usually objectively determinate. This factor represents what the hypothesis (in conjunction with background and auxiliaries) objectively says about the likelihood of possible evidential outcomes of the experimental conditions. So, all reasonable support functions should agree on the values for likelihoods. (Section 5 will treat cases where the likelihoods may lack this kind of objectivity.)
This version of Bayes’ Theorem includes a term that represents the ratio of the likelihood of the experimental conditions on the hypothesis and background information (and auxiliaries) to the “likelihood” of the experimental conditions on the background (and auxiliaries) alone: \(P_{\alpha}[c \pmid h_i\cdot b]/ P_{\alpha}[c \pmid b]\). Arguably the value of this term should be 1, or very nearly 1, since the truth of the hypothesis at issue should not significantly affect how likely it is that the experimental conditions are satisfied. If various alternative hypotheses assign significantly different likelihoods to the experimental conditions themselves, then such conditions should more properly be included as part of the evidential outcome \(e\).
Both the prior probability of the hypothesis and the expectedness tend to be somewhat subjective factors in that various agents from the same scientific community may legitimately disagree on what values these factors should take. Bayesian logicians usually accept the apparent subjectivity of the prior probabilities of hypotheses, but find the subjectivity of the expectedness to be more troubling. This is due at least in part to the fact that in a Bayesian logic of evidential support the value of the expectedness cannot be determined independently of likelihoods and prior probabilities of hypotheses. That is, when, for each member of a collection of alternative hypotheses, the likelihood \(P[e \pmid h_j\cdot b\cdot c]\) has an objective (or intersubjectively agreed) value, the expectedness is constrained by the following equation (where the sum ranges over a mutually exclusive and exhaustive collection of alternative hypotheses \(\{h_1, h_2 , \ldots ,h_m , \ldots \}\), which may be finite or countably infinite):
This equation shows that the values for the prior probabilities together with the values of the likelihoods uniquely determine the value for the expectedness of the evidence . Furthermore, it implies that the value of the expectedness must lie between the largest and smallest of the various likelihood values implied by the alternative hypotheses. However, the precise value of the expectedness can only be calculated this way when every alternative to hypothesis \(h_j\) is specified. In cases where some alternative hypotheses remain unspecified (or undiscovered), the value of the expectedness is constrained in principle by the totality of possible alternative hypotheses, but there is no way to figure out precisely what its value should be.
Troubles with determining a numerical value for the expectedness of the evidence may be circumvented by appealing to another form of Bayes’ Theorem, a ratio form that compares hypotheses one pair at a time:
Bayes’ Theorem: Ratio Form
The clause \(P_{\alpha}[c \pmid h_j\cdot b] = P_{\alpha}[c \pmid h_i\cdot b]\) says that the experimental (or observation) condition described by \(c\) is as likely on \((h_i\cdot b)\) as on \((h_j\cdot b)\) — i.e., the experimental or observation conditions are no more likely according to one hypothesis than according to the other. [ 9 ]
This Ratio Form of Bayes’ Theorem expresses how much more plausible, on the evidence, one hypothesis is than another. Notice that the likelihood ratios carry the full import of the evidence. The evidence influences the evaluation of hypotheses in no other way. The only other factor that influences the value of the ratio of posterior probabilities is the ratio of the prior probabilities. When the likelihoods are fully objective, any subjectivity that affects the ratio of posteriors can only arise via subjectivity in the ratio of the priors.
This version of Bayes’s Theorem shows that in order to evaluate the posterior probability ratios for pairs of hypotheses, the prior probabilities of hypotheses need not be evaluated absolutely; only their ratios are needed. That is, with regard to the priors, the Bayesian evaluation of hypotheses only relies on how much more plausible one hypothesis is than another (due to considerations expressed within b ). This kind of Bayesian evaluation of hypotheses is essentially comparative in that only ratios of likelihoods and ratios of prior probabilities are ever really needed for the assessment of scientific hypotheses. Furthermore, we will soon see that the absolute values of the posterior probabilities of hypotheses entirely derive from the posterior probability ratios provided by the Ratio Form of Bayes’ Theorem.
When the evidence consists of a collection of n distinct experiments or observations, we may explicitly represent this fact by replacing the term ‘\(c\)’ by the conjunction of experimental or observational conditions, \((c_1\cdot c_2\cdot \ldots \cdot c_n)\), and replacing the term ‘\(e\)’ by the conjunction of their respective outcomes, \((e_1\cdot e_2\cdot \ldots \cdot e_n)\). For notational convenience, let’s use the term ‘\(c^n\)’ to abbreviate the conjunction of n the experimental conditions, and we use the term ‘\(e^n\)’ to abbreviate the corresponding conjunction of n their respective outcomes. Relative to any given hypothesis \(h\), the evidential outcomes of distinct experiments or observations will usually be probabilistically independent of one another, and also independent of the experimental conditions for one another. In that case we have:
When the Ratio Form of Bayes’ Theorem is extended to explicitly represent the evidence as consisting of a collection of n of distinct experiments (or observations) and their respective outcomes, it takes the following form.
Bayes’ Theorem: Ratio Form for a Collection of n Distinct Evidence Claims
Furthermore, when evidence claims are probabilistically independent of one another, we have
Let’s consider a simple example of how the Ratio Form of Bayes’ Theorem applies to a collection of independent evidential events. Suppose we possess a warped coin and want to determine its propensity for heads when tossed in the usual way. Consider two hypotheses, \(h_{[p]}\) and \(h_{[q]}\), which say that the propensities for the coin to come up heads on the usual kinds of tosses are \(p\) and \(q\), respectively. Let \(c^n\) report that the coin is tossed n times in the normal way, and let \(e^n\) report that precisely m occurrences of heads has resulted. Supposing that the outcomes of such tosses are probabilistically independent (asserted by \(b\)), the respective likelihoods take the binomial form
with \(r\) standing in for \(p\) and for \(q\), respectively. Then, Equation 9** yields the following formula, where the likelihood ratio is the ratio of the respective binomial terms:
When, for instance, the coin is tossed \(n = 100\) times and comes up heads \(m = 72\) times, the evidence for hypothesis \(h_{[1/2]}\) as compared to \(h_{[3/4]}\) is given by the likelihood ratio
In that case, even if the prior plausibility considerations (expressed within \(b\)) make it 100 times more plausible that the coin is fair than that it is warped towards heads with propensity 3/4 — i.e., even if \(P_{\alpha}[h_{[1/2]} \pmid b] / P_{\alpha}[h_{[3/4]} \pmid b] = 100\) — the evidence provided by these tosses makes the posterior plausibility that the coin is fair only about 6/1000 ths as plausible as the hypothesis that it is warped towards heads with propensity 3/4 :
Thus, such evidence strongly refutes the “fairness hypothesis” relative to the “3/4- heads hypothesis”, provided the assessment of prior prior plausibilities doesn’t make the latter hypothesis too extremely implausible to begin with. Notice, however, that strong refutation is not absolute refutation . Additional evidence could reverse this trend towards the refutation of the fairness hypothesis .
This example employs repetitions of the same kind of experiment—repeated tosses of a coin. But the point holds more generally. If, as the evidence increases, the likelihood ratios
approach 0, then the Ratio Forms of Bayes’ Theorem, Equations \(9*)\) and \(9**)\), show that the posterior probability of \(h_j\) must approach 0 as well, since
Such evidence comes to strongly refute \(h_j\), with little regard for its prior plausibility value. Indeed, Bayesian induction turns out to be a version of eliminative induction , and Equation \(9*\) and \(9**\) begin to illustrate this. For, suppose that \(h_i\) is the true hypothesis, and consider what happens to each of its false competitors, \(h_j\). If enough evidence becomes available to drive each of the likelihood ratios
toward 0 (as n increases), then Equation \(9*\) says that each false \(h_j\) will become effectively refuted — each of their posterior probabilities will approaches 0 (as n increases). As a result, the posterior probability of \(h_i\) must approach 1. The next two equations show precisely how this works.
If we sum the ratio versions of Bayes’ Theorem in Equation \(9*\) over all alternatives to hypothesis \(h_i\) (including the catch-all alternative \(h_K\), if appropriate), we get the Odds Form of Bayes’ Theorem. By definition, the odds against a statement \(A\) given \(B\) is related to the probability of \(A\) given \(B\) as follows:
This notion of odds gives rise to the following version of Bayes’ Theorem:
Bayes’ Theorem: Odds Form
where the factor following the ‘ + ’ sign is only required in cases where a catch-all alternative hypothesis, \(h_K\), is needed.
Recall that when we have a finite collection of concrete alternative hypotheses available, \(\{h_1, h_2 , \ldots ,h_m\}\), but where this set of alternatives is not exhaustive (where additional, unarticulated, undiscovered alternative hypotheses may exist), the catch-all alternative hypothesis \(h_K\) is just the denial of each of the concrete alternatives, \(({\nsim}h_1\cdot{\nsim}h_2\cdot \ldots \cdot{\nsim}h_m)\). Generally, the likelihood of evidence claims relative to a catch-all hypothesis will not enjoy the same kind of objectivity possessed by the likelihoods for concrete alternative hypotheses. So, we leave the subscript \(\alpha\) attached to the likelihood for the catch-all hypothesis to indicate this lack of objectivity.
Although the catch-all hypothesis may lack objective likelihoods, the influence of the catch-all term in Bayes’ Theorem diminishes as additional concrete hypotheses are articulated. That is, as new hypotheses are discovered they are “peeled off” of the catch-all. So, when a new hypothesis \(h_{m+1}\) is formulated and made explicit, the old catch-all hypothesis \(h_K\) is replaced by a new catch-all, \(h_{K*}\), of form \(({\nsim}h_1\cdot \cdot{\nsim}h_2\cdot \ldots \cdot{\nsim}h_{m}\cdot{\nsim}h_{m+1})\); and the prior probability for the new catch-all hypothesis is gotten by diminishing the prior of the old catch-all: \(P_{\alpha}[h_{K*} \pmid b] = P_{\alpha}[h_K \pmid b] - P_{\alpha}[h_{m+1} \pmid b]\). Thus, the influence of the catch-all term should diminish towards 0 as new alternative hypotheses are made explicit. [ 10 ]
If increasing evidence drives towards 0 the likelihood ratios comparing each competitor \(h_j\) with hypothesis \(h_i\), then the odds against \(h_i\), \(\Omega_{\alpha}[{\nsim}h_i \pmid b\cdot c^{n}\cdot e^{n}]\), will approach 0 (provided that priors of catch-all terms, if needed, approach 0 as well, as new alternative hypotheses are made explicit and peeled off). And, as \(\Omega_{\alpha}[{\nsim}h_i \pmid b\cdot c^{n}\cdot e^{n}]\) approaches 0, the posterior probability of \(h_i\) goes to 1. This derives from the fact that the odds against \(h_i\) is related to and its posterior probability by the following formula:
Bayes’ Theorem: General Probabilistic Form
The odds against a hypothesis depends only on the values of ratios of posterior probabilities , which entirely derive from the Ratio Form of Bayes’ Theorem. Thus, we see that the individual value of the posterior probability of a hypothesis depends only on the ratios of posterior probabilities , which come from the Ratio Form of Bayes’ Theorem. Thus, the Ratio Form of Bayes’ Theorem captures all the essential features of the Bayesian evaluation of hypothesis. It shows how the impact of evidence (in the form of likelihood ratios) combines with comparative plausibility assessments of hypotheses (in the form of ratios of prior probabilities) to provide a net assessment of the extent to which hypotheses are refuted or supported via contests with their rivals.
There is a result, a kind of Bayesian Convergence Theorem , that shows that if \(h_i\) (together with \(b\cdot c^n)\) is true, then the likelihood ratios
comparing evidentially distinguishable alternative hypothesis \(h_j\) to \(h_i\) will very probably approach 0 as evidence accumulates (i.e., as n increases). Let’s call this result the Likelihood Ratio Convergence Theorem . When this theorem applies, Equation \(9^*\) shows that the posterior probability of a false competitor \(h_j\) will very probably approach 0 as evidence accumulates, regardless of the value of its prior probability \(P_{\alpha}[h_j \pmid b]\). As this happens to each of \(h_i\)’s false competitors, Equations 10 and 11 say that the posterior probability of the true hypothesis, \(h_i\), will approach 1 as evidence increases. [ 11 ] Thus, Bayesian induction is at bottom a version of induction by elimination , where the elimination of alternatives comes by way of likelihood ratios approaching 0 as evidence accumulates. Thus, when the Likelihood Ratio Convergence Theorem applies, the Criterion of Adequacy for an Inductive Logic described at the beginning of this article will be satisfied: As evidence accumulates, the degree to which the collection of true evidence statements comes to support a hypothesis, as measured by the logic, should very probably come to indicate that false hypotheses are probably false and that true hypotheses are probably true. We will examine this Likelihood Ratio Convergence Theorem in Section 4 . [ 12 ]
A view called Likelihoodism relies on likelihood ratios in much the same way as the Bayesian logic articulated above. However, Likelihoodism attempts to avoid the use of prior probabilities. For an account of this alternative view, see the supplement Likelihood Ratios, Likelihoodism, and the Law of Likelihood . For more discussion of Bayes’ Theorem and its application, see the entries on Bayes’ Theorem and on Bayesian Epistemology in this Encyclopedia .
Given that a scientific community should largely agree on the values of the likelihoods, any significant disagreement among them with regard to the values of posterior probabilities of hypotheses should derive from disagreements over their assessments of values for the prior probabilities of those hypotheses. We saw in Section 3.3 that the Bayesian logic of evidential support need only rely on assessments of ratios of prior probabilities —on how much more plausible one hypothesis is than another. Thus, the logic of evidential support only requires that scientists can assess the comparative plausibilities of various hypotheses. Presumably, in scientific contexts the comparative plausibility values for hypotheses should depend on explicit plausibility arguments, not merely on privately held opinions. (Formally, the logic may represent comparative plausibility arguments by explicit statements expressed within \(b\).) It would be highly unscientific for a member of the scientific community to disregard or dismiss a hypothesis that other members take to be a reasonable proposal with only the comment, “don’t ask me to give my reasons, it’s just my opinion”. Even so, agents may be unable to specify precisely how much more strongly the available plausibility arguments support a hypothesis over an alternative; so prior probability ratios for hypotheses may be vague. Furthermore, agents in a scientific community may disagree about how strongly the available plausibility arguments support a hypothesis over a rival hypothesis; so prior probability ratios may be somewhat diverse as well.
Both the vagueness of comparative plausibilities assessments for individual agents and the diversity of such assessments among the community of agents can be represented formally by sets of support functions, \(\{P_{\alpha}, P_{\beta}, \ldots \}\), that agree on the values for the likelihoods but encompass a range of values for the (ratios of) prior probabilities of hypotheses. Vagueness and diversity are somewhat different issues, but they may be represented in much the same way. Let’s briefly consider each in turn.
Assessments of the prior plausibilities of hypotheses will often be vague—not subject to the kind of precise quantitative treatment that a Bayesian version of probabilistic inductive logic may seem to require for prior probabilities. So, it may seem that the kind of assessment of prior probabilities required to get the Bayesian algorithm going cannot be accomplished in practice. To see how Bayesian inductivists address this worry, first recall the Ratio Form of Bayes’ Theorem, Equation \(9^*\).
Recall that this Ratio Form of the theorem captures the essential features of the logic of evidential support, even though it only provides a value for the ratio of the posterior probabilities. Notice that the ratio form of the theorem easily accommodates situations where we don’t have precise numerical values for prior probabilities. It only depends on our ability to assess how much more or less plausible alternative hypothesis \(h_j\) is than hypothesis \(h_i\)—only the value of the ratio \(P_{\alpha}[h_j \pmid b] / P_{\alpha}[h_i \pmid b]\) need be assessed; the values of the individual prior probabilities are not needed. Such comparative plausibilities are much easier to assess than specific numerical values for the prior probabilities of individual hypotheses. When combined with the ratio of likelihoods , this ratio of priors suffices to yield an assessment of the ratio of posterior plausibilities ,
Although such posterior ratios don’t supply values for the posterior probabilities of individual hypotheses, they place a crucial constraint on the posterior support of hypothesis \(h_j\), since
This Ratio Form of Bayes’ Theorem tolerates a good deal of vagueness or imprecision in assessments of the ratios of prior probabilities. In practice one need only assess bounds for these prior plausibility ratios to achieve meaningful results. Given a prior ratio in a specific interval,
a likelihood ratio
results in a posterior support ratio in the interval
(Technically each probabilistic support function assigns a specific numerical value to each pair of sentences; so when we write an inequality like
we are really referring to a set of probability functions \(P_{\alpha}\), a vagueness set , for which the inequality holds. Thus, technically, the Bayesian logic employs sets of probabilistic support functions to represent the vagueness in comparative plausibility values for hypotheses.)
Observe that if the likelihood ratio values \(\LR^n\) approach 0 as the amount of evidence \(e^n\) increases, the interval of values for the posterior probability ratio must become tighter as the upper bound (\(\LR^n\times r)\) approaches 0. Furthermore, the absolute degree of support for \(h_j\), \(P_{\alpha}[h_j \pmid b\cdot c^{n}\cdot e^{n}]\), must also approach 0.
This observation is really useful. For, it can be shown that when \(h_{i}\cdot b\cdot c^{n}\) is true and \(h_j\) is empirically distinct from \(h_i\), the continual pursuit of evidence is very likely to result in evidential outcomes \(e^n\) that (as n increases) yield values of likelihood ratios \(P[e^n \pmid h_{j}\cdot b\cdot c^{n}] / P[e^n \pmid h_{i}\cdot b\cdot c^{n}]\) that approach 0 as the amount of evidence increases. This result, called the Likelihood Ratio Convergence Theorem , will be investigated in more detail in Section 4 . When that kind of convergence towards 0 for likelihood ratios occurs, the upper bound on the posterior probability ratio also approaches 0, driving the posterior probability of \(h_j\) to approach 0 as well, effectively refuting hypothesis \(h_j\). Thus, false competitors of a true hypothesis will effectively be eliminated by increasing evidence. As this happens, Equations 9* through 11 show that the posterior probability \(P_{\alpha}[h_i \pmid b\cdot c^{n}\cdot e^{n}]\) of the true hypothesis \(h_i\) approaches 1.
Thus, Bayesian logic of inductive support for hypotheses is a form of eliminative induction, where the evidence effectively refutes false alternatives to the true hypothesis. Because of its eliminative nature, the Bayesian logic of evidential support doesn’t require precise values for prior probabilities. It only needs to draw on bounds on the values of comparative plausibility ratios, and these bounds only play a significant role while evidence remains fairly weak. If the true hypothesis is assessed to be comparatively plausible (due to plausibility arguments contained in b ), then plausibility assessments give it a leg-up over alternatives. If the true hypothesis is assessed to be comparatively implausible, the plausibility assessments merely slow down the rate at which it comes to dominate its rivals, reflecting the idea that extraordinary hypotheses require extraordinary evidence (or an extraordinary accumulation of evidence) to overcome their initial implausibilities. Thus, as evidence accumulates, the agent’s vague initial plausibility assessments transform into quite sharp posterior probabilities that indicate their strong refutation or support by the evidence.
When the various agents in a community may widely disagree over the non-evidential plausibilities of hypotheses, the Bayesian logic of evidential support may represent this kind of diversity across the community of agents as a collection of the agents’ vagueness sets of support functions. Let’s call such a collection of support functions a diversity set . That is, a diversity set is just a set of support functions \(P_{\alpha}\) that cover the ranges of values for comparative plausibility assessments for pairs of competing hypotheses
as assessed by the scientific community. But, once again, if accumulating evidence drives the likelihood ratios comparing various alternative hypotheses to the true hypothesis towards 0, the range of support functions in a diversity set will come to near agreement, near 0, on the values for posterior probabilities of false competitors of the true hypothesis. So, not only does such evidence firm up each agent’s vague initial plausibility assessment, it also brings the whole community into agreement on the near refutation of empirically distinct competitors of a true hypothesis. As this happens, the posterior probability of the true hypothesis may approach 1. The Likelihood Ratio Convergence Theorem implies that this kind of convergence to the truth should very probably happen, provided that the true hypothesis is empirically distinct enough from its rivals.
One more point about prior probabilities and Bayesian convergence should be mentioned before proceeding to Section 4 . Some subjectivist versions of Bayesian induction seem to suggest that an agent’s prior plausibility assessments for hypotheses should stay fixed once-and-for-all, and that all plausibility updating should be brought about via the likelihoods in accord with Bayes’ Theorem. Critics argue that this is unreasonable. The members of a scientific community may quite legitimately revise their (comparative) prior plausibility assessments for hypotheses from time to time as they rethink plausibility arguments and bring new considerations to bear. This seems a natural part of the conceptual development of a science. It turns out that such reassessments of the comparative plausibilities of hypotheses poses no difficulty for the probabilistic inductive logic discussed here. Such reassessments may be represented by the addition or modification of explicit statements that modify the background information b . Such reassessments may result in (non-Bayesian) transitions to new vagueness sets for individual agents and new diversity sets for the community. The logic of Bayesian induction (as described here) has nothing to say about what values the prior plausibility assessments for hypotheses should have; and it places no restrictions on how they might change over time. Provided that the series of reassessments of (comparative) prior plausibilities doesn’t happen to diminish the (comparative) prior plausibility value of the true hypothesis towards zero (or, at least, doesn’t do so too quickly), the Likelihood Ratio Convergence Theorem implies that the evidence will very probably bring the posterior probabilities of empirically distinct rivals of the true hypothesis to approach 0 via decreasing likelihood ratios; and as this happens, the posterior probability of the true hypothesis will head towards 1.
(Those interested in a Bayesian account of Enumerative Induction and the estimation of values for relative frequencies of attributes in populations should see the supplement, Enumerative Inductions: Bayesian Estimation and Convergence .)
4. The Likelihood Ratio Convergence Theorem
In this section we will investigate the Likelihood Ratio Convergence Theorem . This theorem shows that under certain reasonable conditions, when hypothesis \(h_i\) (in conjunction with auxiliaries in b ) is true and an alternative hypothesis \(h_j\) is empirically distinct from \(h_i\) on some possible outcomes of experiments or observations described by conditions \(c_k\), then it is very likely that a long enough sequence of such experiments and observations c\(^n\) will produce a sequence of outcomes \(e^n\) that yields likelihood ratios \(P[e^n \pmid h_{j}\cdot b\cdot c^{n}] / P[e^n \pmid h_{i}\cdot b\cdot c^{n}]\) that approach 0, favoring \(h_i\) over \(h_j\), as evidence accumulates (i.e., as n increases). This theorem places an explicit lower bound on the “rate of probable convergence” of these likelihood ratios towards 0. That is, it puts a lower bound on how likely it is, if \(h_i\) is true, that a stream of outcomes will occur that yields likelihood ratio values against \(h_j\) as compared to \(h_i\) that lie within any specified small distance above 0.
The theorem itself does not require the full apparatus of Bayesian probability functions. It draws only on likelihoods. Neither the statement of the theorem nor its proof employ prior probabilities of any kind. So even likelihoodists , who eschew the use of Bayesian prior probabilities, may embrace this result. Given the forms of Bayes’ Theorem, 9*-11 from the previous section, the Likelihood Ratio Convergence Theorem further implies the likely convergence to 0 of the posterior probabilities of false competitors of a true hypothesis. That is, when the ratios \(P[e^n \pmid h_{j}\cdot b\cdot c^{n}] / P[e^n \pmid h_{i}\cdot b\cdot c^{n}]\) approach 0 for increasing n , the Ratio Form of Bayes’ Theorem, Equation 9* , says that the posterior probability of \(h_j\) must also approach 0 as evidence accumulates, regardless of the value of its prior probability. So, support functions in collections representing vague prior plausibilities for an individual agent (i.e., a vagueness set) and representing the diverse range of priors for a community of agents (i.e., a diversity set) will come to agree on the near 0 posterior probability of empirically distinct false rivals of a true hypothesis. And as the posterior probabilities of false competitors fall, the posterior probability of the true hypothesis heads towards 1. Thus, the theorem establishes that the inductive logic of probabilistic support functions satisfies the Criterion of Adequacy (CoA) suggested at the beginning of this article.
The Likelihood Ratio Convergence Theorem merely provides some sufficient conditions for probable convergence. But likelihood ratios may well converge towards 0 (in the way described by the theorem) even when the antecedent conditions of the theorem are not satisfied. This theorem overcomes many of the objections raised by critics of Bayesian convergence results. First, this theorem does not employ second-order probabilities ; it says noting about the probability of a probability. It only concerns the probability of a particular disjunctive sentence that expresses a disjunction of various possible sequences of experimental or observational outcomes. The theorem does not require evidence to consist of sequences of events that, according to the hypothesis, are identically distributed (like repeated tosses of a die). The result is most easily expressed in cases where the individual outcomes of a sequence of experiments or observations are probabilistically independent, given each hypothesis. So that is the version that will be presented in this section. However, a version of the theorem also holds when the individual outcomes of the evidence stream are not probabilistically independent, given the hypotheses. (This more general version of the theorem will be presented in a supplement on the Probabilistic Refutation Theorem , below, where the proof of both versions is provided.) In addition, this result does not rely on supposing that the probability functions involved are countably additive . Furthermore, the explicit lower bounds on the rate of convergence provided by this result means that there is no need to wait for the infinitely long run before convergence occurs (as some critics seem to think).
It is sometimes claimed that Bayesian convergence results only work when an agent locks in values for the prior probabilities of hypotheses once-and-for-all, and then updates posterior probabilities from there only by conditioning on evidence via Bayes Theorem. The Likelihood Ratio Convergence Theorem , however, applies even if agents revise their prior probability assessments over time. Such non-Bayesian shifts from one support function (or vagueness set) to another may arise from new plausibility arguments or from reassessments of the strengths of old ones. The Likelihood Ratio Convergence Theorem itself only involves the values of likelihoods. So, provided such reassessments don’t push the prior probability of the true hypothesis towards 0 too rapidly , the theorem implies that the posterior probabilities of each empirically distinct false competitor will very probably approach 0 as evidence increases. [ 13 ]
To specify the details of the Likelihood Ratio Convergence Theorem we’ll need a few additional notational conventions and definitions. Here they are.
For a given sequence of n experiments or observations \(c^n\), consider the set of those possible sequences of outcomes that would result in likelihood ratios for \(h_j\) over \(h_i\) that are less than some chosen small number \(\varepsilon \gt 0\). This set is represented by the expression,
Placing the disjunction symbol ‘\(\vee\)’ in front of this expression yields an expression,
that we’ll use to represent the disjunction of all outcome sequences \(e^n\) in this set. So,
is just a particular sentence that says, in effect, “one of the sequences of outcomes of the first n experiments or observations will occur that makes the likelihood ratio for \(h_j\) over \(h_i\) less than \(\varepsilon\)”.
The Likelihood Ratio Convergence Theorem says that under certain conditions (covered in detail below), the likelihood of a disjunctive sentence of this sort, given that ‘\(h_{i}\cdot b\cdot c^{n}\)’ is true,
must be at least \(1-(\psi /n)\), for some explicitly calculable term \(\psi\). Thus, the true hypothesis \(h_i\) probabilistically implies that as the amount of evidence, n , increases, it becomes highly likely (as close to 1 as you please) that one of the outcome sequences \(e^n\) will occur that yields a likelihood ratio \(P[e^n \pmid h_{j}\cdot b\cdot c^{n}] / P[e^n \pmid h_{i}\cdot b\cdot c^{n}]\) less than \(\varepsilon\); and this holds for any specific value of \(\varepsilon\) you may choose. As this happens, the posterior probability of \(h_i\)’s false competitor, \(h_j\), must approach 0, as required by the Ratio Form of Bayes’ Theorem, Equation 9* .
The term \(\psi\) in the lower bound of this probability depends on a measure of the empirical distinctness of the two hypotheses \(h_j\) and \(h_i\) for the proposed sequence of experiments and observations \(c^n\). To specify this measure we need to contemplate the collection of possible outcomes of each experiment or observation. So, consider some sequence of experimental or observational conditions described by sentences \(c_1,c_2 ,\ldots ,c_n\). Corresponding to each condition \(c_k\) there will be some range of possible alternative outcomes. Let \(O_{k} = \{o_{k1},o_{k2},\ldots ,o_{kw}\}\) be a set of statements describing the alternative possible outcomes for condition \(c_k\). (The number of alternative outcomes will usually differ for distinct experiments among those in the sequence \(c_1 ,\ldots ,c_n\); so, the value of w may depend on \(c_k\).) For each hypothesis \(h_j\), the alternative outcomes of \(c_k\) in \(O_k\) are mutually exclusive and exhaustive, so we have:
We now let expressions of form ‘\(e_k\)’ act as variables that range over the possible outcomes of condition \(c_k\)—i.e., \(e_k\) ranges over the members of \(O_k\). As before, ‘\(c^n\)’ denotes the conjunction of the first n test conditions, \((c_1\cdot c_2\cdot \ldots \cdot c_n)\), and ‘\(e^n\)’ represents possible sequences of corresponding outcomes, \((e_1\cdot e_2\cdot \ldots \cdot e_n)\). Let’s use the expression ‘ E\(^n\) ’ to represent the set of all possible outcome sequences that may result from the sequence of conditions c\(^n\) . So, for each hypothesis \(h_j\) (including \(h_i)\), \(\sum_{e^n\in E^n} P[e^n \pmid h_{j}\cdot b\cdot c^{n}] = 1\).
Everything introduced in this subsection is mere notational convention. No substantive suppositions (other than the axioms of probability theory) have yet been introduced. The version of the Likelihood Ratio Convergence Theorem I’ll present below does, however, draw on one substantive supposition, although a rather weak one. The next subsection will discuss that supposition in detail.
In most scientific contexts the outcomes in a stream of experiments or observations are probabilistically independent of one another relative to each hypothesis under consideration, or can at least be divided up into probabilistically independent parts. For our purposes probabilistic independence of evidential outcomes on a hypothesis divides neatly into two types.
Definition: Independent Evidence Conditions :
- A sequence of outcomes \(e^k\) is condition-independent of a condition for an additional experiment or observation \(c_{k+1}\), given \(h\cdot b\) together with its own conditions \(c^k\), if and only if \[ P[e^k \pmid h\cdot b\cdot c^{k }\cdot c_{ k+1}] = P[e^k \pmid h\cdot b\cdot c^k] . \]
- An individual outcome \(e_k\) is result-independent of a sequence of other observations and their outcomes \((c^{k-1}\cdot e^{k-1})\), given \(h\cdot b\) and its own condition \(c_k\), if and only if \[ P[e_k \pmid h\cdot b\cdot c_k\cdot(c^{k-1 }\cdot e^{ k-1})] = P[e_k \pmid h\cdot b\cdot c_k] . \]
When these two conditions hold, the likelihood for an evidence sequence may be decomposed into the product of the likelihoods for individual experiments or observations. To see how the two independence conditions affect the decomposition, first consider the following formula, which holds even when neither independence condition is satisfied:
When condition-independence holds, the likelihood of the whole evidence stream parses into a product of likelihoods that probabilistically depend on only past observation conditions and their outcomes. They do not depend on the conditions for other experiments whose outcomes are not yet specified. Here is the formula:
Finally, whenever both independence conditions are satisfied we have the following relationship between the likelihood of the evidence stream and the likelihoods of individual experiments or observations:
(For proofs of Equations 12–14 see the supplement Immediate Consequences of Independent Evidence Conditions .)
In scientific contexts the evidence can almost always be divided into parts that satisfy both clauses of the Independent Evidence Condition with respect to each alternative hypothesis. To see why, let us consider each independence condition more carefully.
Condition-independence says that the mere addition of a new observation condition \(c_{k+1}\), without specifying one of its outcomes , does not alter the likelihood of the outcomes \(e^k\) of other experiments \(c^k\). To appreciate the significance of this condition, imagine what it would be like if it were violated. Suppose hypothesis \(h_j\) is some statistical theory, say, for example, a quantum theory of superconductivity. The conditions expressed in \(c^k\) describe a number of experimental setups, perhaps conducted in numerous labs throughout the world, that test a variety of aspects of the theory (e.g., experiments that test electrical conductivity in different materials at a range of temperatures). An outcome sequence \(e^k\) describes the results of these experiments. The violation of condition-independence would mean that merely adding to \(h_{j}\cdot b\cdot c^{k}\) a statement \(c_{k+1}\) describing how an additional experiment has been set up, but with no mention of its outcome, changes how likely the evidence sequence \(e^k\) is taken to be. What \((h_j\cdot b)\) says via likelihoods about the outcomes \(e^k\) of experiments \(c^k\) differs as a result of merely supplying a description of another experimental arrangement, \(c_{k+1}\). Condition-independence , when it holds, rules out such strange effects.
Result-independence says that the description of previous test conditions together with their outcomes is irrelevant to the likelihoods of outcomes for additional experiments. If this condition were widely violated, then in order to specify the most informed likelihoods for a given hypothesis one would need to include information about volumes of past observations and their outcomes. What a hypothesis says about future cases would depend on how past cases have gone. Such dependence had better not happen on a large scale. Otherwise, the hypothesis would be fairly useless, since its empirical import in each specific case would depend on taking into account volumes of past observational and experimental results. However, even if such dependencies occur, provided they are not too pervasive, result-independence can be accommodated rather easily by packaging each collection of result-dependent data together, treating it like a single extended experiment or observation. The result-independence condition will then be satisfied by letting each term ‘\(c_k\)’ in the statement of the independence condition represent a conjunction of test conditions for a collection of result-dependent tests, and by letting each term ‘\(e_k\)’ (and each term ‘\(o_{ku}\)’) stand for a conjunction of the corresponding result-dependent outcomes. Thus, by packaging result-dependent data together in this way, the result-independence condition is satisfied by those (conjunctive) statements that describe the separate, result-independent chunks. [ 14 ]
The version of the Likelihood Ratio Convergence Theorem we will examine depends only on the Independent Evidence Conditions (together with the axioms of probability theory). It draws on no other assumptions. Indeed, an even more general version of the theorem can be established, a version that draws on neither of the Independent Evidence Conditions . However, the Independent Evidence Conditions will be satisfied in almost all scientific contexts, so little will be lost by assuming them. (And the presentation will run more smoothly if we side-step the added complications needed to explain the more general result.)
From this point on, let us assume that the following versions of the Independent Evidence Conditions hold.
Assumption: Independent Evidence Assumptions . For each hypothesis h and background b under consideration, we assume that the experiments and observations can be packaged into condition statements, \(c_1 ,\ldots ,c_k, c_{k+1},\ldots\), and possible outcomes in a way that satisfies the following conditions:
- Each sequence of possible outcomes \(e^k\) of a sequence of conditions \(c^k\) is condition-independent of additional conditions \(c_{k+1}\)—i.e., \[P[e^k \pmid h\cdot b\cdot c^{k}\cdot c_{k+1}] = P[e^k \pmid h\cdot b\cdot c^k].\]
- Each possible outcome \(e_k\) of condition \(c_k\) is result-independent of sequences of other observations and possible outcomes \((c^{k-1}\cdot e^{k-1})\)—i.e., \[P[e_k \pmid h\cdot b\cdot c_k\cdot(c^{k-1}\cdot e^{k-1})] = P[e_k \pmid h\cdot b\cdot c_k].\]
We now have all that is needed to begin to state the Likelihood Ratio Convergence Theorem .
The Likelihood Ratio Convergence Theorem comes in two parts. The first part applies only to those experiments or observations \(c_k\) within the total evidence stream \(c^n\) for which some of the possible outcomes have 0 likelihood of occurring according to hypothesis \(h_j\) but have non-0 likelihood of occurring according to \(h_i\). Such outcomes are highly desirable. If they occur, the likelihood ratio comparing \(h_j\) to \(h_i\) will become 0, and \(h_j\) will be falsified . So-called crucial experiments are a special case of this, where for at least one possible outcome \(o_{ku}\), \(P[o_{ku} \pmid h_{i}\cdot b\cdot c_{k}] = 1\) and \(P[o_{ku} \pmid h_{j}\cdot b\cdot c_{k}] = 0\). In the more general case \(h_i\) together with b says that one of the outcomes of \(c_k\) is at least minimally probable, whereas \(h_j\) says that this outcome is impossible—i.e., \(P[o_{ku} \pmid h_{i}\cdot b\cdot c_{k}] \gt 0\) and \(P[o_{ku} \pmid h_{j}\cdot b\cdot c_{k}] = 0\). It will be convenient to define a term for this situation.
Definition: Full Outcome Compatibility. Let’s call \(h_j\) fully outcome-compatible with \(h_i\) on experiment or observation \(c_k\) just when , for each of its possible outcomes \(e_k\), if \(P[e_k \pmid h_{i}\cdot b\cdot c_{k}] \gt 0\), then \(P[e_k \pmid h_{j}\cdot b\cdot c_{k}] \gt 0\). Equivalently, \(h_j\) is fails to be fully outcome-compatible with \(h_i\) on experiment or observation \(c_k\) just when , for at least one of its possible outcomes \(e_k\), \(P[e_k \pmid h_{i}\cdot b\cdot c_{k}] \gt 0\) but \(P[e_k \pmid h_{j}\cdot b\cdot c_{k}] = 0\).
The first part of the Likelihood Ratio Convergence Theorem applies to that part of the total stream of evidence (i.e., that subsequence of the total evidence stream) on which hypothesis \(h_j\) fails to be fully outcome-compatible with hypothesis \(h_i\); the second part of the theorem applies to the remaining part of the total stream of evidence, that subsequence of the total evidence stream on which \(h_j\) is fully outcome-compatible with \(h_i\). It turns out that these two kinds of cases must be treated differently. (This is due to the way in which the expected information content for empirically distinguishing between the two hypotheses will be measured for experiments and observations that are fully outcome compatible ; this measure of information content blows up (becomes infinite) for experiments and observations that fail to be fully outcome compatible ). Thus, the following part of the convergence theorem applies to just that part of the total stream of evidence that consists of experiments and observations that fail to be fully outcome compatible for the pair of hypotheses involved. Here, then, is the first part of the convergence theorem.
Likelihood Ratio Convergence Theorem 1—The Falsification Theorem: Suppose that the total stream of evidence \(c^n\) contains precisely m experiments or observations on which \(h_j\) fails to be fully outcome-compatible with \(h_i\). And suppose that the Independent Evidence Conditions hold for evidence stream \(c^n\) with respect to each of these two hypotheses. Furthermore, suppose there is a lower bound \(\delta \gt 0\) such that for each \(c_k\) on which \(h_j\) fails to be fully outcome-compatible with \(h_i\),
—i.e., \(h_i\) together with \(b\cdot c_k\) says , with likelihood at least as large as \(\delta\), that one of the outcomes will occur that \(h_j\) says cannot occur. Then,
which approaches 1 for large m . (For proof see Proof of the Falsification Theorem .)
In other words, we only suppose that for each of m observations, \(c_k, h_i\) says observation \(c_k\) has at least a small likelihood \(\delta\) of producing one of the outcomes \(o_{ku}\) that \(h_j\) says is impossible. If the number m of such experiments or observations is large enough (or if the lower bound \(\delta\) on the likelihoods of getting such outcomes is large enough), and if \(h_i\) (together with \(b\cdot c^n)\) is true, then it is highly likely that one of the outcomes held to be impossible by \(h_j\) will actually occur. If one of these outcomes does occur, then the likelihood ratio for \(h_j\) as compared to over \(h_i\) will become 0. According to Bayes’ Theorem, when this happen, \(h_j\) is absolutely refuted by the evidence—its posterior probability becomes 0.
The Falsification Theorem is quite commonsensical. First, notice that if there is a crucial experiment in the evidence stream, the theorem is completely obvious. That is, suppose for the specific experiment \(c_k\) (in evidence stream \(c^n)\) there are two incompatible possible outcomes \(o_{kv}\) and \(o_{ku}\) such that \(P[o_{kv} \pmid h_{j}\cdot b\cdot c_{k}] = 1\) and \(P[o_{ku} \pmid h_{i}\cdot b\cdot c_{k}] = 1\). Then, clearly, \(P[\vee \{ o_{ku}: P[o_{ku} \pmid h_{j}\cdot b\cdot c_{k}] = 0\} \pmid h_{i}\cdot b\cdot c_{k}] = 1\), since \(o_{ku}\) is one of the \(o_{ku}\) such that \(P[o_{ku} \pmid h_{j}\cdot b\cdot c_{k}] = 0\). So, where a crucial experiment is available, the theorem applies with \(m = 1\) and \(\delta = 1\).
The theorem is equally commonsensical for cases where no crucial experiment is available. To see what it says in such cases, consider an example. Let \(h_i\) be some theory that implies a specific rate of proton decay, but a rate so low that there is only a very small probability that any particular proton will decay in a given year. Consider an alternative theory \(h_j\) that implies that protons never decay. If \(h_i\) is true, then for a persistent enough sequence of observations (i.e., if proper detectors can keep trillions of protons under observation for long enough), eventually a proton decay will almost surely be detected. When this happens, the likelihood ratio becomes 0. Thus, the posterior probability of \(h_j\) becomes 0.
It is instructive to plug some specific values into the formula given by the Falsification Theorem, to see what the convergence rate might look like. For example, the theorem tells us that if we compare any pair of hypotheses \(h_i\) and \(h_j\) on an evidence stream \(c^n\) that contains at least \(m = 19\) observations or experiments, where each has a likelihood \(\delta \ge .10\) of yielding a falsifying outcome , then the likelihood (on \(h_{i}\cdot b\cdot c^{n})\) of obtaining an outcome sequence \(e^n\) that yields likelihood-ratio
will be at least as large as \((1 - (1-.1)^{19}) = .865\). (The reader is invited to try other values of \(\delta\) and m .)
A comment about the need for and usefulness of such convergence theorems is in order, now that we’ve seen one. Given some specific pair of scientific hypotheses \(h_i\) and \(h_j\) one may directly compute the likelihood, given \((h_{i}\cdot b\cdot c^{n})\), that a proposed sequence of experiments or observations \(c^n\) will result in one of the sequences of outcomes that would yield low likelihood ratios. So, given a specific pair of hypotheses and a proposed sequence of experiments, we don’t need a general Convergence Theorem to tell us the likelihood of obtaining refuting evidence. The specific hypotheses \(h_i\) and \(h_j\) tell us this themselves . They tell us the likelihood of obtaining each specific outcome stream, including those that either refute the competitor or produce a very small likelihood ratio for it. Furthermore, after we’ve actually performed an experiment and recorded its outcome, all that matters is the actual ratio of likelihoods for that outcome. Convergence theorems become moot.
The point of the Likelihood Ratio Convergence Theorem (both the Falsification Theorem and the part of the theorem still to come) is to assure us in advance of considering any specific pair of hypotheses that if the possible evidence streams that test hypotheses have certain characteristics which reflect the empirical distinctness of the two hypotheses, then it is highly likely that one of the sequences of outcomes will occur that yields a very small likelihood ratio. These theorems provide finite lower bounds on how quickly such convergence is likely to be. Thus, they show that the CoA is satisfied in advance of our using the logic to test specific pairs of hypotheses against one another.
The Falsification Theorem applies whenever the evidence stream includes possible outcomes that may falsify the alternative hypothesis. However, it completely ignores the influence of any experiments or observations in the evidence stream on which hypothesis \(h_j\) is fully outcome-compatible with hypothesis \(h_i\). We now turn to a theorem that applies to those evidence streams (or to parts of evidence streams) consisting only of experiments and observations on which hypothesis \(h_j\) is fully outcome-compatible with hypothesis \(h_i\). Evidence streams of this kind contain no possibly falsifying outcomes. In such cases the only outcomes of an experiment or observation \(c_k\) for which hypothesis \(h_j\) may specify 0 likelihoods are those for which hypothesis \(h_i\) specifies 0 likelihoods as well.
Hypotheses whose connection with the evidence is entirely statistical in nature will usually be fully outcome-compatible on the entire evidence stream. So, evidence streams of this kind are undoubtedly much more common in practice than those containing possibly falsifying outcomes. Furthermore, whenever an entire stream of evidence contains some mixture of experiments and observations on which the hypotheses are not fully outcome compatible along with others on which they are fully outcome compatible , we may treat the experiments and observations for which full outcome compatibility holds as a separate subsequence of the entire evidence stream, to see the likely impact of that part of the evidence in producing values for likelihood ratios.
To cover evidence streams (or subsequences of evidence streams) consisting entirely of experiments or observations on which \(h_j\) is fully outcome-compatible with hypothesis \(h_i\) we will first need to identify a useful way to measure the degree to which hypotheses are empirically distinct from one another on such evidence. Consider some particular sequence of outcomes \(e^n\) that results from observations \(c^n\). The likelihood ratio \(P[e^n \pmid h_{j}\cdot b\cdot c^{n}] / P[e^n \pmid h_{i}\cdot b\cdot c^{n}]\) itself measures the extent to which the outcome sequence distinguishes between \(h_i\) and \(h_j\). But as a measure of the power of evidence to distinguish among hypotheses, raw likelihood ratios provide a rather lopsided scale, a scale that ranges from 0 to infinity with the midpoint, where \(e^n\) doesn’t distinguish at all between \(h_i\) and \(h_j\), at 1. So, rather than using raw likelihood ratios to measure the ability of \(e^n\) to distinguish between hypotheses, it proves more useful to employ a symmetric measure. The logarithm of the likelihood ratio provides such a measure.
Definition: QI—the Quality of the Information . For each experiment or observation \(c_k\), define the quality of the information provided by possible outcome \(o_{ku}\) for distinguishing \(h_j\) from \(h_i\), given b , as follows (where henceforth we take “logs” to be base-2):
Similarly, for the sequence of experiments or observations \(c^n\), define the quality of the information provided by possible outcome \(e^n\) for distinguishing \(h_j\) from \(h_i\), given b , as follows:
That is, QI is the base-2 logarithm of the likelihood ratio for \(h_i\) over that for \(h_j\).
So, we’ll measure the Quality of the Information an outcome would yield in distinguishing between two hypotheses as the base-2 logarithm of the likelihood ratio. This is clearly a symmetric measure of the outcome’s evidential strength at distinguishing between the two hypotheses. On this measure hypotheses \(h_i\) and \(h_j\) assign the same likelihood value to a given outcome \(o_{ku}\) just when \(\QI[o_{ku} \pmid h_i /h_j \pmid b\cdot c_k] = 0\). Thus, QI measures information on a logarithmic scale that is symmetric about the natural no-information midpoint, 0. This measure is set up so that positive information favors \(h_i\) over \(h_j\), and negative information favors \(h_j\) over \(h_i\).
Given the Independent Evidence Assumptions with respect to each hypothesis, it’s easy to show that the QI for a sequence of outcomes is just the sum of the QIs of the individual outcomes in the sequence:
Probability theorists measure the expected value of a quantity by first multiplying each of its possible values by their probabilities of occurring, and then summing these products. Thus, the expected value of QI is given by the following formula:
Definition: EQI—the Expected Quality of the Information . We adopt the convention that if \(P[o_{ku} \pmid h_{i}\cdot b\cdot c_{k}] = 0\), then the term \(\QI[o_{ku} \pmid h_i /h_j \pmid b\cdot c_k] \times P[o_{ku} \pmid h_{i}\cdot b\cdot c_{k}] = 0\). This convention will make good sense in the context of the following definition because, whenever the outcome \(o_{ku}\) has 0 probability of occurring according to \(h_i\) (together with \(b\cdot c_k)\), it makes good sense to give it 0 impact on the ability of the evidence to distinguish between \(h_j\) and \(h_i\) when \(h_i\) (together with \(b\cdot c_k)\) is true. Also notice that the full outcome-compatibility of \(h_j\) with \(h_i\) on \(c_k\) means that whenever \(P[e_k \pmid h_{j}\cdot b\cdot c_{k}] = 0\), we must have \(P[e_k \pmid h_{i}\cdot b\cdot c_{k}] = 0\) as well; so whenever the denominator would be 0 in the term
the convention just described makes the term
Thus the following notion is well-defined:
For \(h_j\) fully outcome-compatible with \(h_i\) on experiment or observation \(c_k\), define
Also, for \(h_j\) fully outcome-compatible with \(h_i\) on each experiment and observation in the sequence \(c^n\), define
The EQI of an experiment or observation is the Expected Quality of its Information for distinguishing \(h_i\) from \(h_j\) when \(h_i\) is true. It is a measure of the expected evidential strength of the possible outcomes of an experiment or observation at distinguishing between the hypotheses when \(h_i\) (together with \(b\cdot c)\) is true. Whereas QI measures the ability of each particular outcome or sequence of outcomes to empirically distinguish hypotheses, EQI measures the tendency of experiments or observations to produce distinguishing outcomes. It can be shown that EQI tracks empirical distinctness in a very precise way. We return to this in a moment.
It is easily seen that the EQI for a sequence of observations \(c^n\) is just the sum of the EQIs of the individual observations \(c_k\) in the sequence:
(For proof see the supplement Proof that the EQI for \(c^n\) is the sum of the EQI for the individual \(c_k\) .)
This suggests that it may be useful to average the values of the \(\EQI[c_k \pmid h_i /h_j \pmid b]\) over the number of observations n to obtain a measure of the average expected quality of the information among the experiments and observations that make up the evidence stream \(c^n\).
Definition: The Average Expected Quality of Information For \(h_j\) fully outcome-compatible with \(h_i\) on each experiment and observation in the evidence stream \(c^n\), define the average expected quality of information, \(\bEQI\), from \(c^n\) for distinguishing \(h_j\) from \(h_i\), given \(h_i\cdot b\), as follows:
It turns out that the value of \(\EQI[c_k \pmid h_i /h_j \pmid b_{}]\) cannot be less than 0; and it must be greater than 0 just in case \(h_i\) is empirically distinct from \(h_j\) on at least one outcome \(o_{ku}\)—i.e., just in case it is empirically distinct in the sense that \(P[o_{ku} \pmid h_{i}\cdot b\cdot c_{k}] \ne P[o_{ku} \pmid h_{j}\cdot b\cdot c_{k}]\), for at least one outcome \(o_{ku}\). The same goes for the average, \(\bEQI[c^n \pmid h_i /h_j \pmid b]\).
Theorem: Nonnegativity of EQI.
\(\EQI[c_k \pmid h_i /h_j \pmid b_{}] \ge 0\); and \(\EQI[c_k \pmid h_i /h_j \pmid b_{}] \gt 0\) if and only if for at least one of its possible outcomes \(o_{ku}\),
As a result, \(\bEQI[c^n \pmid h_i /h_j \pmid b] \ge 0\); and \(\bEQI[c^n \pmid h_i /h_j \pmid b] \gt 0\) if and only if at least one experiment or observation \(c_k\) has at least one possible outcome \(o_{ku}\) such that
(For proof, see the supplement The Effect on EQI of Partitioning the Outcome Space More Finely—Including Proof of the Nonnegativity of EQI .)
In fact, the more finely one partitions the outcome space \(O_{k} = \{o_{k1},\ldots ,o_{kv},\ldots ,o_{kw}\}\) into distinct outcomes that differ on likelihood ratio values, the larger EQI becomes. [ 15 ] This shows that EQI tracks empirical distinctness in a precise way. The importance of the Non-negativity of EQI result for the Likelihood Ratio Convergence Theorem will become clear in a moment.
We are now in a position to state the second part of the Likelihood Ratio Convergence Theorem . It applies to all evidence streams not containing possibly falsifying outcomes for \(h_j\) when \(h_i\) holds—i.e., it applies to all evidence streams for which \(h_j\) is fully outcome-compatible with \(h_i\) on each \(c_k\) in the stream.
Likelihood Ratio Convergence Theorem 2—The Probabilistic Refutation Theorem.
Suppose the evidence stream \(c^n\) contains only experiments or observations on which \(h_j\) is fully outcome-compatible with \(h_i\)—i.e., suppose that for each condition \(c_k\) in sequence \(c^n\), for each of its possible outcomes possible outcomes \(o_{ku}\), either \(P[o_{ku} \pmid h_{i}\cdot b\cdot c_{k}] = 0\) or \(P[o_{ku} \pmid h_{j}\cdot b\cdot c_{k}] \gt 0\). In addition (as a slight strengthening of the previous supposition), for some \(\gamma \gt 0\) a number smaller than \(1/e^2\) (\(\approx .135\); where e ’ is the base of the natural logarithm), suppose that for each possible outcome \(o_{ku}\) of each observation condition \(c_k\) in \(c^n\), either \(P[o_{ku} \pmid h_{i}\cdot b\cdot c_{k}] = 0\) or
And suppose that the Independent Evidence Conditions hold for evidence stream \(c^n\) with respect to each of these hypotheses. Now, choose any positive \(\varepsilon \lt 1\), as small as you like, but large enough (for the number of observations n being contemplated) that the value of
For \(\varepsilon = 1/2^m\) and \(\gamma = 1/2^q\), this formula becomes,
(For proof see the supplement Proof of the Probabilistic Refutation Theorem .)
This theorem provides sufficient conditions for the likely refutation of false alternatives via exceeding small likelihood ratios. The conditions under which this happens characterize the degree to which the hypotheses involved are empirically distinct from one another. The theorem says that when these conditions are met, according to hypothesis \(h_i\) (taken together with \(b\cdot c^n)\), the likelihood is near 1 that that one of the outcome sequence \(e^n\) will occur for which the likelihood ratio is smaller than \(\varepsilon\) (for any value of \(\varepsilon\) you may choose). The likelihood of getting such an evidential outcome \(e^n\) is quite close to 1—i.e., no more than the amount
below 1. (Notice that this amount below 1 goes to 0 as n increases.)
It turns out that in almost every case (for almost any pair of hypotheses) the actual likelihood of obtaining such evidence (i.e., evidence that has a likelihood ratio value less than \(\varepsilon)\) will be much closer to 1 than this factor indicates. [ 16 ] Thus, the theorem provides an overly cautious lower bound on the likelihood of obtaining small likelihood ratios. It shows that the larger the value of \(\bEQI\) for an evidence stream, the more likely that stream is to produce a sequence of outcomes that yield a very small likelihood ratio value. But even if \(\bEQI\) remains quite small, a long enough evidence stream, n , of such low-grade evidence will, nevertheless, almost surely produce an outcome sequence having a very small likelihood ratio value. [ 17 ]
Notice that the antecedent condition of the theorem, that “either
for some \(\gamma \gt 0\) but less than \(1/e^2\) (\(\approx .135\))”, does not favor hypothesis \(h_i\) over \(h_j\) in any way. The condition only rules out the possibility that some outcomes might furnish extremely strong evidence against \(h_j\) relative to \(h_i\)—by making \(P[o_{ku} \pmid h_{i}\cdot b\cdot c_{k}] = 0\) or by making
less than some quite small \(\gamma\). This condition is only needed because our measure of evidential distinguishability, QI, blows up when the ratio
is extremely small. Furthermore, this condition is really no restriction at all on possible experiments or observations. If \(c_k\) has some possible outcome sentence \(o_{ku}\) that would make
(for a given small \(\gamma\) of interest), one may disjunctively lump \(o_{ku}\) together with some other outcome sentence \(o_{kv}\) for \(c_k\). Then, the antecedent condition of the theorem will be satisfied, but with the sentence ‘\((o_{ku} \vee o_{kv})\)’ treated as a single outcome. It can be proved that the only effect of such “disjunctive lumping” is to make \(\bEQI\) smaller than it would otherwise be (whereas larger values of \(\bEQI\) are more desirable). If the too strongly refuting disjunct \(o_{ku}\) actually occurs when the experiment or observation \(c_k\) is conducted, all the better, since this results in a likelihood ratio
smaller than \(\gamma\) on that particular evidential outcome. We merely failed to take this more strongly refuting possibility into account when computing our lower bound on the likelihood that refutation via likelihood ratios would occur.
The point of the two Convergence Theorems explored in this section is to assure us, in advance of the consideration of any specific pair of hypotheses, that if the possible evidence streams that test them have certain characteristics which reflect their evidential distinguishability, it is highly likely that outcomes yielding small likelihood ratios will result. These theorems provide finite lower bounds on how quickly convergence is likely to occur. Thus, there is no need to wait through some infinitely long run for convergence to occur. Indeed, for any evidence sequence on which the probability distributions are at all well behaved, the actual likelihood of obtaining outcomes that yield small likelihood ratio values will inevitably be much higher than the lower bounds given by Theorems 1 and 2.
In sum, according to Theorems 1 and 2, each hypothesis \(h_i\) says , via likelihoods, that given enough observations, it is very likely to dominate its empirically distinct rivals in a contest of likelihood ratios. The true hypothesis speaks truthfully about this, and its competitors lie. Even a sequence of observations with an extremely low average expected quality of information is very likely to do the job if that evidential sequence is long enough. Thus (by Equation 9* ), as evidence accumulates, the degree of support for false hypotheses will very probably approach 0, indicating that they are probably false; and as this happens, (by Equations 10 and 11) the degree of support for the true hypothesis will approach 1, indicating its probable truth. Thus, the Criterion of Adequacy (CoA) is satisfied.
Up to this point we have been supposing that likelihoods possess objective or agreed numerical values. Although this supposition is often satisfied in scientific contexts, there are important settings where it is unrealistic, where hypotheses only support vague likelihood values, and where there is enough ambiguity in what hypotheses say about evidential claims that the scientific community cannot agree on precise values for the likelihoods of evidential claims. [ 18 ] Let us now see how the supposition of precise, agreed likelihood values may be relaxed in a reasonable way.
Recall why agreement, or near agreement, on precise values for likelihoods is so important to the scientific enterprise. To the extent that members of a scientific community disagree on the likelihoods, they disagree about the empirical content of their hypotheses, about what each hypothesis says about how the world is likely to be. This can lead to disagreement about which hypotheses are refuted or supported by a given body of evidence. Similarly, to the extent that the values of likelihoods are only vaguely implied by hypotheses as understood by an individual agent, that agent may be unable to determine which of several hypotheses is refuted or supported by a given body of evidence.
We have seen, however, that the individual values of likelihoods are not really crucial to the way evidence impacts hypotheses. Rather, as Equations 9–11 show, it is ratios of likelihoods that do the heavy lifting. So, even if two support functions \(P_{\alpha}\) and \(P_{\beta}\) disagree on the values of individual likelihoods, they may, nevertheless, largely agree on the refutation or support that accrues to various rival hypotheses, provided that the following condition is satisfied:
- whenever possible outcome sequence \(e^n\) makes \[\frac{P_{\alpha}[e^n \pmid h_{j}\cdot b\cdot c^{n}]}{P_{\alpha}[e^n \pmid h_{i}\cdot b\cdot c^{n}]} \lt 1,\] it also makes \[\frac{P_{\beta}[e^n \pmid h_{j}\cdot b\cdot c^{n}]}{P_{\beta}[e^n \pmid h_{i}\cdot b\cdot c^{n}]} \lt 1;\]
- whenever possible outcome sequence \(e^n\) makes \[\frac{P_{\alpha}[e^n \pmid h_{j}\cdot b\cdot c^{n}]}{P_{\alpha}[e^n \pmid h_{i}\cdot b\cdot c^{n}]} \gt 1,\] it also makes \[\frac{P_{\beta}[e^n \pmid h_{j}\cdot b\cdot c^{n}]}{P_{\beta}[e^n \pmid h_{i}\cdot b\cdot c^{n}]} \gt 1;\]
- each of these likelihood ratios is either close to 1 for both of these support functions, or is quite far from 1 for both of them. [ 19 ]
When this condition holds, the evidence will support \(h_i\) over \(h_j\) according to \(P_{\alpha}\) just in case it does so for \(P_{\beta}\) as well, although the strength of support may differ. Furthermore, although the rate at which the likelihood ratios increase or decrease on a stream of evidence may differ for the two support functions, the impact of the cumulative evidence should ultimately affect their refutation or support in much the same way.
When likelihoods are vague or diverse, we may take an approach similar to that we employed for vague and diverse prior plausibility assessments. We may extend the vagueness sets for individual agents to include a collection of inductive support functions that cover the range of values for likelihood ratios of evidence claims (as well as cover the ranges of comparative support strengths for hypotheses due to plausibility arguments within b , as represented by ratios of prior probabilities). Similarly, we may extend the diversity sets for communities of agents to include support functions that cover the ranges of likelihood ratio values that arise within the vagueness sets of members of the scientific community.
This broadening of vagueness and diversity sets to accommodate vague and diverse likelihood values makes no trouble for the convergence to truth results for hypotheses. For, provided that the Directional Agreement Condition is satisfied by all support functions in an extended vagueness or diversity set under consideration, the Likelihood Ratio Convergence Theorem applies to each individual support function in that set. For, the the proof of that convergence theorem doesn’t depend on the supposition that likelihoods are objective or have intersubjectively agreed values. Rather, it applies to each individual support function \(P_{\alpha}\). The only possible problem with applying this result across a range of support functions is that when their values for likelihoods differ, function \(P_{\alpha}\) may disagree with \(P_{\beta}\) on which of the hypotheses is favored by a given sequence of evidence. That can happen because different support functions may represent the evidential import of hypotheses differently, by specifying different likelihood values for the very same evidence claims. So, an evidence stream that favors \(h_i\) according to \(P_{\alpha}\) may instead favor \(h_j\) according to \(P_{\beta}\). However, when the Directional Agreement Condition holds for a given collection of support functions, this problem cannot arise. Directional Agreement means that the evidential import of hypotheses is similar enough for \(P_{\alpha}\) and \(P_{\beta}\) that a sequence of outcomes may favor a hypothesis according to \(P_{\alpha}\) only if it does so for \(P_{\beta}\) as well.
Thus, when the Directional Agreement Condition holds for all support functions in a vagueness or diversity set that is extended to include vague or diverse likelihoods, and provided that enough evidentially distinguishing experiments or observations can be performed, all support functions in the extended vagueness or diversity set will very probably come to agree that the likelihood ratios for empirically distinct false competitors of a true hypothesis are extremely small. As that happens, the community comes to agree on the refutation of these competitors, and the true hypothesis rises to the top of the heap. [ 20 ]
What if the true hypothesis has evidentially equivalent rivals? Their posterior probabilities must rise as well. In that case we are only assured that the disjunction of the true hypothesis with its evidentially equivalent rivals will be driven to 1 as evidence lays low its evidentially distinct rivals. The true hypothesis will itself approach 1 only if either it has no evidentially equivalent rivals, or whatever equivalent rivals it does have can be laid low by plausibility arguments of a kind that don’t depend on the evidential likelihoods, but only show up via the comparative plausibility assessments represented by ratios of prior probabilities.
- Enumerative Inductions: Bayesian Estimation and Convergence
- Some Prominent Approaches to the Representation of Uncertain Inference
- Likelihood Ratios, Likelihoodism, and the Law of Likelihood
- Immediate Consequences of the Independent Evidence Conditions
- Proof of the Falsification Theorem
- Proof that the EQI for \(c^n\) is the sum of EQI for the individual \(c_k\)
- The Effect on EQI of Partitioning the Outcome Space More Finely—Including Proof of the Nonnegativity of EQI
- Proof of the Probabilistic Refutation Theorem
- Boole, George, 1854, The Laws of Thought , London: MacMillan. Republished in 1958 by Dover: New York.
- Bovens, Luc and Stephan Hartmann, 2003, Bayesian Epistemology , Oxford: Oxford University Press. doi:10.1093/0199269750.001.0001
- Carnap, Rudolf, 1950, Logical Foundations of Probability , Chicago: University of Chicago Press.
- –––, 1952, The Continuum of Inductive Methods , Chicago: University of Chicago Press.
- –––, 1963, “Replies and Systematic Expositions”, in The Philosophy of Rudolf Carnap , Paul Arthur Schilpp (ed.),La Salle, IL: Open Court.
- Chihara, Charles S., 1987, “Some Problems for Bayesian Confirmation Theory”, British Journal for the Philosophy of Science , 38(4): 551–560. doi:10.1093/bjps/38.4.551
- Christensen, David, 1999, “Measuring Confirmation”, Journal of Philosophy , 96(9): 437–61. doi:10.2307/2564707
- –––, 2004, Putting Logic in its Place: Formal Constraints on Rational Belief , Oxford: Oxford University Press. doi:10.1093/0199263256.001.0001
- De Finetti, Bruno, 1937, “La Prévision: Ses Lois Logiques, Ses Sources Subjectives”, Annales de l’Institut Henri Poincaré , 7: 1–68; translated by Henry E. Kyburg, Jr. as “Foresight. Its Logical Laws, Its Subjective Sources”, in Studies in Subjective Probability , Henry E. Kyburg, Jr. and H.E. Smokler (eds.), Robert E. Krieger Publishing Company, 1980.
- Dowe, David L., Steve Gardner, and Graham Oppy, 2007, “Bayes, Not Bust! Why Simplicity is No Problem for Bayesians”, British Journal for the Philosophy of Science , 58(4): 709–754. doi:10.1093/bjps/axm033
- Dubois, Didier J. and Henri Prade, 1980, Fuzzy Sets and Systems , (Mathematics in Science and Engineering, 144), New York: Academic Press.
- –––, 1990, “An Introduction to Possibilistic and Fuzzy Logics”, in Glenn Shafer and Judea Pearl (eds.), Readings in Uncertain Reasoning , San Mateo, CA: Morgan Kaufmann, 742–761..
- Duhem, P., 1906, La theorie physique. Son objet et sa structure , Paris: Chevalier et Riviere; translated by P.P. Wiener, The Aim and Structure of Physical Theory , Princeton, NJ: Princeton University Press, 1954.
- Earman, John, 1992, Bayes or Bust? A Critical Examination of Bayesian Confirmation Theory , Cambridge, MA: MIT Press.
- Edwards, A.W.F., 1972, Likelihood: an account of the statistical concept of likelihood and its application to scientific inference , Cambridge: Cambridge University Press.
- Edwards, Ward, Harold Lindman, and Leonard J. Savage, 1963, “Bayesian Statistical Inference for Psychological Research”, Psychological Review , 70(3): 193–242. doi:10.1037/h0044139
- Eells, Ellery, 1985, “Problems of Old Evidence”, Pacific Philosophical Quarterly , 66(3–4): 283–302. doi:10.1111/j.1468-0114.1985.tb00254.x
- –––, 2006, “Confirmation Theory”, Sarkar and Pfeifer 2006..
- Eells, Ellery and Branden Fitelson, 2000, “Measuring Confirmation and Evidence”, Journal of Philosophy , 97(12): 663–672. doi:10.2307/2678462
- Field, Hartry H., 1977, “Logic, Meaning, and Conceptual Role”, Journal of Philosophy , 74(7): 379–409. doi:10.2307/2025580
- Fisher, R.A., 1922, “On the Mathematical Foundations of Theoretical Statistics”, Philosophical Transactions of the Royal Society, series A , 222(594–604): 309–368. doi:10.1098/rsta.1922.0009
- Fitelson, Branden, 1999, “The Plurality of Bayesian Measures of Confirmation and the Problem of Measure Sensitivity”, Philosophy of Science , 66: S362–S378. doi:10.1086/392738
- –––, 2001, “A Bayesian Account of Independent Evidence with Applications”, Philosophy of Science , 68(S3): S123–S140. doi:10.1086/392903
- –––, 2002, “Putting the Irrelevance Back Into the Problem of Irrelevant Conjunction”, Philosophy of Science , 69(4): 611–622. doi:10.1086/344624
- –––, 2006, “Inductive Logic”, Sarkar and Pfeifer 2006..
- –––, 2006, “Logical Foundations of Evidential Support”, Philosophy of Science , 73(5): 500–512. doi:10.1086/518320
- –––, 2007, “Likelihoodism, Bayesianism, and Relational Confirmation”, Synthese , 156(3): 473–489. doi:10.1007/s11229-006-9134-9
- Fitelson, Branden and James Hawthorne, 2010, “How Bayesian Confirmation Theory Handles the Paradox of the Ravens”, in Eells and Fetzer (eds.), The Place of Probability in Science , Open Court. [ Fitelson & Hawthorne 2010 preprint available from the author (PDF) ]
- Forster, Malcolm and Elliott Sober, 2004, “Why Likelihood”, in Mark L. Taper and Subhash R. Lele (eds.), The Nature of Scientific Evidence , Chicago: University of Chicago Press.
- Friedman, Nir and Joseph Y. Halpern, 1995, “Plausibility Measures: A User’s Guide”, in UAI 95: Proceedings of the Eleventh Conference on Uncertainty in Artificial Intelligence , 175–184.
- Gaifman, Haim and Marc Snir, 1982, “Probabilities Over Rich Languages, Testing and Randomness”, Journal of Symbolic Logic , 47(3): 495–548. doi:10.2307/2273587
- Gillies, Donald, 2000, Philosophical Theories of Probability , London: Routledge.
- Glymour, Clark N., 1980, Theory and Evidence , Princeton, NJ: Princeton University Press.
- Goodman, Nelson, 1983, Fact, Fiction, and Forecast , 4 th edition, Cambridge, MA: Harvard University Press.
- Hacking, Ian, 1965, Logic of Statistical Inference , Cambridge: Cambridge University Press.
- –––, 1975, The Emergence of Probability: a Philosophical Study of Early Ideas about Probability, Induction and Statistical Inference , Cambridge: Cambridge University Press. doi:10.1017/CBO9780511817557
- –––, 2001, An Introduction to Probability and Inductive Logic , Cambridge: Cambridge University Press. doi:10.1017/CBO9780511801297
- Hájek, Alan, 2003a, “What Conditional Probability Could Not Be”, Synthese , 137(3):, 273–323. doi:10.1023/B:SYNT.0000004904.91112.16
- –––, 2003b, “Interpretations of the Probability Calculus”, in the Stanford Encyclopedia of Philosophy , (Summer 2003 Edition), Edward N. Zalta (ed.), URL = < https://plato.stanford.edu/archives/sum2003/entries/probability-interpret/ >
- –––, 2005, “Scotching Dutch Books?” Philosophical Perspectives , 19 (Epistemology): 139–151. doi:10.1111/j.1520-8583.2005.00057.x
- –––, 2007, “The Reference Class Problem is Your Problem Too”, Synthese , 156(3): 563–585. doi:10.1007/s11229-006-9138-5
- Halpern, Joseph Y., 2003, Reasoning About Uncertainty , Cambridge, MA: MIT Press.
- Harper, William L., 1976, “Rational Belief Change, Popper Functions and Counterfactuals”, in Harper and Hooker 1976: 73–115. doi:10.1007/978-94-010-1853-1_5
- Harper, William L. and Clifford Alan Hooker (eds.), 1976, Foundations of Probability Theory, Statistical Inference, and Statistical Theories of Science, volume I Foundations and Philosophy of Epistemic Applications of Probability Theory , (The Western Ontario Series in Philosophy of Science, 6a), Dordrecht: Reidel. doi:10.1007/978-94-010-1853-1
- Hawthorne, James, 1993, “Bayesian Induction is Eliminative Induction”, Philosophical Topics , 21(1): 99–138. doi:10.5840/philtopics19932117
- –––, 1994,“On the Nature of Bayesian Convergence”, PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association 1994 , 1: 241–249. doi:10.1086/psaprocbienmeetp.1994.1.193029
- –––, 2005, “ Degree-of-Belief and Degree-of-Support : Why Bayesians Need Both Notions”, Mind , 114(454): 277–320. doi:10.1093/mind/fzi277
- –––, 2009, “The Lockean Thesis and the Logic of Belief”, in Franz Huber and Christoph Schmidt-Petri (eds.), Degrees of Belief , (Synthese Library, 342), Dordrecht: Springer, pp. 49–74. doi:10.1007/978-1-4020-9198-8_3
- Hawthorne, James and Luc Bovens, 1999, “The Preface, the Lottery, and the Logic of Belief”, Mind , 108(430): 241–264. doi:10.1093/mind/108.430.241
- Hawthorne, James and Branden Fitelson, 2004, “Discussion: Re-solving Irrelevant Conjunction With Probabilistic Independence”, Philosophy of Science , 71(4): 505–514. doi:10.1086/423626
- Hellman, Geoffrey, 1997, “Bayes and Beyond”, Philosophy of Science , 64(2): 191–221. doi:10.1086/392548
- Hempel, Carl G., 1945, “Studies in the Logic of Confirmation”, Mind , 54(213): 1–26, 54(214):97–121. doi:10.1093/mind/LIV.213.1 doi:10.1093/mind/LIV.214.97
- Horwich, Paul, 1982, Probability and Evidence , Cambridge: Cambridge University Press. doi:10.1017/CBO9781316494219
- Howson, Colin, 1997, “A Logic of Induction”, Philosophy of Science , 64(2): 268–290. doi:10.1086/392551
- –––, 2000, Hume’s Problem: Induction and the Justification of Belief , Oxford: Oxford University Press. doi:10.1093/0198250371.001.0001
- –––, 2002, “Bayesianism in Statistics“, in Swinburne 2002: 39–71. doi:10.5871/bacad/9780197263419.003.0003
- –––, 2007, “Logic With Numbers”, Synthese , 156(3): 491–512. doi:10.1007/s11229-006-9135-8
- Howson, Colin and Peter Urbach, 1993, Scientific Reasoning: The Bayesian Approach , La Salle, IL: Open Court. [3rd edition, 2005.]
- Huber, Franz, 2005a, “Subjective Probabilities as Basis for Scientific Reasoning?” British Journal for the Philosophy of Science , 56(1): 101–116. doi:10.1093/phisci/axi105
- –––, 2005b, “What Is the Point of Confirmation?” Philosophy of Science , 72(5): 1146–1159. doi:10.1086/508961
- Jaynes, Edwin T., 1968, “Prior Probabilities”, IEEE Transactions on Systems Science and Cybernetics , SSC–4(3): 227–241. doi:10.1109/TSSC.1968.300117
- Jeffrey, Richard C., 1983, The Logic of Decision , 2nd edition, Chicago: University of Chicago Press.
- –––, 1987, “Alias Smith and Jones: The Testimony of the Senses”, Erkenntnis , 26(3): 391–399. doi:10.1007/BF00167725
- –––, 1992, Probability and the Art of Judgment , New York: Cambridge University Press. doi:10.1017/CBO9781139172394
- –––, 2004, Subjective Probability: The Real Thing , Cambridge: Cambridge University Press. doi:10.1017/CBO9780511816161
- Jeffreys, Harold, 1939, Theory of Probability , Oxford: Oxford University Press.
- Joyce, James M., 1998, “A Nonpragmatic Vindication of Probabilism”, Philosophy of Science , 65(4): 575–603. doi:10.1086/392661
- –––, 1999, The Foundations of Causal Decision Theory , New York: Cambridge University Press. doi:10.1017/CBO9780511498497
- –––, 2003, “Bayes’ Theorem”, in the Stanford Encyclopedia of Philosophy , (Summer 2003 Edition), Edward N. Zalta (ed.), URL = < https://plato.stanford.edu/archives/win2003/entries/bayes-theorem/ >
- –––, 2004, “Bayesianism”, in Alfred R. Mele and Piers Rawling (eds.), The Oxford Handbook of Rationality , Oxford: Oxford University Press, pp. 132–153. doi:10.1093/0195145399.003.0008
- –––, 2005, “How Probabilities Reflect Evidence”, Philosophical Perspectives , 19: 153–179. doi:10.1111/j.1520-8583.2005.00058.x
- Kaplan, Mark, 1996, Decision Theory as Philosophy , Cambridge: Cambridge University Press.
- Kelly, Kevin T., Oliver Schulte, and Cory Juhl, 1997, “Learning Theory and the Philosophy of Science”, Philosophy of Science , 64(2): 245–267. doi:10.1086/392550
- Keynes, John Maynard, 1921, A Treatise on Probability , London: Macmillan and Co.
- Kolmogorov, A.N., 1956, Foundations of the Theory of Probability ( Grundbegriffe der Wahrscheinlichkeitsrechnung , 2 nd edition, New York: Chelsea Publishing Company.
- Koopman, B.O., 1940, “The Bases of Probability”, Bulletin of the American Mathematical Society , 46(10): 763–774. Reprinted in H. Kyburg and H. Smokler (eds.), 1980, Studies in Subjective Probability , 2nd edition, Huntington, NY: Krieger Publ. Co. [ Koopman 1940 available online ]
- Kyburg, Henry E., Jr., 1974, The Logical Foundations of Statistical Inference , Dordrecht: Reidel. doi:10.1007/978-94-010-2175-3
- –––, 1977, “Randomness and the Right Reference Class”, Journal of Philosophy , 74(9): 501–520. doi:10.2307/2025794
- –––, 1978, “An Interpolation Theorem for Inductive Relations”, Journal of Philosophy , 75:93–98.
- –––, 2006, “Belief, Evidence, and Conditioning”, Philosophy of Science , 73(1): 42–65. doi:10.1086/510174
- Lange, Marc, 1999, “Calibration and the Epistemological Role of Bayesian Conditionalization”, Journal of Philosophy , 96(6): 294–324. doi:10.2307/2564680
- –––, 2002, “Okasha on Inductive Scepticism”, The Philosophical Quarterly , 52(207): 226–232. doi:10.1111/1467-9213.00264
- Laudan, Larry, 1997, “How About Bust? Factoring Explanatory Power Back into Theory Evaluation”, Philosophy of Science , 64(2): 206–216. doi:10.1086/392553
- Lenhard Johannes, 2006, “Models and Statistical Inference: The Controversy Between Fisher and Neyman-Pearson”, British Journal for the Philosophy of Science , 57(1): 69–91. doi:10.1093/bjps/axi152
- Levi, Isaac, 1967, Gambling with Truth: An Essay on Induction and the Aims of Science , New York: Knopf.
- –––, 1977, “Direct Inference”, Journal of Philosophy , 74(1): 5–29. doi:10.2307/2025732
- –––, 1978, “Confirmational Conditionalization”, Journal of Philosophy , 75(12): 730–737. doi:10.2307/2025516
- –––, 1980, The Enterprise of Knowledge: An Essay on Knowledge, Credal Probability, and Chance , Cambridge, MA: MIT Press.
- Lewis, David, 1980, “A Subjectivist’s Guide to Objective Chance”, in Richard C. Jeffrey, (ed.), Studies in Inductive Logic and Probability , vol. 2, Berkeley: University of California Press, 263–293.
- Maher, Patrick, 1993, Betting on Theories , Cambridge: Cambridge University Press.
- –––, 1996, “Subjective and Objective Confirmation”, Philosophy of Science , 63(2): 149–174. doi:10.1086/289906
- –––, 1997, “Depragmatized Dutch Book Arguments”, Philosophy of Science , 64(2): 291–305. doi:10.1086/392552
- –––, 1999, “Inductive Logic and the Ravens Paradox”, Philosophy of Science , 66(1): 50–70. doi:10.1086/392676
- –––, 2004, “Probability Captures the Logic of Scientific Confirmation”, in Christopher Hitchcock (ed.), Contemporary Debates in Philosophy of Science , Oxford: Blackwell, 69–93.
- –––, 2005, “Confirmation Theory”, The Encyclopedia of Philosophy , 2nd edition, Donald M. Borchert (ed.), Detroit: Macmillan.
- –––, 2006a, “The Concept of Inductive Probability”, Erkenntnis , 65(2): 185–206. doi:10.1007/s10670-005-5087-5
- –––, 2006b, “A Conception of Inductive Logic”, Philosophy of Science , 73(5): 513–523. doi:10.1086/518321
- –––, 2010, “Bayesian Probability”, Synthese , 172(1): 119–127. doi:10.1007/s11229-009-9471-6
- Mayo, Deborah G., 1996, Error and the Growth of Experimental Knowledge , Chicago: University of Chicago Press.
- –––, 1997, “Duhem’s Problem, the Bayesian Way, and Error Statistics, or ‘What’s Belief Got to do with It?’”, Philosophy of Science , 64(2): 222–244. doi:10.1086/392549
- Mayo Deborah and Aris Spanos, 2006, “Severe Testing as a Basic Concept in a Neyman-Pearson Philosophy of Induction“, British Journal for the Philosophy of Science , 57(2): 323–357. doi:10.1093/bjps/axl003
- McGee, Vann, 1994, “Learning the Impossible”, in E. Eells and B. Skyrms (eds.), Probability and Conditionals: Belief Revision and Rational Decision , New York: Cambridge University Press, 179–200.
- McGrew, Timothy J., 2003, “Confirmation, Heuristics, and Explanatory Reasoning”, British Journal for the Philosophy of Science , 54: 553–567.
- McGrew, Lydia and Timothy McGrew, 2008, “Foundationalism, Probability, and Mutual Support”, Erkenntnis , 68(1): 55–77. doi:10.1007/s10670-007-9062-1
- Neyman, Jerzy and E.S. Pearson, 1967, Joint Statistical Papers , Cambridge: Cambridge University Press.
- Norton, John D., 2003, “A Material Theory of Induction”, Philosophy of Science , 70(4): 647–670. doi:10.1086/378858
- –––, 2007, “Probability Disassembled”, British Journal for the Philosophy of Science , 58(2): 141–171. doi:10.1093/bjps/axm009
- Okasha, Samir, 2001, “What Did Hume Really Show About Induction?”, The Philosophical Quarterly , 51(204): 307–327. doi:10.1111/1467-9213.00231
- Popper, Karl, 1968, The Logic of Scientific Discovery , 3 rd edition, London: Hutchinson.
- Quine, W.V., 1953, “Two Dogmas of Empiricism”, in From a Logical Point of View , New York: Harper Torchbooks. Routledge Encyclopedia of Philosophy, Version 1.0, London: Routledge
- Ramsey, F.P., 1926, “Truth and Probability”, in Foundations of Mathematics and other Essays , R.B. Braithwaite (ed.), Routledge & P. Kegan,1931, 156–198. Reprinted in Studies in Subjective Probability , H. Kyburg and H. Smokler (eds.), 2 nd ed., R.E. Krieger Publishing Company, 1980, 23–52. Reprinted in Philosophical Papers , D.H. Mellor (ed.), Cambridge: University Press, Cambridge, 1990,
- Reichenbach, Hans, 1938, Experience and Prediction: An Analysis of the Foundations and the Structure of Knowledge , Chicago: University of Chicago Press.
- Rényi, Alfred, 1970, Foundations of Probability , San Francisco, CA: Holden-Day.
- Rosenkrantz, R.D., 1981, Foundations and Applications of Inductive Probability , Atascadero, CA: Ridgeview Publishing.
- Roush, Sherrilyn , 2004, “Discussion Note: Positive Relevance Defended”, Philosophy of Science , 71(1): 110–116. doi:10.1086/381416
- –––, 2006, “Induction, Problem of”, Sarkar and Pfeifer 2006..
- –––, 2006, Tracking Truth: Knowledge, Evidence, and Science , Oxford: Oxford University Press.
- Royall, Richard M., 1997, Statistical Evidence: A Likelihood Paradigm , New York: Chapman & Hall/CRC.
- Salmon, Wesley C., 1966, The Foundations of Scientific Inference , Pittsburgh, PA: University of Pittsburgh Press.
- –––, 1975, “Confirmation and Relevance”, in H. Feigl and G. Maxwell (eds.), Induction, Probability, and Confirmation , (Minnesota Studies in the Philosophy of Science, 6), Minneapolis: University of Minnesota Press, 3–36.
- Sarkar, Sahotra and Jessica Pfeifer (eds.), 2006, The Philosophy of Science: An Encyclopedia , 2 volumes, New York: Routledge.
- Savage, Leonard J., 1954, The Foundations of Statistics , John Wiley (2nd ed., New York: Dover 1972).
- Savage, Leonard J., et al., 1962, The Foundations of Statistical Inference , London: Methuen.
- Schlesinger, George N., 1991, The Sweep of Probability , Notre Dame, IN: Notre Dame University Press.
- Seidenfeld, Teddy, 1978, “Direct Inference and Inverse Inference”, Journal of Philosophy , 75(12): 709–730. doi:10.2307/2025515
- –––, 1992, “R.A. Fisher’s Fiducial Argument and Bayes’ Theorem”, Statistical Science , 7(3): 358–368. doi:10.1214/ss/1177011232
- Shafer, Glenn, 1976, A Mathematical Theory of Evidence , Princeton, NJ: Princeton University Press.
- –––, 1990, “Perspectives on the Theory and Practice of Belief Functions”, International Journal of Approximate Reasoning , 4(5–6): 323–362. doi:10.1016/0888-613X(90)90012-Q
- Skyrms, Brian, 1984, Pragmatics and Empiricism , New Haven, CT: Yale University Press.
- –––, 1990, The Dynamics of Rational Deliberation , Cambridge, MA: Harvard University Press.
- –––, 2000, Choice and Chance: An Introduction to Inductive Logic , 4 th edition, Belmont, CA: Wadsworth, Inc.
- Sober, Elliott, 2002, “Bayesianism—Its Scope and Limits”, in Swinburne 2002: 21–38. doi:10.5871/bacad/9780197263419.003.0002
- Spohn, Wolfgang, 1988, “Ordinal Conditional Functions: A Dynamic Theory of Epistemic States”, in William L. Harper and Brian Skyrms (eds.), Causation in Decision, Belief Change, and Statistics , vol. 2, Dordrecht: Reidel, 105–134. doi:10.1007/978-94-009-2865-7_6
- Strevens, Michael, 2004, “Bayesian Confirmation Theory: Inductive Logic, or Mere Inductive Framework?” Synthese , 141(3): 365–379. doi:10.1023/B:SYNT.0000044991.73791.f7
- Suppes, Patrick, 2007, “Where do Bayesian Priors Come From?” Synthese , 156(3): 441–471. doi:10.1007/s11229-006-9133-x
- Swinburne, Richard, 2002, Bayes’ Theorem , Oxford: Oxford University Press. doi:10.5871/bacad/9780197263419.001.0001
- Talbot, W., 2001, “Bayesian Epistemology”, in the Stanford Encyclopedia of Philosophy , (Fall 2001 Edition), Edward N. Zalta (ed.), URL = < https://plato.stanford.edu/archives/fall2001/entries/epistemology-bayesian/ >
- Teller, Paul, 1976, “Conditionalization, Observation, and Change of Preference”, in Harper and Hooker 1976: 205–259. doi:10.1007/978-94-010-1853-1_9
- Van Fraassen, Bas C., 1983, “Calibration: A Frequency Justification for Personal Probability ”, in R.S. Cohen and L. Laudan (eds.), Physics, Philosophy, and Psychoanalysis: Essays in Honor of Adolf Grunbaum , Dordrecht: Reidel. doi:10.1007/978-94-009-7055-7_15
- Venn, John, 1876, The Logic of Chance , 2 nd ed., Macmillan and co; reprinted, New York, 1962.
- Vineberg, Susan, 2006, “Dutch Book Argument”, Sarkar and Pfeifer 2006..
- Vranas, Peter B.M., 2004, “Hempel’s Raven Paradox: A Lacuna in the Standard Bayesian Solution”, British Journal for the Philosophy of Science , 55(3): 545–560. doi:10.1093/bjps/55.3.545
- Weatherson, Brian, 1999, “Begging the Question and Bayesianism”, Studies in History and Philosophy of Science [Part A] , 30(4): 687–697. doi:10.1016/S0039-3681(99)00020-5
- Williamson, Jon, 2007, “Inductive Influence”, British Journal for Philosophy of Science , 58(4): 689–708. doi:10.1093/bjps/axm032
- Zadeh, Lotfi A., 1965, “Fuzzy Sets”, Information and Control , 8(3): 338–353. doi:10.1016/S0019-9958(65)90241-X
- –––, 1978, “Fuzzy Sets as a Basis for a Theory of Possibility”, Fuzzy Sets and Systems , vol. 1, 3–28.
How to cite this entry . Preview the PDF version of this entry at the Friends of the SEP Society . Look up topics and thinkers related to this entry at the Internet Philosophy Ontology Project (InPhO). Enhanced bibliography for this entry at PhilPapers , with links to its database.
- Confirmation and Induction . Really nice overview by Franz Huber in the Internet Encyclopedia of Philosophy .
- Inductive Logic , (in PDF), by Branden Fitelson, Philosophy of Science: An Encyclopedia , (J. Pfeifer and S. Sarkar, eds.), Routledge. An extensive encyclopedia article on inductive logic.
- Teaching Theory of Knowledge: Probability and Induction . A very extensive outline of issues in Probability and Induction, each topic accompanied by a list of relevant books and articles (without links), compiled by Brad Armendt and Martin Curd.
- Probabilistic Confirmation Theory and Bayesian Reasoning . An annotated bibliography of influential works compiled by Timothy McGrew.
- Bayesian Networks Without Tears , (in PDF), by Eugene Charniak (Computer Science and Cognitive Science, Brown University). An introductory article on Bayesian inference.
- Miscellany of Works on Probabilistic Thinking . A collection of on-line articles on Subjective Probability and probabilistic reasoning by Richard Jeffrey and by several other philosophers writing on related issues.
- Fitelson’s course on Confirmation Theory . Main page of Branden Fitelson’s course on Confirmation Theory. The Syllabus provides an extensive list of links to readings. The Notes, Handouts, & Links page has Fitelson’s weekly course notes and some links to useful internet resources on confirmation theory.
- Fitelson’s course on Probability and Induction . Main page of Branden Fitelson’s course on Probability and Induction. The Syllabus provides an extensive list of links to readings on the subject. The Notes & Handouts page has Fitelson’s powerpoint slides for each of his lectures and some links to handouts for the course. The Links page contains links to some useful internet resources.
Bayes’ Theorem | epistemology: Bayesian | probability, interpretations of
Acknowledgments
Thanks to Alan Hájek, Jim Joyce, and Edward Zalta for many valuable comments and suggestions. The editors and author also thank Greg Stokley and Philippe van Basshuysen for carefully reading an earlier version of the entry and identifying a number of typographical errors.
Copyright © 2018 by James Hawthorne < hawthorne @ ou . edu >
Support SEP
Mirror sites.
View this site from another server:
- Info about mirror sites
The Stanford Encyclopedia of Philosophy is copyright © 2021 by The Metaphysics Research Lab , Department of Philosophy, Stanford University
Library of Congress Catalog Data: ISSN 1095-5054
Purdue Online Writing Lab Purdue OWL® College of Liberal Arts
Using Logic

Welcome to the Purdue OWL
This page is brought to you by the OWL at Purdue University. When printing this page, you must include the entire legal notice.
Copyright ©1995-2018 by The Writing Lab & The OWL at Purdue and Purdue University. All rights reserved. This material may not be published, reproduced, broadcast, rewritten, or redistributed without permission. Use of this site constitutes acceptance of our terms and conditions of fair use.
This resource covers using logic within writing—logical vocabulary, logical fallacies, and other types of logos-based reasoning.
Logical Vocabulary
Before using logic to reach conclusions, it is helpful to know some important vocabulary related to logic.
Premise: Proposition used as evidence in an argument. Conclusion: Logical result of the relationship between the premises. Conclusions serve as the thesis of the argument. Argument: The assertion of a conclusion based on logical premises. Syllogism: The simplest sequence of logical premises and conclusions, devised by Aristotle. Enthymeme: A shortened syllogism which omits the first premise, allowing the audience to fill it in. For example, "Socrates is mortal because he is a human" is an enthymeme which leaves out the premise "All humans are mortal." Induction: A process through which the premises provide some basis for the conclusion. Deduction: A process through which the premises provide conclusive proof for the conclusion.
Reaching Logical Conclusions
Reaching logical conclusions depends on the proper analysis of premises. The goal of a syllogism is to arrange premises so that only one true conclusion is possible.
Example A: Consider the following premises:
Premise 1: Non-renewable resources do not exist in infinite supply. Premise 2: Coal is a non-renewable resource.
From these two premises, only one logical conclusion is available:
Conclusion: Coal does not exist in infinite supply.
Example B: Often logic requires several premises to reach a conclusion.
Premise 1: All monkeys are primates. Premise 2: All primates are mammals. Premise 3: All mammals are vertebrate animals. Conclusions: Monkeys are vertebrate animals.
Example C: Logic allows specific conclusions to be drawn from general premises. Consider the following premises:
Premise 1: All squares are rectangles. Premise 2: Figure 1 is a square. Conclusion: Figure 1 is also a rectangle.
Syllogistic Fallacies
The syllogism is a helpful tool for organizing persuasive logical arguments. However, if used carelessly, syllogisms can instill a false sense of confidence in unfounded conclusions. The examples in this section demonstrate how this can happen.
Example D: Logic requires decisive statements in order to work. Therefore, this syllogism is false:
Premise 1: Some quadrilaterals are squares. Premise 2: Figure 1 is a quadrilateral. Conclusion: Figure 1 is a square.
This syllogism is false because not enough information is provided to allow a verifiable conclusion. Figure 1 could just as likely be a rectangle, which is also a quadrilateral.
Example E: Logic can also mislead when it is based on premises that an audience does not accept. For instance:
Premise 1: People with red hair are not good at checkers. Premise 2: Bill has red hair. Conclusion: Bill is not good at checkers.
Within the syllogism, the conclusion is logically valid. However, the syllogism itself is only true if an audience accepts Premise 1, which is very unlikely. This is an example of how logical statements can appear accurate while being completely false.
Example F: Logical conclusions also depend on which factors are recognized and ignored by the premises. Therefore, premises that are correct but that ignore other pertinent information can lead to incorrect conclusions.
Premise 1: All birds lay eggs. Premise 2: Platypuses lay eggs. Conclusion: Platypuses are birds.
It is true that all birds lay eggs. However, it is also true that some animals that are not birds lay eggs. These include fish, amphibians, reptiles, and a small number of mammals (like the platypus and echidna). To put this another way: laying eggs is not a defining characteristic of birds. Thus, the syllogism, which assumes that because all birds lay eggs, only birds lay eggs, produces an incorrect conclusion.
A better syllogism might look like this:
Premise 1: All mammals have fur. Premise 2: Platypuses have fur. Conclusion: Platypuses are mammals.
Fur is indeed one of the defining characteristics of mammals —in other words, there are not non-mammal animals who also have fur. Thus, the conclusion here is more firmly-supported.
In sum, though logic is a very powerful argumentative tool and is far preferable to a disorganized argument, logic does have limitations. It must also be effectively developed from a syllogism into a written piece.
- Give this article
Advertisement
Supported by
Justice Dept. Investigated Clinton Foundation Until Trump’s Final Days
President Donald J. Trump and his allies tried to cast the Clinton Foundation as corrupt. But the yearslong investigation sputtered to its conclusion without charges.
Send any friend a story
As a subscriber, you have 10 gift articles to give each month. Anyone can read what you share.

By Adam Goldman
Reporting from Washington
The Justice Department kept open the investigation into Hillary Clinton’s family foundation for nearly all of President Donald J. Trump’s administration, with prosecutors closing the case without charges just days before he left office.
Newly released documents and interviews with former department officials show that the investigation stretched long past when F.B.I. agents and prosecutors knew it was a dead end. The conclusion of the case, which centered on the Clinton Foundation’s dealings with foreign donors when Mrs. Clinton served as secretary of state under President Barack Obama, has not previously been reported.
Mr. Trump, who campaigned on a promise to “lock her up,” spent much of his four-year term pressuring the F.B.I. and the Justice Department to target political rivals. After being accused by the president’s allies of serving as part of a deep-state cabal working against him, F.B.I. officials insisted that the department acknowledge in writing that there was no case to bring.
The closing documents, which were obtained by The New York Times as part of a Freedom of Information Act lawsuit, spelled the end to an investigation that top prosecutors had expressed doubts about from the beginning. Still, it became a rallying cry for Republicans who believed the F.B.I. would ultimately turn up evidence of corruption and damage Mrs. Clinton’s political fortunes.
The foundation became attack fodder for Republicans in 2015 after the conservative author Peter Schweizer published the book “ Clinton Cash: The Untold Story of How and Why Foreign Governments and Businesses Helped Make Bill and Hillary Rich ,” an investigation of donations that foreign entities made to the family organization. Mr. Schweizer is the president of the Government Accountability Institute, where Stephen K. Bannon, Mr. Trump’s former chief strategist, was a founder and the executive chairman.
A spokesman for the Clinton Foundation, Craig Minassian, said that the organization had been “subjected to politically motivated allegations with no basis in fact.”
Republicans seized on the accusations in Mr. Schweizer’s book, accusing Mrs. Clinton of supporting the interests of foundation donors as part of a quid pro quo.
Specifically, critics focused on the foundation’s receipt of large donations in exchange for supporting the sale of Uranium One , a Canadian company with ties to mining stakes in the United States, to a Russian nuclear agency. The deal was approved in 2010 by the Committee on Foreign Investment in the United States when Mrs. Clinton, as secretary of state, had a voting seat on the panel.
Mr. Schweizer’s research caught the eye of F.B.I. agents in Washington, who in 2016 opened a preliminary investigation based solely on “unvetted hearsay information” in the book, according to the final report by John H. Durham , the Trump-era special counsel who led an investigation into the bureau’s inquiry into possible ties between Mr. Trump’s campaign and Russia.
The F.B.I. in New York and Little Rock, Ark., also opened investigations that relied on information from confidential source reporting, according to Mr. Durham.
Mr. Durham also compared the handling of the Clinton Foundation investigation to the F.B.I.’s treatment of the Russia investigation. As part of his inquiry, Mr. Durham questioned Mrs. Clinton last spring. “Secretary Clinton was voluntarily interviewed by Special Counsel Durham on May 11, 2022,” said David E. Kendall, her lawyer. “No topics were off limits. She answered every question.”
The Justice Department did not think much of the foundation investigations, frustrating F.B.I. agents. Raymond N. Hulser, a prosecutor in charge of the public integrity section at the time, told Mr. Durham that the Washington case that was based on the book lacked predication.
Indeed, some prosecutors at the time believed the book had been discredited.
The investigation became a source of friction at the F.B.I. as agents believed the Justice Department had stymied their work.
That tension spilled into public view and had far-reaching consequences.
Andrew G. McCabe, then the F.B.I.’s deputy director, was accused of leaking information about the case to a Wall Street Journal reporter and later lying about it to the Justice Department’s inspector general. The episode helped prompt his dismissal in 2018 and a failed effort by the department to prosecute him.
In August 2016, the three foundation cases were consolidated under the supervision of agents in New York. Agents were authorized to seek subpoenas from the U.S. attorneys’ offices in Manhattan and Brooklyn, but prosecutors declined to issue them. The investigation seemed to go dormant.
Ultimately, the F.B.I. moved the case to Little Rock. In 2017, after prosecutors there requested help, the deputy attorney general’s office said the Justice Department would support the case.
Eventually, prosecutors secured a subpoena for the charity in early 2018 and the F.B.I. detailed personnel to examine donor records. Investigators also interviewed the former chief financial officer for the foundation.
Career prosecutors in Little Rock then closed the case, notifying the F.B.I.’s office there in two letters in January 2021. But in a toxic atmosphere in which Mr. Trump had long accused the F.B.I. of bias, the top agent in Little Rock wanted it known that career prosecutors, not F.B.I. officials, were behind the decision.
In August 2021, the F.B.I. received what is known as a declination memo from prosecutors and as a result considered the matter closed.
“All of the evidence obtained during the course of this investigation has been returned or otherwise destroyed,” according to the F.B.I.
Jo Becker contributed reporting.
Adam Goldman reports on the F.B.I. and national security from Washington, D.C., and is a two-time Pulitzer Prize winner. He is the coauthor of “Enemies Within: Inside the NYPD's Secret Spying Unit and bin Laden's Final Plot Against America.” @ adamgoldmanNYT
- Latest Headlines
- English Edition Edition English 中文 (Chinese) 日本語 (Japanese)
- Print Edition
- More More Other Products from WSJ Buy Side from WSJ WSJ Shop WSJwine
This copy is for your personal, non-commercial use only. Distribution and use of this material are governed by our Subscriber Agreement and by copyright law. For non-personal use or to order multiple copies, please contact Dow Jones Reprints at 1-800-843-0008 or visit www.djreprints.com.
https://www.wsj.com/articles/john-durham-report-special-counsel-fbi-donald-trump-hillary-clinton-russia-peter-strzok-christopher-steele-d4a01d4e
Why the Durham Report Matters to Democracy
It is a damning account of the corruption of the fbi and its accomplices..
Copyright © 2023 Dow Jones & Company, Inc. All Rights Reserved. 87990cbe856818d5eddac44c7b1cdeb8
- TurboTax : Save up to $15 with TurboTax coupon May 2023
- The Motley Fool : Epic Bundle - 3x Expert Stock Recommendations
- H&R Block Tax : 15% Off DIY Online Tax Filing Services | H&R Block Coupon
- Top Resume : 10% TopResume Discount Code for expert resume-writing services
- eBay : 30% off eBay coupon
- Groupon : Exclusive: 30% OFF Groupon Promo Code
Most Popular news
That $6,000 vacation isn’t worth it right now.
Wagner Chief’s Feud With Russian Military Cracks Putin’s Image of Control
Why Tipping Prompts Are Suddenly Everywhere
Debt-Ceiling Fight Comes Down to Spending: Freeze or Cut?
Target to Pull Some LGBT-Themed Merchandise After Backlash
Most Popular opinion
Opinion: inflation has peaked—get ready for deflation.
Opinion: These Aren’t Tommy Lasorda’s Dodgers
Opinion: A Lawsuit to Protect Pensions From Climate Politics
Opinion: Will Illinois Still ‘Invest in Kids’?
Opinion: Stupid Debt-Ceiling Tricks
Recommended Videos
Copyright © 2023 Dow Jones & Company, Inc. All Rights Reserved
- International edition
- Australia edition

Inflation is on the wane, but a longed-for sharp drop is far from in the bag
Energy bills and a tight labour market continue to cloud predictions of a steep decline by the end of the year
To tackle inflation , the Bank of England is keeping a close eye on the UK’s tight jobs market, ultra-high energy prices and the crippling cost of food .
These are the three main factors that have sent inflation rocketing over the last 18 months. The question is, has Britain reached a turning point and how quickly will inflation fall?
This week, official figures, to be announced on Wednesday, are expected to show the overall rate of inflation, excluding housing costs, falling to single figures for the first time in seven months. Forecasters expect April’s rate will have dropped by as much as two percentage points from the 10.1% reported in March, reflecting a sharp fall in energy prices and an easing of pressures in the labour market.
The consumer prices index (CPI) measure of inflation is forecast to fall further this year as 12 consecutive interest-rate rises by the central bank take their toll on the disposable incomes of households and businesses.
However, predictions made at the beginning of the year that inflation would slump to 3% by Christmas have been torn up in recent weeks after food continued to become more expensive.
In its most recent economic outlook, the Bank of England said inflation would remain stickier than its officials had estimated back in February, prompting an increase in the forecast for CPI from 3.25% in the last three months of the year to 5.1%.
In another blow to hopes of a sharp slide in inflation, the consultancy Oxford Economics said the easing of the labour market could prove to be shortlived.
In the most recent figures covering the three months to the end of April, job vacancies fell by 55,000 to just over 1m, the 10th consecutive quarterly drop. The number of people of working age not in the labour market and considered “inactive” rather than unemployed decreased by 0.4 percentage points to 21% in the three months to March from the previous quarter.
The consultancy’s chief UK economist, Andrew Goodwin, said much of the fall in inactivity was due to students working part time, which was not going to close the skills gap in sectors screaming for highly trained staff. And anyway, it was a trend that had most likely already run out of steam, he added.
Goodwin said the composition of those inactive was also changing, after April’s figures showed the number of people unable to work due to long-term sickness had reached a fresh record of 2.55 million .
Ministers seem to be underestimating the health shock to the UK workforce, which many experts believe is not going away any time soon.
Goodwin said the Office for National Statistics figures were questionable after a steep decline in the number of people and businesses responding to ONS surveys since the pandemic hit. “But if the data is accurate, it suggests that it might be hard to sustain the recent loosening in labour market conditions, particularly if demand for labour holds up,” he said.
Another factor preventing a steep inflation drop is the likelihood of energy prices remaining elevated for the rest of the year. About 27 million domestic energy customers are braced for the latest Ofgem price cap, to be announced on Thursday, which will set the price of energy bills from 1 July 2023.
Predictions are that Ofgem will cap average household energy bills at about £2,053 a year from 1 July, a reduction from the current average £2,500 under the government’s energy price guarantee.
However, analysts at Cornwall Insight said that rather than energy prices falling back to 2019 levels, future price caps will set average energy bills at £2,098 from 1 October and £2,163 from 1 January 2024, which will effectively halt the much-hoped-for slump in energy bills in its tracks.
The analysts said that while the price of gas had collapsed on world markets, long-term contracts signed by energy providers mean the savings will take time to filter through, leaving prices at double where they were before the pandemic struck.
According to the End Fuel Poverty Coalition, a campaign group, prices will be 62% above the level seen before the invasion of Ukraine.
Inflation may be coming down, but possibly not by much this year, and prices will still be high, especially for those on low incomes who spend more on energy and food as a proportion of their income.
- Observer business agenda
- Energy bills
- UK cost of living crisis
- Household bills
- Work & careers
- Interest rates

IMF’s anti-tax cut stance may give Rishi Sunak cover to resist Tory right

Bank of England says inflation has ‘turned corner’ despite soaring food prices

From CPI to stagflation: how the UK tracks price rises and what key inflation terms mean

IMF says UK no longer heading for recession but urges against tax cuts

UK food price inflation at third-highest level since 2008

China’s war chest: how the fight for semiconductors reveals the outlines of a future conflict

Sunak to urge G7 support for collective action against ‘economic coercion’

Central banks ‘risk losing trust if they cannot tame inflation’
Most viewed.

IMAGES
VIDEO
COMMENTS
Validity and soundness. Validity and soundness are two criteria for assessing deductive reasoning arguments. Validity. In this context, validity is about the way the premises relate to each other and the conclusion. This is a different concept from research validity.. An argument is valid if the premises logically support and relate to the conclusion.
Deductive reasoning, also known as deduction, is a basic form of reasoning. It starts out with a general statement, ... Conclusion: Humans have backbones. Major premise: All birds lay eggs.
Deductive Reasoning. Deduction is generally defined as "the deriving of a conclusion by reasoning." Its specific meaning in logic is "inference in which the conclusion about particulars follows necessarily from general or universal premises."Simply put, deduction—or the process of deducing—is the formation of a conclusion based on generally accepted statements or facts.
deduction: [noun] an act of taking away. something that is or may be subtracted.
Both deduction and induction are used to prove hypotheses and support the scientific method. Deduction requires us to look at how closely a premises and the conclusion are related. If the premises are backed by evidence and experiment, then the conclusion is more likely to be true. In the scientific method, scientists form a hypothesis.
Deductive reasoning gives you a certain and conclusive answer to your original question or theory. A deductive argument is only valid if the premises are true. And the arguments are sound when the conclusion, following those valid arguments, is true. To me, this sounds a bit more like the scientific method.
deduction, in logic, a rigorous proof, or derivation, of one statement (the conclusion) from one or more statements (the premises)—i.e., a chain of statements, each of which is either a premise or a consequence of a statement occurring earlier in the proof. This usage is a generalization of what the Greek philosopher Aristotle called the syllogism, but a syllogism is now recognized as merely ...
In lang=en terms the difference between conclusion and deduction. is that conclusion is an estoppel or bar by which a person is held to a particular position while deduction is a process of reasoning that moves from the general to the specific, in which a conclusion follows necessarily from the premises presented, so that the conclusion cannot ...
Deduction: A deduction is any item or expenditure subtracted from gross income to reduce the amount of income subject to income tax . It is also referred to as an "allowable deduction." For ...
In reading comprehension, deduction is the act of drawing logical conclusions based on the information given in a text, using one's personal experiences and knowledge of the world. Deduction is often taught alongside inference, which is defined as any interpretation of the text that goes beyond the literal information given.
a conclusion is a final understanding and reasoning behind the action or problem. once you find the conclusion, the reason behind, it you can fix the problem. deduction is a process used to find a conclusion and answer. for example scientists use different ways to deduct other possible reasons behind something before they can find the correct conclusion.
⚡ Quick summary. Inductive reasoning (also called induction) involves forming general theories from specific observations.Observing something happen repeatedly and concluding that it will happen again in the same way is an example of inductive reasoning.Deductive reasoning (also called deduction) involves forming specific conclusions from general premises, as in: everyone in this class is an ...
The conclusion of this argument might seem to follow with certainty, but additional evidence can be added to increase the probability of the truth of the conclusion. ... is termed Deduction. In its due place, it is a highly important part of every science; but it has no value when the fundamental principles, on which the whole of the ...
Deductive reasoning is the mental process of drawing deductive inferences. An inference is deductively valid if its conclusion follows logically from its premises, i.e. if it is impossible for the premises to be true and the conclusion to be false. For example, the inference from the premises "all men are mortal" and " Socrates is a man" to the ...
Deductive and Inductive Arguments. In philosophy, an argument consists of a set of statements called premises that serve as grounds for affirming another statement called the conclusion. Philosophers typically distinguish arguments in natural languages (such as English) into two fundamentally different types: deductive and inductive.Each type of argument is said to have characteristics that ...
Induction vs. Deduction In writing, argument is used in an attempt to convince the reader of the truth or falsity of some proposal or thesis. Two of the methods used are induction and deduction. Induction: A process of reasoning (arguing) which infers a general conclusion based on individual cases, examples, specific bits of evidence, and other specific
Deduction is inference deriving logical conclusions from premises known or assumed to be true, with the laws of valid inference being studied in logic. Induction is inference from particular premises to a universal conclusion. Abduction is inference to the best explanation." - Mauro ALLEGRANZA.
The conclusions of deductive reasoning can only be true if all the premises set in the inductive study are true and the terms are clear. Example. All dogs have fleas (premise) Benno is a dog (premise) Benno has fleas (conclusion) Based on the premises we have, the conclusion must be true. However, if the first premise turns out to be false, the ...
Inferences can be deductive, inductive, or abductive. Deductive inferences are the strongest because they can guarantee the truth of their conclusions. Inductive inferences are the most widely used, but they do not guarantee the truth and instead deliver conclusions that are probably true. Abductive inferences also deal in probability.
An inductive logic is a logic of evidential support. In a deductive logic, the premises of a valid deductive argument logically entail the conclusion, where logical entailment means that every logically possible state of affairs that makes the premises true must make the conclusion true as well. Thus, the premises of a valid deductive argument provide total support for the conclusion.
Deduction: A process through which the premises provide conclusive proof for the conclusion. Reaching Logical Conclusions. Reaching logical conclusions depends on the proper analysis of premises. The goal of a syllogism is to arrange premises so that only one true conclusion is possible. Example A: Consider the following premises:
The conclusion of the case, which centered on the Clinton Foundation's dealings with foreign donors when Mrs. Clinton served as secretary of state under President Barack Obama, has not ...
"The franchise tag for linebackers includes some edge rushers and is projected to come in at more than $20 million for 2024, which would be more than any off-ball linebacker in football is getting.
News Corp is a global, diversified media and information services company focused on creating and distributing authoritative and engaging content and other products and services.
This week, official figures, to be announced on Wednesday, are expected to show the overall rate of inflation, excluding housing costs, falling to single figures for the first time in seven months.
"I think everyone takes points off one another, generally, so if you go down, it's not a foregone conclusion you're going to come straight back up, regardless of who you are. It is a very ...