Likelihood function

Written by So8res, et al. last updated
Requires: Likelihood, Math 2
Teaches: Likelihood

Let's say you have a piece of evidence and a set of hypotheses Each assigns some likelihood to The function that reports this likelihood for each is known as a "likelihood function."

For example, let's say that the evidence is = "Mr. Boddy was killed with a candlestick," and the hypotheses are = "Miss Scarlett did it," = "Colonel Mustard did it," and = "Mrs. Peacock did it." Furthermore, if Miss Scarlett was the murderer, she's 20% likely to have used a candlestick. By contrast, if Colonel Mustard did it, he's 10% likely to have used a candlestick, and if Mrs. Peacock did it, she's only 1% likely to have used a candlestick. In this case, the likelihood function is

For an example using a continuous function, consider a possibly-biased coin whose bias to come up heads on any particular coinflip might be anywhere between and . Suppose we observe the coin to come up heads, tails, and tails. We will denote this evidence The likelihood function over each hypothesis = "the coin is biased to come up heads portion of the time" for is:

There's no reason to normalize likelihood functions so that they sum to 1 — they aren't probability distributions, they're functions expressing each hypothesis' propensity to yield the observed evidence. For example, if the evidence was really obvious ( = "the sun rose this morning,") it might be the case that almost all hypotheses have a very high likelihood, in which case the sum of the likelihood function will be much more than 1.

Likelihood functions carry absolute likelihood information, and therefore, they contain information that relative likelihoods do not. Namely, absolute likelihoods can be used to check a hypothesis for strict confusion.

Parents: