Likelihood function
The likelihood of a set of parameter values, θ, given outcomes x, is equal to the probability of those observed outcomes given those parameter values, that is
- .
Discrete probability distribution[edit]
Let X be a random variable with a discrete probability distribution p depending on a parameter θ. Then the functionconsidered as a function of θ, is called the likelihood function (of θ, given the outcome x of X). Sometimes the probability on the value x of X for the parameter value θ is written as; often written as to emphasize that this differs from which is not a conditional probability, because θ is a parameter and not a random variable.Continuous probability distribution[edit]
Let X be a random variable with a continuous probability distribution with density function f depending on a parameter θ. Then the functionconsidered as a function of θ, is called the likelihood function (of θ, given the outcome x of X). Sometimes the density function for the value x of X for the parameter value θ is written as , but should not be confused with which should not be considered a conditional probability density.- The only reason to use the log-likelihood instead of the plain old likelihood is mathematical convenience, because it lets you turn multiplication into addition. The plain old likelihood is P(parameters | data), i.e. assuming your data is fixed and you vary the parameters of your model. Maximizing this is one way to do parameter estimation and is known as maximum likelihood.
No comments:
Post a Comment