Statlect: Log-Likelihood
Statlect: Log-Likelihood
StatLect
Index > Glossary
Log-likelihood
by Marco Taboga, PhD
The log-likelihood is, as the term suggests, the natural logarithm of the likelihood.
Definition
The following elements are needed to rigorously define the log-likelihood function:
https://www.statlect.com/glossary/log-likelihood 1/6
7/6/2020 Log-likelihood
when the joint probability mass (or density) function is considered as a function of
for fixed (i.e., for the sample we have observed), it is called likelihood (or
likelihood function) and it is denoted by . So,
if is discrete and
if is continuous.
Given all these elements, the log-likelihood function is the function defined by
Example
whose entries are draws from a normal distribution. The probability density
function of a generic draw is
where and are the parameters (mean and variance) of the normal distribution.
With the notation used in the previous section, the parameter vector is
The parametric family being considered is the set of all normal distributions (that can
be obtained by varying the parameters and ).
https://www.statlect.com/glossary/log-likelihood 2/6
7/6/2020 Log-likelihood
In order to stress the fact that the probability density depends on the two parameters,
we write
because the joint density of a set of independent variables is equal to the product of
their marginal densities (see the lecture on Independent random variables).
https://www.statlect.com/glossary/log-likelihood 3/6
7/6/2020 Log-likelihood
that is, by finding the parameter that maximizes the log-likelihood of the observed
sample . This is the same as maximizing the likelihood function because the
natural logarithm is a strictly increasing function.
One may wonder why the log of the likelihood function is taken. There are several
good reasons. To understand them, suppose that the sample is made up of
independent observations (as in the example above). Then, the logarithm transforms
a product of densities into a sum. This is very convenient because:
the asymptotic properties of sums are easier to analyze (one can apply Laws of
Large Numbers and Central Limit Theorems to these sums; see the proofs of
consistency and asymptotic normality of the maximum likelihood estimator);
products are not numerically stable: they tend to converge quickly to zero or to
infinity, depending on whether the densities of the single observations are on
average less than or greater than 1; sums are instead more stable from a
numerical standpoint; this is important because the maximum likelihood problem is
often solved numerically on computers where limited machine precision does not
allow to distinguish a very small number from zero and a very large number from
infinity.
More examples
More example of how to derive log-likelihood functions can be found in the lectures
on:
https://www.statlect.com/glossary/log-likelihood 4/6
7/6/2020 Log-likelihood
More details
The log-likelihood and its properties are discussed in a more detailed manner in the
lecture on maximum likelihood estimation.
https://www.statlect.com/glossary/log-likelihood 5/6
7/6/2020 Log-likelihood
The book
Most of the learning materials found on this website are now available in a traditional
textbook format.
Learn more
Explore About
Set estimation About Statlect
Convergence in distribution Contacts
Independent events Cookies, privacy and terms of use
Glossary entries
Discrete random variable
Almost sure
Type I error
Null hypothesis
Binomial coefficient
Type II error
Share
https://www.statlect.com/glossary/log-likelihood 6/6