Likelihood

Definition: Likelihood and Probability

Consider a Statistical Model parameterized by a parameter and an observation of samples . We have, the probability of observing under parameter is . For the same value, we define it as the likelihood of the true parameter being given observation :

  • 💡 That is, the likelihood of a parameter is the probability of observing the data given that parameter. Note that the likelihood is a function of the parameter .

For continuous Random Variables, sometimes (we will see later) it is more convenient to define the likelihood using the observation’s Probability Density Function:

If is the true value of the parameter, should be large. That’s why we call the likelihood, and we usually want to maximize the likelihood. Since mass-density transformation does not affect the maximization operation, the likelihood is usually defined using the density function.

Log-Likelihood

For iid samples , the likelihood function is of the form . Thus, it is usually more convenient to deal with the log of the likelihood function, turning the product into a sum:

Since the log function is monotonically increasing, this transformation again does not affect the maximization operation.

Score Function

The derivation of the log-likelihood function is called the score function: