Likelihood
Definition: Likelihood and Probability
Consider a Statistical Model parameterized by a parameter and an observation of samples . We have, the probability of observing under parameter is . For the same value, we define it as the likelihood of the true parameter being given observation :
- 💡 That is, the likelihood of a parameter is the probability of observing the data given that parameter. Note that the likelihood is a function of the parameter .
For continuous Random Variables, sometimes (we will see later) it is more convenient to define the likelihood using the observation’s Probability Density Function:
If is the true value of the parameter, should be large. That’s why we call the likelihood, and we usually want to maximize the likelihood. Since mass-density transformation does not affect the maximization operation, the likelihood is usually defined using the density function.
Log-Likelihood
For iid samples , the likelihood function is of the form . Thus, it is usually more convenient to deal with the log of the likelihood function, turning the product into a sum:
Since the log function is monotonically increasing, this transformation again does not affect the maximization operation.
Score Function
The derivation of the log-likelihood function is called the score function: