There are more complicated definitions, like ...
The density function of continuous random variables, , that depend on model parameters, , and written as, is the probability density function of , given, or is the likelihood of , given , often written as .
(Which definition do you prefer? Me too.)
Figure 1 - Some values of are more likely than others.
We engineers recognize the figure above as a familiar "probability density function," and see it as a function of the random variable, x . For a given x, the ordinate, y, is the probability that x can take on that value.
Well, that's not quite right. As it turns out the probability that x can take on EXACTLY some value x0 is precisely zero! For example, if x0 is exactly 1, then the value of x=1.00000000000000001 is excluded. You can see that a vanishingly small difference is still a difference and so the probability of exactly some value must be zero.
Wait! Don't panic! The probability that x is within
some non-zero distance of x0 is the ordinate, y,
times the interval of x that we say is close enough,
So the probability of x being within that interval is the integral of
the probability curve from
x-(x /2), to x+(x /2). (You can also see that if x is zero then the product of zero and the ordinate, y, is zero too.)
So what? There is another way to interpret the figure. Rather than consider x as unknown and y as the probability that x is within some interval, we could also consider x as known, and y as the probability (or "likelihood") that we put the curve in the right place, i.e., that we have the right value for , the distribution's location parameter, and that we have the right shape parameter,, too.
The figure above compares the likelihood values for =1.5 (wide vertical lines) and =3.2 (narrow lines) at x=0.6, 0.8, 1.0, 1.2, and 1.4. In this example, a value of 3.2 for the Weibull shape parameter,, is more likely than a value of 1.5, given the known values of x.
We would multiply the likelihoods for each x observation to compute an overall likelihood for . Since multiplication of likelihoods can be messy, in practice we sum the logs of the likelihoods. (The maximum loglikelihood will occur at the same value for as the maximum for the likelihood itself because the logarithm is a monotonic function.)
This is the maximum likelihood criterion. It says that we should choose values for the Weibull model parameters, and , that maximize the likelihood (probability) that the experiment turned out the way that it did. Although maximum likelihood estimators (MLEs) are sometimes biased, they often more than make up for that by having smaller variability, and thus are superior to other methods for estimating model parameters. (More on that topic here.)
The statistical literature often uses "likelihood" and "loglikelihood" interchangeably, which can be confusing to the statistical newcomer, but in practice it is rather easy to distinguish the two based on context.