Understanding Maximum Likelihood Estimation: Unpacking the Basics

In summary, the likelihood function determines the plausibility of parameters given the observed random variables. The observed random variables themselves must be generated from a probability distribution, but the logic becomes circular if the observed random variables are generated from a probability distribution and then the likelihood of the distribution parameters is determined based on the observed random variables.
  • #1
FallenApple
566
61
I'm getting a bit lost on some of the basics. So a Likelihood function determines the plausibility of parameters given the observed random variables. This is fine and all, but something seems a bit off. The observed random variables themselves must be generated from a probability distribution as well. So the logic becomes circular. Is there something I'm not seeing?
 
Physics news on Phys.org
  • #2
FallenApple said:
I'm getting a bit lost on some of the basics. So a Likelihood function determines the plausibility of parameters given the observed random variables. This is fine and all, but something seems a bit off. The observed random variables themselves must be generated from a probability distribution as well. So the logic becomes circular. Is there something I'm not seeing?
Random variables can be generated from probability distributions, but also from physical processes. If you first generate values of random variables from a probability distribution and then find likelihoods of distribution parameters based on those values, then yes, you have created a circular process. It is not circular, though, if you measure some physical process and use likelihood functions to help construct a mathematical model of the process.
 
  • Like
Likes FallenApple
  • #3
tnich said:
Random variables can be generated from probability distributions, but also from physical processes. If you first generate values of random variables from a probability distribution and then find likelihoods of distribution parameters based on those values, then yes, you have created a circular process. It is not circular, though, if you measure some physical process and use likelihood functions to help construct a mathematical model of the process.
Thanks, that really cleared up all of the confusion.
 
  • #4
FallenApple said:
So the logic becomes circular. Is there something I'm not seeing?

It isn't clear what line of reasoning you're thinking about when you say "the logic".

So a Likelihood function determines the plausibility of parameters given the observed random variables.
What is your definition of "plausibility"? The likihood function does not determine the "probability" of the parameters given the observed random variables - if that's what you're thinking. It also does not determine the "liklihood" of the parameters. It's better to think of the liklhood function as giving the liklihood of the data for given values of the parmeters - as opposed to the liklihood of the parameters for given values of the data.

If we are considering a family of probability distributions and each member of the family is specified by giving specific values to some parameters then the liklihood function gives the "liklihood" of the data as a function of the parameters and the data. The phrase "liklihood of the data" is used instead of "probability of the data" because it is incorrect to say that evaluating a probability density function produces a probability. Evaluating a probability density function, in the case of a continuous distribution, gives a "probability density". not a "probability". For example, the probability density of a random variable U that is uniformly distributed on [0,1] is the constant function f(x) = 1. The fact that f(1/3) = 1 does imply that the probability that the value 1/3 occurs is 1. "Liklihood of" is a way to say "probability density of".

One procedure for estimating parameters from given values of data is to use the values of the parameters that maximize the value of the liklihood function. It should be emphasized that (like many things in statistics - e.g. hypothesis testing) this is a procedure - i.e. merely one procedure out of several possible procedures, not a technique that can be proven to be the unique optimal way to do things. If your remark about "the logic becomes circular" indicates skeptism about a proof that maximum liklihood estimation is optimal, your skeptism is correct. However, if you are studying a respectable textbook, I doubt the textbook says that the Maximuj Liklihood estimation procedure is an optimal way to estimate parameters in all cases. There can be theorems along those lines ,but they deal with specific cases - and they have to define the specific function we are trying to optimize.
 
  • Like
Likes FallenApple
  • #5
Good question. The short answer is that there is no circular logic because there is no hard "logic" at all that applies.

The maximum likelihood estimator allows you to determine the model parameter that makes the given data most likely. There is no formal "logic" that will say that the parameter is correct. So you are wise to be cautious. If you also obtain a confidence interval for the parameter, you can see how unusual it would be to get that data if the true parameter was, in fact, outside of that interval. Even then, you would have no hard logic to say whether it is in or out of the confidence interval -- only hypothetical probabilities.
 
  • Like
Likes FallenApple

1. What is a likelihood function?

A likelihood function is a statistical tool used to determine the probability of obtaining a certain set of data based on a specific statistical model. It is often used in hypothesis testing and parameter estimation.

2. How does a likelihood function work?

A likelihood function takes in a set of data and a statistical model and calculates the probability of obtaining that data based on the model. It uses a mathematical formula to compare the observed data to the expected data and produces a value that represents the likelihood of the model being true.

3. What is the difference between a likelihood function and a probability distribution?

A likelihood function is a function of the parameters of a statistical model, whereas a probability distribution is a function of the possible outcomes of a random variable. In other words, a likelihood function describes the relationship between the data and the parameters of the model, while a probability distribution describes the relationship between the data and the possible outcomes.

4. How is a likelihood function used in parameter estimation?

In parameter estimation, a likelihood function is used to find the values of the model parameters that make the observed data most likely to occur. This is done by maximizing the likelihood function, which means finding the parameter values that produce the highest likelihood value.

5. Can a likelihood function be used for any type of model?

Yes, a likelihood function can be used for any type of statistical model, including linear regression, logistic regression, and Gaussian distributions. However, the specific form of the likelihood function will vary depending on the model being used.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
6
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
11
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
8
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
957
  • Set Theory, Logic, Probability, Statistics
Replies
7
Views
458
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
2K
Back
Top