Generative vs discriminative model

  • Context: Graduate 
  • Thread starter Thread starter pamparana
  • Start date Start date
  • Tags Tags
    Model
Click For Summary
SUMMARY

This discussion centers on the differences between generative and discriminative models in machine learning. Generative models aim to model the joint distribution p(x,y), while discriminative models focus on estimating the conditional distribution p(y|x). The conversation highlights the complexity of generative models and the misconception that discriminative models require estimating the joint distribution. Additionally, the role of Markov random fields in representing joint distributions is addressed, emphasizing the nuances in understanding these modeling approaches.

PREREQUISITES
  • Understanding of joint and conditional distributions in probability theory
  • Familiarity with Bayes' theorem and its applications
  • Knowledge of generative models and discriminative models in machine learning
  • Basic concepts of Markov random fields and their representation of distributions
NEXT STEPS
  • Study the differences between generative and discriminative models in detail
  • Explore the application of Bayes' theorem in machine learning contexts
  • Read the paper on the advantages of learning P(x,y) as an intermediate step in modeling P(y|x) available at http://ai.stanford.edu/~ang/papers/nips01-discriminativegenerative.pdf
  • Investigate the properties and applications of Markov random fields in probabilistic modeling
USEFUL FOR

Machine learning practitioners, data scientists, and researchers interested in understanding the theoretical foundations and practical applications of generative and discriminative models.

pamparana
Messages
123
Reaction score
0
Hello everyone,

I have a question about generative and discriminative models. As I understand it generative models aim to model the joint distribution p(x,y) of the input x and output y and the discriminative approach estimates the conditional distribution p(y|x). I understand that the generative approach is more complex but I have a small niggle...

Applying Bayes rule, we have the conditional distribution

p(y|x) = p(x|y) p(y)/p(x)

which is equal to:

P(y|x) = p(x, y)/p(x)

So it seems when we want to model discriminative modelling, we have to estimate this joint distribution (in he numerator) as well. I am sure this is wrong and I have a flaw in my understanding but I cannot seem to figure it out. Is there something about this modelling process that I am not understanding. Perhaps we do not estimate it in this way?

Also, a markov random field defined over an undirected graph, is it a joint distribution?

I would really appreciate any help anyone can give me on this.

Thanks,
Luc
 
Physics news on Phys.org
P(x|y) has less information than P(x,y), because you can get the former from the latter, but not the latter from the former. For certain data sets, one can estimate parameters for a model of P(x|y) without the intermediate step of modelling P(x,y).

There is an interesting discussion on the advantages and disadvantages of explicitly learning P(x,y) as an intermediate step in modelling P(x|y) in http://ai.stanford.edu/~ang/papers/nips01-discriminativegenerative.pdf.
 
Last edited:

Similar threads

  • · Replies 30 ·
2
Replies
30
Views
5K
Replies
3
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 23 ·
Replies
23
Views
4K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 3 ·
Replies
3
Views
1K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K