How should a physicist learn Machine Learning?

AI Thread Summary
A master’s student working on physics and neural networks seeks to deepen their understanding of neural networks while using TensorFlow and Keras for their thesis. They express a need for accessible resources that provide foundational knowledge without overwhelming detail, as they cannot commit to lengthy texts. Suggestions include introductory materials that foster intuition about neural networks, such as a recommended quick guide and a series of videos by 3Blue1Brown, which effectively explain the concepts. Key points discussed emphasize that deep learning primarily involves curve fitting and classification, with neural networks acting as complex function parameterizers. The conversation highlights the challenges of overfitting and error estimation in deep learning, noting that while traditional methods like linear and logistic regression serve as foundational concepts, the surprising capabilities of deep learning warrant further exploration. Overall, the focus remains on finding concise, effective resources to build a solid understanding of neural networks.
Phylosopher
Messages
139
Reaction score
26
TL;DR Summary
Doing my master thesis in physics using machine learning (geared toward physics not machine learning). I can not afford learning neural networks "the regular way" because lack of time. What should I start with?
Hello everyone,
I am currently a master student working on physics and neural networks. I have already started producing neural network results (I use tensorflow and keras) so I know how to program the basic things that I am required to do, the problem is that I do not understand them well.

I do not have a good solid background when it comes to the neural networks themselves, and as a master student working on his thesis, I can not/I do not want to write my thesis blindly. I want to truly understand the behavior of my program.

Problem is, I can not afford reading a 600 page book on ML! My idea was to read the papers that are mentioned in the tensorflow pages once I use a specific function or class(Example: Here). But as I said, I do not have a solid background to read them properly.
So, my question is, what resources you think I should read first before I delve deeper in these papers? I need the bare minimum so I can accelerate my learning.

Things I have bookmarked so far that I think are useful:: Quick hand on introduction, Intro book, More detailed Intro book ... Do you think these are good starting points? Do you have better suggestions?
 
Technology news on Phys.org
Quick hand on introduction - I can thoroughly recommend this, it should help you develop a deep intuition for the process and could be all you need.

Intro book - I'm not familiar with this but looking at the ToC it might be a good fit.

More detailed Intro book - this seems to have a much wider scope than you need but might be good context.
 
  • Like
Likes FactChecker, jedishrfu and atyy
This short, four video set is the best explanation of neural networks that I've seen in my three years of study - 3Blue1Brown Neural Networks.
 
  • Like
Likes jedishrfu, Greg Bernhardt, nsaspook and 3 others
The minimum ideas of machine learning are that deep learning is essentially curve fitting or classification. A neural network simply parameterizes a very complicated function. A traditional algorithm for curve fitting is linear regression. A traditional algorithm for classification is logistic regression.

Compared to basic statistical methods, an odd thing about neural networks is that there are many more parameters than data points. There are tricks that prevent overfitting. However, why overfitting can be avoided is not yet well understood.

Also, methods of estimating errors in deep learning are not well understood. However, there are reasonable attempts.
 
  • Like
Likes Borg and Greg Bernhardt
atyy said:
The minimum ideas of machine learning are that deep learning is essentially curve fitting or classification
That is the essential lesson. It is possible that more than one method could lead to the same or a very similar curve fit. It's possible even if those other methods are not similar to deep learning.

So if you are just using the fit, or just using a tool to create the fit, it is not essential to learn the methods used.
 
atyy said:
The minimum ideas of machine learning are that deep learning is essentially curve fitting or classification. A neural network simply parameterizes a very complicated function. A traditional algorithm for curve fitting is linear regression. A traditional algorithm for classification is logistic regression.
Although I wouldn't want to make deep learning sound too ordinary. Even for someone very used to curve fitting, there are many aspects and capabilities of deep learning that are surprising, powerful, and fascinating.
 
Thread 'Is this public key encryption?'
I've tried to intuit public key encryption but never quite managed. But this seems to wrap it up in a bow. This seems to be a very elegant way of transmitting a message publicly that only the sender and receiver can decipher. Is this how PKE works? No, it cant be. In the above case, the requester knows the target's "secret" key - because they have his ID, and therefore knows his birthdate.
I tried a web search "the loss of programming ", and found an article saying that all aspects of writing, developing, and testing software programs will one day all be handled through artificial intelligence. One must wonder then, who is responsible. WHO is responsible for any problems, bugs, deficiencies, or whatever malfunctions which the programs make their users endure? Things may work wrong however the "wrong" happens. AI needs to fix the problems for the users. Any way to...

Similar threads

Back
Top