How should a physicist learn Machine Learning?

AI Thread Summary
A master’s student working on physics and neural networks seeks to deepen their understanding of neural networks while using TensorFlow and Keras for their thesis. They express a need for accessible resources that provide foundational knowledge without overwhelming detail, as they cannot commit to lengthy texts. Suggestions include introductory materials that foster intuition about neural networks, such as a recommended quick guide and a series of videos by 3Blue1Brown, which effectively explain the concepts. Key points discussed emphasize that deep learning primarily involves curve fitting and classification, with neural networks acting as complex function parameterizers. The conversation highlights the challenges of overfitting and error estimation in deep learning, noting that while traditional methods like linear and logistic regression serve as foundational concepts, the surprising capabilities of deep learning warrant further exploration. Overall, the focus remains on finding concise, effective resources to build a solid understanding of neural networks.
Phylosopher
Messages
139
Reaction score
26
TL;DR Summary
Doing my master thesis in physics using machine learning (geared toward physics not machine learning). I can not afford learning neural networks "the regular way" because lack of time. What should I start with?
Hello everyone,
I am currently a master student working on physics and neural networks. I have already started producing neural network results (I use tensorflow and keras) so I know how to program the basic things that I am required to do, the problem is that I do not understand them well.

I do not have a good solid background when it comes to the neural networks themselves, and as a master student working on his thesis, I can not/I do not want to write my thesis blindly. I want to truly understand the behavior of my program.

Problem is, I can not afford reading a 600 page book on ML! My idea was to read the papers that are mentioned in the tensorflow pages once I use a specific function or class(Example: Here). But as I said, I do not have a solid background to read them properly.
So, my question is, what resources you think I should read first before I delve deeper in these papers? I need the bare minimum so I can accelerate my learning.

Things I have bookmarked so far that I think are useful:: Quick hand on introduction, Intro book, More detailed Intro book ... Do you think these are good starting points? Do you have better suggestions?
 
Technology news on Phys.org
Quick hand on introduction - I can thoroughly recommend this, it should help you develop a deep intuition for the process and could be all you need.

Intro book - I'm not familiar with this but looking at the ToC it might be a good fit.

More detailed Intro book - this seems to have a much wider scope than you need but might be good context.
 
  • Like
Likes FactChecker, jedishrfu and atyy
This short, four video set is the best explanation of neural networks that I've seen in my three years of study - 3Blue1Brown Neural Networks.
 
  • Like
Likes jedishrfu, Greg Bernhardt, nsaspook and 3 others
The minimum ideas of machine learning are that deep learning is essentially curve fitting or classification. A neural network simply parameterizes a very complicated function. A traditional algorithm for curve fitting is linear regression. A traditional algorithm for classification is logistic regression.

Compared to basic statistical methods, an odd thing about neural networks is that there are many more parameters than data points. There are tricks that prevent overfitting. However, why overfitting can be avoided is not yet well understood.

Also, methods of estimating errors in deep learning are not well understood. However, there are reasonable attempts.
 
  • Like
Likes Borg and Greg Bernhardt
atyy said:
The minimum ideas of machine learning are that deep learning is essentially curve fitting or classification
That is the essential lesson. It is possible that more than one method could lead to the same or a very similar curve fit. It's possible even if those other methods are not similar to deep learning.

So if you are just using the fit, or just using a tool to create the fit, it is not essential to learn the methods used.
 
atyy said:
The minimum ideas of machine learning are that deep learning is essentially curve fitting or classification. A neural network simply parameterizes a very complicated function. A traditional algorithm for curve fitting is linear regression. A traditional algorithm for classification is logistic regression.
Although I wouldn't want to make deep learning sound too ordinary. Even for someone very used to curve fitting, there are many aspects and capabilities of deep learning that are surprising, powerful, and fascinating.
 
Dear Peeps I have posted a few questions about programing on this sectio of the PF forum. I want to ask you veterans how you folks learn program in assembly and about computer architecture for the x86 family. In addition to finish learning C, I am also reading the book From bits to Gates to C and Beyond. In the book, it uses the mini LC3 assembly language. I also have books on assembly programming and computer architecture. The few famous ones i have are Computer Organization and...
I have a quick questions. I am going through a book on C programming on my own. Afterwards, I plan to go through something call data structures and algorithms on my own also in C. I also need to learn C++, Matlab and for personal interest Haskell. For the two topic of data structures and algorithms, I understand there are standard ones across all programming languages. After learning it through C, what would be the biggest issue when trying to implement the same data...

Similar threads

Back
Top