Sensor data in Deep Learning: questions on a simple level

Click For Summary

Discussion Overview

The discussion revolves around understanding how Deep Learning and AI handle sensor data, particularly from a simplified perspective. Participants explore various conceptual frameworks and resources to clarify these ideas, touching on topics such as Model Theory, algorithm selection, and the iterative nature of machine learning processes.

Discussion Character

  • Exploratory
  • Conceptual clarification
  • Debate/contested
  • Homework-related
  • Mathematical reasoning

Main Points Raised

  • One participant seeks a simplified explanation of how Deep Learning processes sensor data, proposing two conceptual models based on logic and Model Theory.
  • Another participant finds the initial inquiry vague and suggests a video series as a resource for better understanding.
  • A participant acknowledges the usefulness of the recommended videos and expresses a desire to refine their questions about AI from a Model Theory perspective.
  • Recommendations for books aimed at beginners in machine learning are provided, highlighting gaps in understanding internal workings of algorithms.
  • Some participants discuss the artistic aspect of machine learning, emphasizing the selection of appropriate algorithms for specific tasks and the iterative nature of refining models.
  • There are mentions of using thresholds and loss functions in algorithm iteration, with some participants noting the complexity and flexibility involved in machine learning strategies.
  • Concerns are raised about the challenges of achieving rigor in explanations while navigating the complexities of machine learning processes.
  • One participant draws parallels between the iterative process in machine learning and more complex tasks like playing chess or Go.

Areas of Agreement / Disagreement

Participants express a range of views on the nature of machine learning, with some emphasizing its artistic aspects and others focusing on the technical details. There is no consensus on a single approach or understanding of how to simplify the explanation of sensor data handling in Deep Learning.

Contextual Notes

Participants acknowledge the limitations of their explanations, including the vagueness of initial inquiries and the complexity of the subject matter, which may depend on specific definitions and assumptions.

nomadreid
Gold Member
Messages
1,773
Reaction score
256
First of all: I am not a programmer, I cannot read code, only have a background in pure mathematics, am looking for a very simplified summary ("for Dummies") of the way Deep Learning and similar AI routines handle sensor data. I looked at a few articles such as http://papers.www2017.com.au.s3-website-ap-southeast-2.amazonaws.com/proceedings/p351.pdf, but these are written for those who already understand the basic idea and wish to go into details. So, the question is such: I imagine (in my naïve approach) the computer-without-sensors to be composed of a (first-order) logic, and then the sensor data might be attached in one of two different ways (please hold your laughter for the moment):
(a) the coded sensor data provides constant values for variables in the logic, so that the whole thing remains purely syntactical, with truth predicates either implicit or explicit, or
(b) the sensor data is the universe (in the sense of a model, i.e., an interpretation, that is, semantics) in the sense of Model Theory), and correspondence (the interpretation function in the sense of Model Theory) of the values with the syntactical symbols, guided by a given lattice of truth values (evaluations), all this then changing à la Kripke except much faster.
I emphasize that I understand that the true situation is likely not to be able to be fit into either one of these characterizations, but I present it as a start from which a knowledgeable person may start a rough explanation (without quoting lines of code).
Thanks.
 
Technology news on Phys.org
This is really very vague. I read universe, sensors, deep learning and give me a summary and all I can think of is this video sequence by 3blue1brown on YouTube:

 
  • Like
Likes   Reactions: nomadreid
Thank you, jedishrfu. On your recommendation, I have watched the four videos in this sequence. It addresses my question (obliquely, but in the right direction) , which, I recognize was exceedingly vague (due to my zero-background in AI), and this gives me a basis upon which to better and more precisely formulate my further questions in my attempt to understand more of AI (including deep learning) from a purely Model Theory viewpoint. So your nudge was very helpful.
 
  • Like
Likes   Reactions: anorlunda
There is actually a Machine Learning for Dummies book that covers things in greater detail. Its a few years old and the industry is moving but looks pretty good.

Another excellent book is Hands On Machine Learning by Aurelian Geron. It covers the basics and Google Tensorflow v1 and could actually make you into an ML novice.

In general they teach the algorithm theory and how to use the algorithms but don't tell you how they work internally.
 
Last edited:
  • Like
Likes   Reactions: nomadreid
Thanks, jedishrfu. Upon your recommendation, I have (legally) downloaded the Dummies book, and will keep the other recommendations in mind. I will be looking more for the motivations behind the techniques than for the techniques themselves.
 
ML is more of an art than science. Practioners select the appropriate algorithm for the task at hand. Sometimes neural nets make sense and sometimes genetic algorithms or k nearest neughbor.

What you discover though is there is nothing magic as they just divide up a search space and eventually find a reasonable though not necessarily the best answer.
 
  • Like
Likes   Reactions: nomadreid
The way I understand it, you iterate your algorithm, be it regression, etc. using a threshold value , loss function, until your algorithm has succeeded meaning the loss is below the threshold value and it then stops. E ample, you conduct your regression , compute your loss. If the loss is below the threshold, you're done. If not, you tweak your parameters and iterate until your loss is below the threshold value.
 
  • Like
Likes   Reactions: nomadreid
That’s somewhat simplistic strategy, rather you might tighten your threshold if a lot examples pass the criteria. Of course, you risk centering on a saddle point but such is life.

ML is still a flexible art where you start with schemes that work and then adapt them to your problem and try not to get lost in the data analysis woods. There is a drive to make ML more scientific but I haven’t seen much success in explainable AI or in other methods.
 
  • Like
Likes   Reactions: nomadreid
Well,yes, it is too hard to give a more precise explanation without rigor and getting lost in the trees. Each aspect has its own complexity. You may also tweak or even change the algorithm, the loss function , etc. So I sacrificed rigor in order to shed some light, and hopefully it did accomplish that goal.
 
  • Like
Likes   Reactions: nomadreid
  • #10
The idea of tightening the threshold, etc. is explained pretty well for text recognition in that series of videos which appeared in Post #2. For more complex tasks, such as playing chess or Go, I presume the idea of taking a program that more or less works and tweaking it according to its successes and failures follows the same lines.
 
Last edited:
  • Like
Likes   Reactions: WWGD

Similar threads

  • · Replies 64 ·
3
Replies
64
Views
3K
Replies
29
Views
6K
  • · Replies 14 ·
Replies
14
Views
3K
  • · Replies 8 ·
Replies
8
Views
4K
  • · Replies 37 ·
2
Replies
37
Views
7K
  • · Replies 2 ·
Replies
2
Views
477
Replies
10
Views
5K
  • · Replies 2 ·
Replies
2
Views
4K
  • · Replies 8 ·
Replies
8
Views
3K