Neural Networks Question about the Hebbs and Delta Rule

Click For Summary
SUMMARY

This discussion focuses on the application of the Hebbian and Delta rules in neural networks, referencing Jeff Heaton's book "Neural Networks in Java." The Hebbian rule is defined as Δwij = μ · ai · aj, where the weight adjustment is calculated based on the activations of two neurons. The Delta rule is expressed as Δwij = 2 · μ · xi · (ideal - actual)j, which incorporates the error between the expected and actual outputs. The user seeks clarification on the activation terminology and the correct application of these rules.

PREREQUISITES
  • Understanding of Hebbian learning principles
  • Familiarity with the Delta rule in neural networks
  • Knowledge of activation functions (e.g., ReLU, sigmoid)
  • Basic proficiency in neural network weight adjustment calculations
NEXT STEPS
  • Study the mathematical derivation of the Hebbian rule and its applications in neural networks
  • Learn about the Delta rule and its implementation in training neural networks
  • Explore various activation functions and their effects on neural network performance
  • Review practical examples of weight adjustment in neural networks using Python libraries like TensorFlow or PyTorch
USEFUL FOR

Students, researchers, and practitioners in machine learning and artificial intelligence, particularly those interested in neural network training methodologies and weight adjustment techniques.

Peter_Newman
Messages
155
Reaction score
11
Good afternoon,

I am currently working on Neural Networks and I am reading an introduction by Jeff Heaton (Neural Networks in Java).

Now there are two tasks there whose solutions interest me. The first task is about applying the Hebbs rule. In the book it is given wrong because of a typo but I just googled the Hebbs rule and found it in "correct" form:

##\Delta w_{ij} = \mu \cdot a_i \cdot a_j##

##\Delta w_{ij}## weigth for the connection from neuron ##i## to ##j##
##\mu## learning rate
##a_i, a_j## activation of each neuron

In the first task it says now: Use the Hebbs rule to calculate the adjustment of the weighting, given as specifications: Two neurons N1 and N2, N1 to N2 weight: 3, N1 Activation: 2, N2 Activation: 6

I have now applied the rule bluntly, additionally I have to update the old weighting, therefore ##w_{new} = w_{old} + \Delta w_{ij}## If I do this I will come up:
##w_{new} = 3 + 1*2*6## I have assumed here that the learning rate ##\mu## is 1!

In the second task it says: Use the delta rule to calculate the adjustment of the weighting, given as specifications: Two neurons N1 and N2, N1 to N2 weight: 3, N1 Activation: 2, N2 Activation: 6, Expected: 5.

The delta rule is given in the book as follows:

##\Delta w_{ij} = 2\cdot\mu \cdot x_i \cdot (ideal-actual)_j##

The following is then added:
##\Delta w_{ij}## weigth for the connection from neuron ##i## to ##j##
##\mu## learning rate
The variable ideal represents the desired output of the ##j## neuron. The variable actual represents the actual output of the ##j## neuron. As a result (ideal - actual) is the error. ##x_i## Input for the actual neuron one is looking for (from Video)Alternatively, I found a video of Jeff Heaton, in which he explains this at this point (from minute 5:00, see )

I'm not sure about this task, because overall the term "activation" confuses me a bit. But if I understand the formula correctly, then this is ##w_{new} = w_{old} + \Delta w_{ij}##, where ##\Delta w_{ij}= 2\cdot \mu\cdot x_i\cdot(ideal-actual)_j## From this follows (for me): ##w_{new} = w_{old} + 2\cdot \mu\cdot x_i\cdot(ideal-actual)_j = 2\cdot 1\cdot 2\cdot (5-6)## I have assumed here that the learning rate ##\mu## is 1!

I'm not sure that's right. I'd be curious to hear your opinions.

I also found the book on Google Books: I have included the corresponding page right here: Google Books Link

Important: This is not homework! I bought the book out of interest and I just read it and do the tasks in the book.
 
Technology news on Phys.org
The activation is the output of each node, it is a function of the weights, activation for

ni,j = f(∑wk ⋅ ni-1,k) ,

where f(x) is the activation function, eg relu, sigmoid, tanh, elu, gelu, softmax, parametric relu.
 
We have many threads on AI, which are mostly AI/LLM, e.g,. ChatGPT, Claude, etc. It is important to draw a distinction between AI/LLM and AI/ML/DL, where ML - Machine Learning and DL = Deep Learning. AI is a broad technology; the AI/ML/DL is being developed to handle large data sets, and even seemingly disparate datasets to rapidly evaluated the data and determine the quantitative relationships in order to understand what those relationships (about the variaboles) mean. At the Harvard &...

Similar threads

  • · Replies 2 ·
Replies
2
Views
2K
Replies
3
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 1 ·
Replies
1
Views
3K
Replies
1
Views
3K
  • · Replies 9 ·
Replies
9
Views
5K
  • · Replies 0 ·
Replies
0
Views
1K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 6 ·
Replies
6
Views
2K