Understanding Mean Squared Error in Neural Networks

AI Thread Summary
Mean Squared Error (MSE) is used in neural networks because it provides a non-zero value that reflects the magnitude of errors, unlike the mean error which can average to zero with alternating positive and negative values. MSE is calculated by squaring the errors, ensuring that all values contribute positively to the final result, thus offering a more useful metric for model evaluation. This approach is similar to using Root Mean Squared (RMS) values in electrical engineering, where squaring negative and positive values yields a meaningful average. The learning algorithm in neural networks concludes when the MSE reaches a predefined threshold, indicating satisfactory model performance. Understanding MSE is crucial for effective training and evaluation of neural network models.
hisham.i
Messages
176
Reaction score
2
In neural network the learning algorithm ends when the mean squared error value is less than or equal to a value we have precised.
But i don't understand why we are comparing with the mean squared error and not the mean error?
What does the mean squared error represent?
 
Engineering news on Phys.org
I'm not sure of the exact answer you require, but if it's anything like using RMS values it's pretty straight forward.

Let's say you have two error values: -1 and 1

The mean error value is 0.
The mean squared error is 1.

If I remember correctly, it is because you have alternating positive and negative values and no matter what you do you end up with 0 as the mean if you simply take the average of the exact values.

For example, with alternating current of 230V (UK standard supply) you have a sin wave with a maximum of +320V and a minimum of -320V. If you average these values you get an average voltage out of your wall socket of 0V - this is of no use to you.

So you use an RMS (Root Mean Squared) value to get a useful value.

In this case you have +320V and -320V. So you square them (+3202 and -320V-2).
Add the squared values together (+320V2+-320V2).
Square root them and then take the mean (Sqrt(+320V2+-320V2))/2.

This then gives you the RMS voltage. For the UK this is ~230V.

So by using a value such as your "squared error" you get a useful answer instead of 0 every time.
 
hisham.i said:
In neural network the learning algorithm ends when the mean squared error value is less than or equal to a value we have precised.
But i don't understand why we are comparing with the mean squared error and not the mean error?
What does the mean squared error represent?

I'm sure you mean the root mean squared error meaning the root of the averaged sum of the squared differences from the averaged value.
 
Thread 'Weird near-field phenomenon I get in my EM simulation'
I recently made a basic simulation of wire antennas and I am not sure if the near field in my simulation is modeled correctly. One of the things that worry me is the fact that sometimes I see in my simulation "movements" in the near field that seems to be faster than the speed of wave propagation I defined (the speed of light in the simulation). Specifically I see "nodes" of low amplitude in the E field that are quickly "emitted" from the antenna and then slow down as they approach the far...
Hello dear reader, a brief introduction: Some 4 years ago someone started developing health related issues, apparently due to exposure to RF & ELF related frequencies and/or fields (Magnetic). This is currently becoming known as EHS. (Electromagnetic hypersensitivity is a claimed sensitivity to electromagnetic fields, to which adverse symptoms are attributed.) She experiences a deep burning sensation throughout her entire body, leaving her in pain and exhausted after a pulse has occurred...
Back
Top