How Do Artificial Neural Networks Use Output Data to Approximate Functions?

AI Thread Summary
Artificial neural networks (ANNs) use output nodes that apply a threshold function to a linear combination of inputs from the previous layer to approximate functions. For example, in a single-layer ANN, the output is directly derived from the input through a linear combination. When attempting to approximate a function, the outputs from the nodes, after applying a threshold function like sigmoid, can be transformed to fit the desired range of the target function. To approximate functions with different ranges, such as Sin[x]+30, a scaling or transformation of the output values is necessary to align them with the target function's output range. Ultimately, the output layer's results serve as the approximations for the functions being modeled.
phoenixthoth
Messages
1,600
Reaction score
2
the ouput nodes are a threshold function of a linear combination of inputs from the previous layer; in a single layered artificial neural network, the output is a linear combination of the input and if the ann has a n_1-n_2-...-n_j-n_(j+1) structure then the n_(j+1) output layers are all a threshold function of a linear combination of the outputs geerated from layer j which has n_j nodes. what is done with the outputs to obtain an approximation to a function?

say you have three output nodes and after you apply the threshold function, let's say it's sigmoid in this case, you get three outputs 0.2, 0.3, and 0.9. what do you do with those numbers if you're trying to approximate a function g?

or let's say you have one ouput node and after you apply the threshold function f to the dot product of the current weight vector and the reults of the previous layer, and you get 0.2. what do you do with that output to approximate a function?

or in general, how would you approximate the function x^2 using an ann? say on the interval [0,1] or [-1,1]...

in all the references I've seen, they go into depth about the error functions, back propogation and how to update the weights, blah blah blah but I'm failing to see where they explain how exactly one uses an ann to fit data.

my semi-ultimate goal would be to use an ann to approximate the fractional iterates of a function...
 
Mathematics news on Phys.org
Generally, your output layer has one node for each function you are trying to approximate. The output of that node is your approximation.
 
the output node is the result of applying a threshold function to a linear combination of outputs of the previous layer or, in the case of no hidden layers, the inputs. however, the threshold function has a limited range such as (0,1) or even {0,1}. so how does one get from the output given by an output node to something suitable to approximate a function with a different range? like if i wanted to approximate Sin[x]+30, how would i actually use the output of a threshold function to do this?
 
Insights auto threads is broken atm, so I'm manually creating these for new Insight articles. In Dirac’s Principles of Quantum Mechanics published in 1930 he introduced a “convenient notation” he referred to as a “delta function” which he treated as a continuum analog to the discrete Kronecker delta. The Kronecker delta is simply the indexed components of the identity operator in matrix algebra Source: https://www.physicsforums.com/insights/what-exactly-is-diracs-delta-function/ by...
Fermat's Last Theorem has long been one of the most famous mathematical problems, and is now one of the most famous theorems. It simply states that the equation $$ a^n+b^n=c^n $$ has no solutions with positive integers if ##n>2.## It was named after Pierre de Fermat (1607-1665). The problem itself stems from the book Arithmetica by Diophantus of Alexandria. It gained popularity because Fermat noted in his copy "Cubum autem in duos cubos, aut quadratoquadratum in duos quadratoquadratos, et...
Thread 'Imaginary Pythagorus'
I posted this in the Lame Math thread, but it's got me thinking. Is there any validity to this? Or is it really just a mathematical trick? Naively, I see that i2 + plus 12 does equal zero2. But does this have a meaning? I know one can treat the imaginary number line as just another axis like the reals, but does that mean this does represent a triangle in the complex plane with a hypotenuse of length zero? Ibix offered a rendering of the diagram using what I assume is matrix* notation...
Back
Top