Understanding the Cost Function in Machine Learning: A Practical Guide

Click For Summary

Discussion Overview

The discussion revolves around understanding the differentiation of a cost function in the context of machine learning, specifically related to a loss function used in neural networks. Participants are exploring the mathematical formulation and notation used in a referenced paper.

Discussion Character

  • Technical explanation
  • Debate/contested

Main Points Raised

  • Emma presents a loss function defined as L=wE, where E=(G-Gest)^2 and G=F'F, and expresses difficulty in understanding the differentiation of this function.
  • One participant questions the notation used, asking if all functions depend on time and why G-G is not equal to zero, suggesting a potential misunderstanding of the notation.
  • Another participant reiterates the confusion regarding the notation and proposes an alternative expression for the loss function.
  • Emma clarifies that G and Gest refer to the output and estimated output of a neural network, and that the primes denote transposition.
  • A participant expresses the need for more context to understand the mathematical aspects being discussed.
  • Emma provides a link to a specific problem described in a paper for further reference.

Areas of Agreement / Disagreement

Participants express varying levels of understanding and confusion regarding the mathematical notation and formulation. There is no consensus on the interpretation of the loss function or the differentiation process.

Contextual Notes

The discussion highlights potential ambiguities in notation and the need for additional context to fully grasp the mathematical concepts involved. Specific assumptions about the functions and their dependencies remain unclear.

emmasaunders12
Messages
43
Reaction score
0
Could someone please help me work through the differentiation in a paper (not homework), I am having trouble finding out how they came up with their cost function.

The loss function is L=wE, where E=(G-Gest)^2 and G=F'F

The derivative of the loss function wrt F is proportional to F'(G-Gest)

Can't seem to figure it out.

Thanks

Emma
 
Last edited:
Physics news on Phys.org
I have some trouble to understand you:

Do all functions depend on, say time ##t##, which the primes refer to? And why isn't ##G-G=0##? I first thought it could be the strange notation of a function, but then you defined a single ##G##. And last, could it be ##L \propto F(G-G)'##?
 
fresh_42 said:
I have some trouble to understand you:

Do all functions depend on, say time ##t##, which the primes refer to? And why isn't ##G-G=0##? I first thought it could be the strange notation of a function, but then you defined a single ##G##. And last, could it be ##L \propto F(G-G)'##?

Thanks for the response, its the loss function of a neural network, so I've corrected to G and Gest, primes refer to transpose
 
emmasaunders12 said:
Thanks for the response, its the loss function of a neural network, so I've corrected to G and Gest, primes refer to transpose
Perhaps someone else can help, but without a lot more context I have no idea what mathematically we are dealing with here.
 
PeroK said:
Perhaps someone else can help, but without a lot more context I have no idea what mathematically we are dealing with here.

The specific problem is described on page 4 here https://arxiv.org/pdf/1505.07376v3.pdf
 

Similar threads

  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 3 ·
Replies
3
Views
4K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 39 ·
2
Replies
39
Views
6K
  • · Replies 6 ·
Replies
6
Views
3K
  • · Replies 2 ·
Replies
2
Views
1K
  • Sticky
  • · Replies 16 ·
Replies
16
Views
12K
Replies
2
Views
936
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 48 ·
2
Replies
48
Views
12K