Metric for rating funtion accuracy from R^m to R.

Click For Summary

Discussion Overview

The discussion revolves around methods for rating the accuracy of functions that map from R^m to R, particularly in the context of generating pseudo-random expressions to approximate a given data set. Participants explore statistical methods for evaluating the accuracy of these functions based on expected and actual outputs.

Discussion Character

  • Exploratory, Technical explanation, Debate/contested

Main Points Raised

  • One participant seeks a method to rate function accuracy using a list of expected and actual outputs, suggesting a statistical approach.
  • Another participant proposes the sum of squared errors as a standard method for evaluating accuracy, noting its analytical convenience.
  • It is mentioned that some practitioners prefer using the sum of absolute errors to avoid underestimating the impact of bad solutions, especially in the presence of noise.
  • A participant expresses a preference for the absolute value metric due to the random nature of their solutions, indicating a concern about accurately reflecting the quality of poor solutions.
  • There is a question raised about the implications of using the absolute value metric, specifically regarding the presence of noise in the data.

Areas of Agreement / Disagreement

Participants have differing views on the best metric to use for evaluating function accuracy, with some advocating for the sum of squared errors and others for the sum of absolute errors. The discussion remains unresolved regarding which method is superior.

Contextual Notes

Participants have not fully addressed the assumptions underlying their chosen metrics, nor have they clarified how noise in the data might affect their evaluations.

TylerH
Messages
729
Reaction score
0
I'm writing program in which I generate pseudo random expressions with the hope of finding one that closely approximates the given data set. The functions map from R^m (an m-tuple of reals) to a real. What I need is a way to rate the functions by their accuracy. Are there any known methods for doing this? Maybe something from stats.

Ideally this would be a method that would take a list of expected outputs and actual outputs and return a real number. A uniform distribution would be good, but not required.
 
Physics news on Phys.org
A standard thing is to do something like
\sum_{j} \left( y_j - z_j \right)^2

where the yj are the actual data that you measured, and the zj are the predictions that you made (this is called the sum of squared errors). It's used mostly because it's easy to work with sums of squares when doing analytical calculations (like trying to minimize it). If you're doing a numerical method, people will often instead use the sum of the absolute values of the errors, especially if they consider the property "makes sure there are no significant outliers at the cost of being off by slightly more on average" to be a negative quality
 
I went with the abs value one because, by nature of the fact they're all random, there are going to be a lot of bad solutions. So I thought a metric which doesn't underestimate their badness would be better. Thanks.
 
TylerH said:
I went with the abs value one because, by nature of the fact they're all random, there are going to be a lot of bad solutions. So I thought a metric which doesn't underestimate their badness would be better. Thanks.

I don't really understand what this post is trying to get at... do you mean that your data has a lot of noise in it?
 

Similar threads

  • · Replies 0 ·
Replies
0
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 11 ·
Replies
11
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 11 ·
Replies
11
Views
3K
  • · Replies 24 ·
Replies
24
Views
7K
  • · Replies 11 ·
Replies
11
Views
2K
  • · Replies 19 ·
Replies
19
Views
1K
  • · Replies 24 ·
Replies
24
Views
5K
  • · Replies 39 ·
2
Replies
39
Views
5K