Testing Randomness in a Set of 200+ Data Points

In summary, the person is seeking help with understanding statistics and testing for abnormal data points in a set of 200 positive integers. They plan to use this information in a blog post and need guidance on how to determine if a data point is too large to be random or if a source is producing abnormally high numbers over a period of ten years. The suggested method is to compute the mean and standard deviation of the data points and compare them to a typical distribution, such as the Gaussian distribution. The person is advised to specify a cutoff for what is considered an "abnormally high" value before analyzing the data. Further steps include graphing the data over time and examining individual years for unusual values.
  • #1
qspeechc
844
15
Hi everyone.

It's been years since I've done any stats, so I need a bit of help, please. I want to include it in a blog post I'm going to do (not here on PF), so I don't want to give away too many details :p I apologise for my terrible understanding of stats, please be patient!

Anyway, over ten years I have 20 data points for each year, i.e. 200 in total, which are positive integers. In practice they are never higher than 2000, although conceivably they could be. The assumption is that each number is generated randomly.

1) How do I test if a given data point is too large to be random, given that the other numbers tend to be smaller?

2) A 'source' produces one data point a year, how can I test if this source is producing abnormally high numbers over the ten years?

Thank you for any help.
 
Mathematics news on Phys.org
  • #2
For both cases I would compute the mean and standard deviation of the distribution of your 200 data points.
1) You can check if they follow a typical distribution (most notably the Gaussian distribution). If yes, everything that you would not expect given this distribution might be some real effect. You cannot be sure without a clear model, but you can get a good idea with that method.
2) Check the mean and expected deviation of this mean, see if it is compatible with the first distribution.
 
  • Like
Likes qspeechc
  • #3
If all the values are integers they can't be "normally distributed", even if they are symmetric. (The normal distribution is just a convenient description of patters often found in data anyway, no data is truly normal.) But even if they are symmetric, the process outlined above would indicate only that a value is an "outlier" - that doesn't disqualify it as being non-random, simply identifies it as unusual in size.

As a first step you need to specify what qualifies as "abnormally high" (do you have a specific cutoff for that? If not, then saying something like any value more than 3 standard deviations above the mean, or more than 1.5IQR above the third quartile, is needed). Once this is done you might
* graph the data over time and look to see which years, if any, have unusally large values
* look at a plot of each year's data (boxplot?) to check just that group

But again, first making a more specific description of what you mean is where you need to begin.
 

What is randomness and why is it important in data analysis?

Randomness refers to the lack of any discernible pattern or predictability in a set of data points. It is important in data analysis because it allows us to make unbiased and accurate conclusions about a population based on a sample of data.

How do you determine if a set of 200+ data points is random?

There are several statistical tests that can be used to determine if a set of data points is random. These include the Chi-Square test, Kolmogorov-Smirnov test, and Runs test. These tests assess the distribution and patterns within the data set to determine if it is consistent with randomness.

What factors can affect the randomness of a data set?

There are several factors that can affect the randomness of a data set, including the sampling method used, the size of the sample, and the characteristics of the population being studied. Other factors such as human error or bias can also influence the randomness of a data set.

What are the limitations of testing randomness in a set of 200+ data points?

One limitation is that statistical tests can only assess the randomness of a sample, not the entire population. Additionally, these tests may also be affected by the assumptions made about the data and the chosen significance level. It is also important to note that a data set may appear random, but still contain underlying patterns or relationships that are not captured by the statistical tests.

How can the results of testing randomness be interpreted?

The results of testing randomness should be interpreted in the context of the specific data set and the chosen statistical test. If the test indicates that the data set is random, this means that there is no evidence of a significant pattern or relationship within the data. However, if the test indicates that the data set is not random, further investigation may be needed to determine the source of any patterns or relationships found.

Similar threads

Replies
11
Views
833
  • Precalculus Mathematics Homework Help
Replies
17
Views
1K
  • General Math
Replies
1
Views
1K
  • Calculus and Beyond Homework Help
Replies
14
Views
2K
  • Programming and Computer Science
Replies
11
Views
997
  • Calculus and Beyond Homework Help
Replies
4
Views
1K
  • Precalculus Mathematics Homework Help
Replies
3
Views
1K
Replies
9
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
15
Views
2K
Back
Top