Q - test: How can I find the Q critical value at 95% Confidence?

  • Context: High School 
  • Thread starter Thread starter Another
  • Start date Start date
  • Tags Tags
    Test Value
Click For Summary

Discussion Overview

The discussion centers around the use of the Q test for outlier detection in datasets, specifically addressing how to find the Q critical value at 95% confidence for a dataset of 180 observations. Participants explore various methods for identifying outliers, including graphical techniques and the implications of sample size on the appropriateness of the Q test.

Discussion Character

  • Debate/contested
  • Technical explanation
  • Conceptual clarification

Main Points Raised

  • One participant seeks guidance on determining the Q critical value for a dataset of 180 observations at 95% confidence.
  • Another participant expresses strong disagreement with the Q test, advocating for graphical methods to assess outliers instead, citing concerns about the assumptions of normality inherent in the Q test.
  • Some participants suggest that the Q test is only suitable for small sample sizes and argue that with 180 observations, alternative methods should be considered.
  • Graphical methods, such as box and whisker plots, are mentioned as useful tools for identifying potential outliers without outright rejection based on distributional assumptions.
  • There is a discussion about the appropriate use of the Q test, with one participant stating it should only be used once to reject at most one observation.
  • Another participant emphasizes that box plots can be used repeatedly to identify data points for further investigation, but cautions against rejecting data solely based on box plot results.

Areas of Agreement / Disagreement

Participants generally disagree on the appropriateness of the Q test for larger datasets, with some advocating for its use and others recommending alternative methods. There is no consensus on the best approach to outlier detection in this context.

Contextual Notes

The discussion highlights differing views on the assumptions underlying the Q test and the implications of sample size on its validity. There are unresolved questions regarding the effectiveness of various outlier detection methods and the conditions under which they should be applied.

Another
Messages
104
Reaction score
5
TL;DR
I want to use Q -test to truncate the data. But I have number of data 180 ( n = 180 )

How can i find Q critical value at 95 % Confidence ,When number of data equal to 180 ?
I want to use Q -test to truncate the data. But I have number of data 180 ( n = 180 )
How can i find Q critical value at 95 % Confidence ,When number of data equal to 180 ?
 
Physics news on Phys.org
I fundamentally disagree with the Q test and similar outlier rejection techniques. I tend to use graphical methods to visually assess possible outliers. For any possible outlier I look at the data and see if there was a transcription error or a recorded experimental "glitch" or flaw. If there is a transcription error then I correct it, and if there was a recorded "glitch" then I reject it. I never reject data simply based on distributional assumptions.

In particular, the Q test assumes normality. An outlier can tell you that your data is not normal. If you assume normality anyway and reject the outlier then you are ignoring important information telling you that your assumption is wrong.

That said, the Q test is only intended for small numbers of observations. 180 is too many. You should pick a different test (or use a different approach alltogether).
 
  • Informative
  • Like
Likes   Reactions: Klystron and hutchphd
Dale said:
I fundamentally disagree with the Q test and similar outlier rejection techniques. I tend to use graphical methods to visually assess possible outliers. For any possible outlier I look at the data and see if there was a transcription error or a recorded experimental "glitch" or flaw. If there is a transcription error then I correct it, and if there was a recorded "glitch" then I reject it. I never reject data simply based on distributional assumptions.

In particular, the Q test assumes normality. An outlier can tell you that your data is not normal. If you assume normality anyway and reject the outlier then you are ignoring important information telling you that your assumption is wrong.

That said, the Q test is only intended for small numbers of observations. 180 is too many. You should pick a different test (or use a different approach alltogether).

What do you think, if I try this program?
https://miniwebtool.com/outlier-calculator/
 
I use something similar in my graphical methods. One of the graphical methods I use is a box and whisker plot which uses this calculation to determine if any data points should be plotted individually. I can then look at those individually plotted points and see if there is a transcription error or an experimental reason to reject the data. So I do use that method, not to directly reject data, but to spot data that I should look at in more detail.
 
Dale said:
I use something similar in my graphical methods. One of the graphical methods I use is a box and whisker plot which uses this calculation to determine if any data points should be plotted individually. I can then look at those individually plotted points and see if there is a transcription error or an experimental reason to reject the data. So I do use that method, not to directly reject data, but to spot data that I should look at in more detail.
When number of sample is small , we can use Q - test to select outliers data. If we find that there is still outlier data , we can use Q-test to confirm reject this data again.

When number of sample is large , I want to use Box Plot method to select outliers data.
If I find that there is still outlier data , I can use Box Plot method again ?
 
Another said:
When number of sample is small , we can use Q - test to select outliers data. If we find that there is still outlier data , we can use Q-test to confirm reject this data again.
No. The Q test should never be used more than once and is only used to reject at most one observation.

Another said:
When number of sample is large , I want to use Box Plot method to select outliers data.
If I find that there is still outlier data , I can use Box Plot method again ?
You can use Box Plots as often as you like, but I would not reject any data based only on the Box Plot itself. All that should do is identify data points to investigate. If you investigate and find that there is no transcription error and no experimental problem, then you should keep the outlier.
 
Last edited:

Similar threads

  • · Replies 9 ·
Replies
9
Views
3K
  • · Replies 7 ·
Replies
7
Views
3K
Replies
6
Views
3K
  • · Replies 5 ·
Replies
5
Views
4K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 6 ·
Replies
6
Views
1K
  • · Replies 22 ·
Replies
22
Views
4K
  • · Replies 10 ·
Replies
10
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K