Statistically identifying cheaters at pub trivia? Could it be done?

  • Context: Undergrad 
  • Thread starter Thread starter enkrypt0r
  • Start date Start date
Click For Summary

Discussion Overview

The discussion revolves around the feasibility of statistically identifying cheaters in pub trivia games, particularly focusing on the use of algorithms to analyze teams' answers to questions. Participants explore various methods, including the use of trap questions and the challenges posed by small sample sizes and the nature of free-form answers.

Discussion Character

  • Exploratory
  • Debate/contested
  • Technical explanation

Main Points Raised

  • One participant suggests that it may be possible to flag teams based on their performance, particularly if they consistently answer questions correctly while others do not.
  • Another participant argues that without a clear symptom of cheating, it may be statistically invalid to distinguish between skill and cheating.
  • Some propose the idea of designing trap questions specifically aimed at identifying cheaters, although the effectiveness of this method is debated.
  • A participant mentions the need for data from both known cheaters and non-cheaters to establish a baseline for comparison.
  • Concerns are raised about the assumption that there is a limit to how well a person can perform, suggesting that this assumption is debatable.
  • One participant highlights the potential influence of team expertise in specific subjects, which could skew results and complicate the identification of cheating.
  • Another suggests the use of technology to detect cell phone usage during trivia, although practical implementation in a pub setting is questioned.

Areas of Agreement / Disagreement

Participants express a range of opinions, with no clear consensus on the methods for identifying cheaters. Some agree on the potential of trap questions, while others remain skeptical about the statistical validity of distinguishing skill from cheating.

Contextual Notes

The discussion highlights limitations such as the small sample size of teams and questions, the binary nature of scoring, and the need for a more extensive dataset to draw meaningful conclusions.

enkrypt0r
Messages
10
Reaction score
0
Forgive my ignorance; I'm a programmer and relatively math-minded, but I haven't taken statistics since high school.

From what I understand, there are many different ways to do this for things such as standardized tests, given to thousands of students. I'm wondering if it would be feasible to have a sort of algorithm which, given information about which questions each team got right and wrong, would send up a red flag if something didn't jive.

For the sake of example, "cheating" here just constitutes using a cell phone or some other method of getting the answer. Don't worry about teams collaborating, copying off of or listening in on one another, or anything like that. We'll assume that the questions expect free-form answers, i.e. no multiple choice. Either they wrote "Waterloo" or they didn't. It's binary.

The big hurdle here (for the place I play trivia, at least) would be statistical certainty. There are generally only 6-10 teams, and each team generally has roughly five people. There are 50-60 questions in a given night. This seems like so few points to draw data from, but (as a programmer) I'm used to much larger datasets and I don't really remember how correlation coefficients work. Would it be possible to draw any meaningful conclusions (or at least hints) given such a small data set?

I would think that if a team is consistently the only one to get an answer correct, that would be a big indicator. It would, however, also suggest that the team is just better at trivia. Is there a way to sort out when their knowledge is unlikely to be unassisted?

Would this be at all doable given the parameters I've outlined, or is there just no way to tell given *only* information about which teams answers which questions correctly and incorrectly? I realize that a better way to catch cheaters is probably instinct and observation.

I realize I'm relatively ignorant in this area, so please feel free to correct my thoughts and assumptions.

Thanks!
 
Physics news on Phys.org
I don't think that there is a statistically valid way to distinguish between skill and cheating. Unless you can think of some symptom of cheating that would show up, how could it be done?
 
I think you could design trap questions aimed at finding those cheating, but otherwise I don't see how.
 
enkrypt0r said:
We'll assume that the questions expect free-form answers, i.e. no multiple choice. Either they wrote "Waterloo" or they didn't. It's binary.

The scoring may be binary, but aren't there many possible wrong answers?

You'd need to collect data from teams known not to cheat and teams known to cheat. To collect such data, you could form a "simulated" teams that answered the questions but didn't actually enter the competition. You could have some simulated teams cheat and some not cheat.
 
The better a person is, the more his answers will agree with the cheaters. The only way to distinguish between the two is to assume that there is a limit to how good a person can be. Put in some questions that no one could answer without cheating. But that is a very debatable assumption.

The only way I know to catch the cheaters is to ask the same or similar questions in person, when you know they can not cheat and see if they are really that good. That is probably not what you had in mind.
 
I agree with the general idea here. A naturally good team would answer correct on more questions, but ought also to answer correct on each question with a probability which scales accordingly to the general proportion of teams who got it right. If a team consistently deviates from this trend, it is suspicious. There are some flaws to this though, suppose it is a team of mathematicians, then it is very likely they will get the questions of mathematical content correct no matter the general difficulty of the question.
 
Last edited:
Borek said:
I think you could design trap questions aimed at finding those cheating, but otherwise I don't see how.
Perhaps you could arrange that some questions are much easier than others (of otherwise similar difficulty) if one has access to the net. You could then look for a correlation on that basis. Wouldn't stand up in a court of law, though.
 
enkrypt0r said:
For the sake of example, "cheating" here just constitutes using a cell phone or some other method of getting the answer.

I don't know if the pub would go to these extreme lengths, but I have a solution!

In my calculus course, a professor claimed that he would use a device to detect whether cell phones were turned on during the final exam. I am ignornant of this technology, but assuming he wasn't bluffing, such a device could be implemented.

Of course, a participant could claim to take a bathroom break, and use their phone in the bathroom. The pub could suspend a player's participation in the game, for one question, whenever they leave the cellphone-free-zone.
 
Borek said:
I think you could design trap questions aimed at finding those cheating, but otherwise I don't see how.

I think this is the most truthful answer. You lack data for a statistical identification. But like Borek suggested - such question, could be a very good way to find cheaters.
 

Similar threads

  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 131 ·
5
Replies
131
Views
10K
  • · Replies 13 ·
Replies
13
Views
2K
  • · Replies 16 ·
Replies
16
Views
2K
  • · Replies 4 ·
Replies
4
Views
5K
Replies
3
Views
3K
  • · Replies 11 ·
Replies
11
Views
4K
  • · Replies 23 ·
Replies
23
Views
4K
  • · Replies 24 ·
Replies
24
Views
4K
  • · Replies 6 ·
Replies
6
Views
2K