- Gnome Stew - https://gnomestew.com -

A Case Study In Dice Stat Tests Part 2: Review of Honest Dice Analysis

In last week’s article we looked at a general approach to doing stat analysis on dice. This week we’re going to look at the analysis that was done in the Kickstarted for Honest Dice | Precision Machined Metal Dice You Can Trust [1]” and see how it stacks up to this ideal. As is the case with many things, there isn’t one right way to do a statistical test, analysis, and presentation but there are quite a few wrong ways to do it. I’m just going to go down the general structure from last week with notes where applicable.

I will leave with one final bit of evidence that Honest Dice are in fact good quality dice. The results calculated show the Honest Die in each die type to have the highest p-value in the group. Let’s assume for a moment that all dice of each type have the same distribution. If x dice all have the exact same distribution, it seems intuitive that if we do a goodness of fit test for each die, each die has an equal chance (1/x) of getting the highest p-value. Thus if there is no difference in the d20s, the chance of the Honest Die getting the highest p-value would be 1/4. For the d6s that would be 1/2. For the d4s that would be 1/3. Since we assume each type of die is all the same we can also safely assume the results across die type would be independent. If these assumptions are true then the probability of getting the highest p-value in all three die types would be the individual probabilities multiplied together. This would be 1/4 * 1/2 * 1/3 = 1/24 ~ .042. Thus given H0: All dice are all the same, the probability of observing the collection of Honest Die p-values we observed is .042. This is lower than a reasonable .05 threshold for significance (.05 is the usual default. It makes sense to use it here). Thus evidence exists to reject H0: all the dice of each type share a distribution.

Last week was part 1: General Approach [3].

Next week is part 3: Suggested Analysis [4].