Prime Day Part 5 – looking at lucky and unlucky players

  • by

In the last part I argued that lumping all data together would diminish the signal we get. Statistical tools allow us to look at a selection of data and generated a p-value that gives us an idea how significant a finding is.

Most of you will know the famous quote – Lies, damn lies and statistics

Slicing and dicing data to find a s significant p-value is something rightly frowned upon by anyone doing proper statistical work. A hypothesis should be made ahead of the experiment to be truly valid.

Slicing and dicing data gives you multiple chances to you need to correct the p-value accordingly. Relying on extreme data is fraught with danger as you invite to base your findings on possible false data.

I’m putting all of this out here as a disclaimer. I’m not blind or oblivious to these issues. The data I had access to is flawed. The experiment (with hindsight) that gathered the data is flawed. I therefore can’t just look at a p value < 0.05 and call it a day. But p-values still are a powerful tool if I put them into context.

So let us start with the most unlucky players. This was some work I did very early on when I didn’t even had an idea what happened. I was mainly looking for a chance that seemed sensible. I define the most unlucky player as the player who needed to check the highest number of Pokemon to find x shiny. I then look at the number of shiny found and calculate the p-value associated with this data point.

Shiny foundPokemon checkedout of n playersp value1 in 25p value1 in 50p value 1 in 100p value 1 in 150
2085010.0060.8081.0001.000
1182110.0000.1050.8730.990
9100050.0000.0050.4570.863
868030.0000.0730.7560.959
785070.0000.0050.3850.789
61197100.0000.0000.0460.315
52062190.0000.0000.0000.006
4900210.0000.0000.0540.284
31100260.0000.0000.0050.065
2992560.0000.0000.0030.039
1800790.0000.0000.0030.030
07501080.0000.0000.00050.007

Looking at this data we can clearly rule out a drop rate of 1 in 25 and 1 in 50. The poor player who reported 5 shiny out of 2062 would have to be among the most unlucky persons who ever played. Assuming a 1 in 50 rate the chance of this happening is 1 in 997947802242 – yes – you read correctly – 1 in 997 trillion. so either this data point if wrong – or our hypothesis is wrong.

And now we look at the most lucky players. I excluded the last row – there can be a lucky player who found zero shiny.

Shiny found Pokemon checked out of n players p value1 in 25 p value1 in 50 p value 1 in 100 p value 1 in 150
2085010.9940.1920.0000.000
1182411.0000.8970.1290.011
918250.1950.0040.0000.000
824030.6240.0540.0010.000
715070.2530.0110.0000.000
6200100.6920.1090.0040.000
5140190.4900.0630.0030.000
4130210.5980.1210.0100.002
339260.0690.0080.0010.000
232560.1350.0260.0040.001
17790.0290.0080.0020.001

These values clearly rule out any shiny rate smaller then 1 in 50. The data is less extreme – but the 20 in 850 and 9 in 182 still corresponds to a 1 in 1.9 million chance.

We now have a problem. If our data is correct, then we can have neither a high nor a low shiny rate as the lucky and unlucky players just don’t fit.

Some of you will point out that this is just false data. After all – this was survey data. Therefore some deliberately false data or bias is to be expected. And some of that surely has happened.

But just declaring the data invalid when your hypothesis doesn’t fit is lazy. There is another reason why I believe at least most of the data is actually correct. The SilphRoad Research team found a very similar overall drop rate. We regard the SilphRoad Research data as the gold standard and tend to trust it. This data also showed similar patterns as above. Some researchers who seemed ‘too lucky’ and some who seemed ‘too unlucky’.