twabby
Registered User
- Mar 9, 2010
- 14,198
- 15,793
Are they really? Just based on what gets posted on twitter whenever a player signs/gets traded, I can't remember ever seeing the models differ significantly in their evaluation of a player. Either they converge because they're all perfect or because they all use similar inputs and methodology and thus are all subject to the same issues and biases. I know which one I'd put my money on.
They are all restricted to the same dataset as far as I'm aware: the NHL's real-time scoring system (RTSS) data that was introduced in the 2007-08 season. It's why a lot of these databases and models have 07-08 as the starting point. That's when they started tracking location events such as shots, and that's also when shift information was available. Before that the only data provided publicly were the simply boxcar counting stats.
I think these models all share a similarish methodology in that they are trying to regress for the outputs from the RTSS data. In other words, given the entirety of the output data that we have (all of the shots, penalties, goals, hits, blocks, etc.), what values for the inputs (e.g. player quality) explain these data the best? Sure it's possible that Jack Johnson has just been getting incredibly unlucky year after year and that's why his on-ice results have been so bad. But it's incredibly unlikely. Similarly, it's possible that Alex Ovechkin has seen a ton of shots, chances, and goals go in against his team because he's just had rotten luck for 7 seasons straight. But it seems the more likely explanation is that he's just a hugely negative impactful player defensively.