Okay I've got a few more things to add to this.
At first glance they didn't seem highly correlated, but when I added the snap counts for the players I noticed that a lot of the outliers were guys that hadn't played much
At first glance they didn't seem highly correlated, but when I added the snap counts for the players I noticed that a lot of the outliers were guys that hadn't played much
Which makes sense. Madden won't rate a guy in the 90s if he only played 2 games, no matter how great they were, while PFF will be perfectly fine grading it higher, and just toss out those small sample sizes during analysis
How many snaps does a person have to play before Madden and PFF come to an agreement? Well it turns out that largely depends on if they play offense or defense.
If you add a minimum of 400 snaps played, you're left with 397 offensive players (about 12 per team), and their PFF grade explains over half of the variance in their Madden rating
Important to note that none of this actually matters, and that's okay, because sometimes you just want to practice your data skills in R with a fun project.
I used janitor::clean_names() for the first time here and it changed my life
I used janitor::clean_names() for the first time here and it changed my life
Your work doesn't always have to be groundbreaking, go do it anyway if it's fun imo
another really good point, and a big reason Madden and PFF are so far apart on Tannehill in the first chart in this thread https://twitter.com/LeeSharpeNFL/status/1286033513565716480