Thrilled to see a NeurIPS 2020 paper that eeks out a bit more performance on benchmark image datasets by...removing mislabeled examples from the train set!

"Identifying Mislabeled Data using the Area Under the
Margin Ranking": https://arxiv.org/pdf/2001.10528.pdf (1/4)
When I do DS, I think about: "Whenever model performance is bad, we (scientists and practitioners) shouldn’t only resort to investigating model architecture and parameters. We should also be thinking about “culprits” of bad performance in the data." https://www.shreya-shankar.com/making-ml-work/  (2/4)
In industry, I learned so much from my DS colleagues by word-of-mouth: for example, oftentimes I get bigger performance boosts by dropping mislabeled examples or intelligently imputing features in "incomplete" data points rather than mindless hyperparameter tuning. (3/4)
It's quite amazing to see that this could work on benchmark datasets 1+ *decade* after release :)

h/t @AlexTamkin for sharing the paper with me and authors @gpleiss, @Tianyi_Zh, @side1track1, and Kilian Q. Weinberger (4/4)
You can follow @sh_reya.
Tip: mention @twtextapp on a Twitter thread with the keyword “unroll” to get a link to it.

Latest Threads Unrolled:

By continuing to use the site, you are consenting to the use of cookies as explained in our Cookie Policy to improve your experience.