Timezone: »
Not all examples are created equal, but standard deep neural network training protocols treat each training point uniformly. Each example is propagated forward and backwards through the network the same amount of times, independent of how much the example contributes to the learning protocol. Recent work has proposed ways to accelerate training by deviating from this uniform treatment. Popular methods entail up-weighting examples that contribute more to the loss with the intuition that examples with low loss have already been learned by the model, so their marginal value to the training procedure should be lower. This view assumes that updating the model with high loss examples will be beneficial to the model. However, this may not hold for noisy, real-world data. In this paper, we theorize and then empirically demonstrate that loss-based acceleration methods degrade in scenarios with noisy and corrupted data. Our work suggests measures of example difficulty need to correctly separate out noise from other types of challenging examples.
Author Information
Niel Hu (ML Collective)
Xinyu Hu (Uber)
Rosanne Liu (ML Collective; Google)
Sara Hooker (Google Brain)
Jason Yosinski (Deep Collective)
More from the Same Authors
-
2022 : Invited Talk #2 - Collaborations with ML researchers »
Rosanne Liu -
2021 Workshop: Workshop on Computational Approaches to Mental Health @ ICML 2021 »
Niranjani Prasad · Caroline Weis · Shems Saleh · Rosanne Liu · Jake Vasilakes · Agni Kumar · Tianlin Zhang · Ida Momennejad · Danielle Belgrave -
2021 Social: Open Collaboration in ML Research »
Brenda Ng · Alexander Gu · Jason Yosinski · Rosanne Liu · Luis Granados · Suzana Ilic -
2020 : Brainstorming & Closing »
Mayoore Jaiswal · Ryan Lowe · Jesse Dodge · Jessica Forde · Rosanne Liu -
2020 : Q&A: Pascale Fung »
Pascale FUNG · Rosanne Liu -
2020 Workshop: MLRetrospectives: A Venue for Self-Reflection in ML Research »
Jessica Forde · Jesse Dodge · Mayoore Jaiswal · Rosanne Liu · Ryan Lowe · Rosanne Liu · Joelle Pineau · Yoshua Bengio -
2020 Poster: Estimating Q(s,s') with Deep Deterministic Dynamics Gradients »
Ashley Edwards · Himanshu Sahni · Rosanne Liu · Jane Hung · Ankit Jain · Rui Wang · Adrien Ecoffet · Thomas Miconi · Charles Isbell · Jason Yosinski -
2019 Poster: Metropolis-Hastings Generative Adversarial Networks »
Ryan Turner · Jane Hung · Eric Frank · Yunus Saatchi · Jason Yosinski -
2019 Oral: Metropolis-Hastings Generative Adversarial Networks »
Ryan Turner · Jane Hung · Eric Frank · Yunus Saatchi · Jason Yosinski