Abstract

Deep learning with noisy labels is practically challenging, as the capacity of deep models is so high that they can totally memorize these noisy labels sooner or later during training. Nonetheless, recent studies on the memorization effects of deep neural networks show that they would first memorize training data of clean labels and then those of noisy labels. Therefore in this paper, we propose a new deep learning paradigm called ''Co-teaching'' for combating with noisy labels. Namely, we train two deep neural networks simultaneously, and let them teach each other given every mini-batch: firstly, each network feeds forward all data and selects some data of possibly clean labels; secondly, two networks communicate with each other what data in this mini-batch should be used for training; finally, each network back propagates the data selected by its peer network and updates itself. Empirical results on noisy versions of MNIST, CIFAR-10 and CIFAR-100 demonstrate that Co-teaching is much superior to the state-of-the-art methods in the robustness of trained deep models.

Keywords

MemorizationMNIST databaseComputer scienceDeep neural networksRobustness (evolution)Artificial intelligenceDeep learningArtificial neural networkNoisy dataTraining setMachine learningTraining (meteorology)Data modelingDatabaseMathematics

Affiliated Institutions

Related Publications

Network In Network

Abstract: We propose a novel deep network structure called In Network (NIN) to enhance model discriminability for local patches within the receptive field. The conventional con...

2014 arXiv (Cornell University) 1037 citations

Publication Info

Year
2018
Type
article
Volume
31
Pages
8527-8537
Citations
1215
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

1215
OpenAlex

Cite This

Bo Han, Quanming Yao, Xingrui Yu et al. (2018). Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy Labels. Neural Information Processing Systems , 31 , 8527-8537. https://doi.org/10.5555/3327757.3327944

Identifiers

DOI
10.5555/3327757.3327944