UCL Discovery Stage
UCL home » Library Services » Electronic resources » UCL Discovery Stage

Knowledge Distillation Meets Label Noise Learning: Ambiguity-Guided Mutual Label Refinery

Jiang, Runqing; Yan, Yan; Xue, Jing-Hao; Chen, Si; Wang, Nannan; Wang, Hanzi; (2023) Knowledge Distillation Meets Label Noise Learning: Ambiguity-Guided Mutual Label Refinery. IEEE Transactions on Neural Networks and Learning Systems 10.1109/tnnls.2023.3335829. (In press). Green open access

[thumbnail of RunqingJiang-TNNLS-2023-accepted.pdf]
Preview
Text
RunqingJiang-TNNLS-2023-accepted.pdf - Accepted Version

Download (1MB) | Preview

Abstract

Knowledge distillation (KD), which aims at transferring the knowledge from a complex network (a teacher) to a simpler and smaller network (a student), has received considerable attention in recent years. Typically, most existing KD methods work on well-labeled data. Unfortunately, real-world data often inevitably involve noisy labels, thus leading to performance deterioration of these methods. In this article, we study a little-explored but important issue, i.e., KD with noisy labels. To this end, we propose a novel KD method, called ambiguity-guided mutual label refinery KD (AML-KD), to train the student model in the presence of noisy labels. Specifically, based on the pretrained teacher model, a two-stage label refinery framework is innovatively introduced to refine labels gradually. In the first stage, we perform label propagation (LP) with small-loss selection guided by the teacher model, improving the learning capability of the student model. In the second stage, we perform mutual LP between the teacher and student models in a mutual-benefit way. During the label refinery, an ambiguity-aware weight estimation (AWE) module is developed to address the problem of ambiguous samples, avoiding overfitting these samples. One distinct advantage of AML-KD is that it is capable of learning a high-accuracy and low-cost student model with label noise. The experimental results on synthetic and real-world noisy datasets show the effectiveness of our AML-KD against state-of-the-art KD methods and label noise learning (LNL) methods. Code is available at https://github.com/Runqing-forMost/ AML-KD.

Type: Article
Title: Knowledge Distillation Meets Label Noise Learning: Ambiguity-Guided Mutual Label Refinery
Open access status: An open access version is available from UCL Discovery
DOI: 10.1109/tnnls.2023.3335829
Publisher version: http://dx.doi.org/10.1109/tnnls.2023.3335829
Language: English
Additional information: This version is the author accepted manuscript. For information on re-use, please refer to the publisher’s terms and conditions.
Keywords: Noise measurement, Training, Annotations, Knowledge engineering, Feature extraction, Computational modeling, Estimation
UCL classification: UCL
UCL > Provost and Vice Provost Offices > UCL BEAMS
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Maths and Physical Sciences
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Maths and Physical Sciences > Dept of Statistical Science
URI: https://discovery-pp.ucl.ac.uk/id/eprint/10183782
Downloads since deposit
2,450Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item