判别式
计算机科学
选择(遗传算法)
一般化
机器学习
样品(材料)
人工智能
选择偏差
采样(信号处理)
集合(抽象数据类型)
编码(集合论)
取样偏差
班级(哲学)
噪音(视频)
数据挖掘
训练集
样本量测定
统计
数学
化学
色谱法
数学分析
滤波器(信号处理)
图像(数学)
计算机视觉
程序设计语言
作者
Qi Wei,Lei Feng,Haobo Wang,Bo An
出处
期刊:Cornell University - arXiv
日期:2024-01-01
标识
DOI:10.48550/arxiv.2401.13360
摘要
Learning with noisy labels aims to ensure model generalization given a label-corrupted training set. The sample selection strategy achieves promising performance by selecting a label-reliable subset for model training. In this paper, we empirically reveal that existing sample selection methods suffer from both data and training bias that are represented as imbalanced selected sets and accumulation errors in practice, respectively. However, only the training bias was handled in previous studies. To address this limitation, we propose a noIse-Tolerant Expert Model (ITEM) for debiased learning in sample selection. Specifically, to mitigate the training bias, we design a robust network architecture that integrates with multiple experts. Compared with the prevailing double-branch network, our network exhibits better performance of selection and prediction by ensembling these experts while training with fewer parameters. Meanwhile, to mitigate the data bias, we propose a mixed sampling strategy based on two weight-based data samplers. By training on the mixture of two class-discriminative mini-batches, the model mitigates the effect of the imbalanced training set while avoiding sparse representations that are easily caused by sampling strategies. Extensive experiments and analyses demonstrate the effectiveness of ITEM. Our code is available at this url \href{https://github.com/1998v7/ITEM}{ITEM}.
科研通智能强力驱动
Strongly Powered by AbleSci AI