Switching: understanding the class-reversed sampling in tail sample memorization

记忆 一般化 计算机科学 采样(信号处理) 人工智能 代表(政治) 班级(哲学) 机器学习 样品(材料) 过程(计算) 模式识别(心理学) 数学 滤波器(信号处理) 操作系统 数学分析 政治 数学教育 色谱法 化学 法学 计算机视觉 政治学
作者
Chi Zhang,Benyi Hu,Yuhang Liuzhang,Le Wang,Li Liu,Yuehu Liu
出处
期刊:Machine Learning [Springer Nature]
卷期号:111 (3): 1073-1101
标识
DOI:10.1007/s10994-021-06087-3
摘要

Long-tailed visual recognition poses significant challenges to traditional machine learning and emerging deep networks due to its inherent class imbalance. Existing reweighting and re-sampling methods, although effective, lack a fundamental theory while leaving the paradoxical effects of long tail unsolved, where network failing with head classes under-represented and tail classes overfitted. In this paper, we investigate long-tailed recognition from a memorization-generalization point of view, which not only unravels the whys of previous methods, but also derives a new principled solution. Specifically, we first empirically identify the regularity of classes under long-tailed distributions, finding that long-tailed challenge is essentially a trade-off between the representation of high-regularity head classes and generalization to low-regularity tail classes. To memorize tail samples without seriously damaging the representation of head samples, we propose a simple yet effective sampling strategy for ordinary mini-batch SGD optimization process, Switching, which switches from instance-balanced sampling to class-reversed sampling for only once at small learning rate. By theoretical analysis, we show that the upper bound on the generalization error of the proposed sampling strategy is lower than instance-balanced sampling conditionally. In our experiments, the proposed method can reach feasible performance more efficiently than current methods. Further experiments validate the superiority of the proposed Switching strategy, implying that the long-tailed learning trade-off could be parsimoniously tackled only in the memorization stage with a small learning rate and over-exposure of tail samples.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
更新
大幅提高文件上传限制,最高150M (2024-4-1)

科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
刚刚
情怀应助lily采纳,获得10
3秒前
Hakunamax发布了新的文献求助10
3秒前
5秒前
youbin完成签到 ,获得积分10
5秒前
玄同发布了新的文献求助10
6秒前
_xySH完成签到 ,获得积分10
7秒前
高高如风发布了新的文献求助10
8秒前
香蕉觅云应助认真的世倌采纳,获得10
8秒前
领导范儿应助albert采纳,获得10
9秒前
11秒前
隐形鸣凤发布了新的文献求助20
11秒前
聪明的棉花糖完成签到 ,获得积分10
14秒前
14秒前
大模型应助愤怒的自我采纳,获得10
17秒前
23秒前
自然的绮梅完成签到 ,获得积分10
26秒前
优秀从凝完成签到,获得积分10
28秒前
六十完成签到,获得积分10
28秒前
Han完成签到 ,获得积分10
35秒前
Hello应助科研通管家采纳,获得30
36秒前
烟花应助科研通管家采纳,获得10
36秒前
36秒前
传奇3应助科研通管家采纳,获得10
36秒前
36秒前
39秒前
39秒前
40秒前
小小王完成签到 ,获得积分10
42秒前
lucy完成签到,获得积分10
43秒前
44秒前
44秒前
49秒前
49秒前
欣喜的机器猫完成签到,获得积分20
49秒前
ywongmath完成签到,获得积分10
51秒前
小螃蟹完成签到,获得积分10
52秒前
科研修沟发布了新的文献求助10
54秒前
乔心发布了新的文献求助10
55秒前
57秒前
高分求助中
The three stars each : the Astrolabes and related texts 1070
Manual of Clinical Microbiology, 4 Volume Set (ASM Books) 13th Edition 1000
Sport in der Antike 800
De arte gymnastica. The art of gymnastics 600
少脉山油柑叶的化学成分研究 530
Sport in der Antike Hardcover – March 1, 2015 500
Boris Pesce - Gli impiegati della Fiat dal 1955 al 1999 un percorso nella memoria 500
热门求助领域 (近24小时)
化学 材料科学 医学 生物 有机化学 工程类 生物化学 纳米技术 物理 内科学 计算机科学 化学工程 复合材料 遗传学 基因 物理化学 催化作用 电极 光电子学 量子力学
热门帖子
关注 科研通微信公众号,转发送积分 2405772
求助须知:如何正确求助?哪些是违规求助? 2103798
关于积分的说明 5310313
捐赠科研通 1831301
什么是DOI,文献DOI怎么找? 912494
版权声明 560646
科研通“疑难数据库(出版商)”最低求助积分说明 487860