Invariant Feature Learning for Generalized Long-Tailed Classification
ECCV 2022 (TEL AVIV 2022)
- Kaihua Tang
- Mingyuan Tao
- Jiaxin Qi
- Zhenguang Liu
- Hanwang Zhang
- Nanyang Technological University
- Damo Academy, Alibaba Group
- Zhejiang University
Abstract
Existing long-tailed classification (LT) methods only focus on tackling the class-wise imbalance that head classes have more samples than tail classes, but overlook the attribute-wise imbalance. In fact, even if the class is balanced, samples within each class may still be long-tailed due to the varying attributes. Note that the latter is fundamentally more ubiquitous and challenging than the former because attributes are not just implicit for most datasets, but also combinatorially complex, thus prohibitively expensive to be balanced. Therefore, we introduce a novel research problem: Generalized Long-Tailed classification (GLT), to jointly consider both kinds of imbalances. By "generalized", we mean that a GLT method should naturally solve the traditional LT, but not vice versa. Not surprisingly, we find that most class-wise LT methods degenerate in our proposed two benchmarks: ImageNet-GLT and MSCOCO-GLT. We argue that it is because they over-emphasize the adjustment of class distribution while neglecting to learn attribute-invariant features. To this end, we propose an Invariant Feature Learning (IFL) method as the first strong baseline for GLT. IFL first discovers environments with divergent intra-class distributions from the imperfect predictions and then learns invariant features across them. Promisingly, as an improved feature backbone, IFL boosts all the LT line-up: one/two-stage re-balance, augmentation, and ensemble.
Poster
Short Presentation (5min) [Slides]
Citation
Acknowledgements
This project is partially supported by Alibaba-NTU Singapore Joint Research Institute (JRI), and AI
Singapore (AISG) Research Programme. We also feel grateful to the computational resources provided
by Damo Academy.
The website template is borrowed from Mip-NeRF and Text2Light.