Published on:
8 November 2023
Primary Category:
Computer Vision and Pattern Recognition
Paper Authors:
Quentin Bouniot
Enforces assumptions from multi-task representation learning theory to improve meta-learning algorithms
Introduces localization information in contrastive learning objectives for pretraining object detectors
Shows improved generalization and sample efficiency when training models with few annotated images
Proposes regularization techniques to guide meta-learning algorithms
Leverages object proposals in images as positives examples for contrastive learning
Learning from few images for image recognition
This paper proposes methods to learn image recognition models more efficiently from limited annotated data. The authors develop techniques in meta-learning and transfer learning to improve model generalization and sample efficiency when training with few images per class. Key ideas include enforcing assumptions from multi-task representation learning theory in meta-learning algorithms, and introducing localization information in contrastive learning objectives for pretraining object detectors.
No comments yet, be the first to start the conversation...
Sign up to comment on this paper