Questions For Contrastive Learning : "What Makes?" (Representation Learning Of Images Summer 2020 Feature 4)
3 main points
✔️ Pursuing the Conditions of the View to Improve the Performance of Contrastive Learning
✔️ Pursuing what kind of information is included in a representation that is useful for downstream tasks
✔️ Get to the bottom of whether InfoMax is really useful
What makes for good views for contrastive learning
written by Yonglong Tian, Chen Sun, Ben Poole, Dilip Krishnan, Cordelia Schmid, Phillip Isola
(Submitted on 20 May 2020)
Comments: Accepted at ECCV2020
Subjects: Computer Vision and Pattern Recognition (cs.CV)
Paper Official Code COMM Code
What makes instance discrimination good for transfer learning?
written by Nanxuan Zhao, Zhirong Wu, Rynson W.H. Lau, Stephen Lin
(Submitted on 11 Jun 2020)
Comments: Published by arXiv
Subjects: Computer Vision and Pattern Recognition (cs.CV)
Paper Official Code COMM Code
On Mutual Information Maximization for Representation Learning
written by Michael Tschannen, Josip Djolonga, Paul K. Rubenstein, Sylvain Gelly, Mario Lucic
(Submitted on 31 Jul 2019 (v1), last revised 23 Jan 2020 (this version, v2))
Comments: Accepted at ICLR2020
Subjects: Computer Vision and Pattern Recognition (cs.CV)
Paper Official Code Colab Code
The writer's special project entitled "Learning to Express Images Summer 2020" introduces various methods of unsupervised learning.
Part 1. Image GPT for domain knowledge-free and unsupervised learning, and image generation is amazing!
Part 2. Contrastive Learning's Two Leading Methods SimCLR And MoCo, And The Evolution Of Each
Part 3. SOTA With Contrastive Learning And Clustering!
Part 4. Questions For Contrastive Learning : "What Makes?"
Part 5. The Versatile And Practical DeepMind Unsupervised Learning Method
Having survived two AI winters and gaining expressive power with the massive image dataset ImageNet, AI in images blossomed in 2012 in a big way. However, this required significant costs for the human labeling of images. In contrast, BERT, which made such a huge social impact in 2018 that natural language processing became a concern for fake news, is also a major feature of the vast amount of data available as it is.
Contrastive learning is a form of unsupervised learning that uses a mechanism for comparing data to each other instead of costly labeling and is capable of training large amounts of data as is. It has been successfully applied to images and has already surpassed the performance of ImageNet-trained models and, like BERT, is expected to have a future impact in the imaging field.
So far, in Part 2 and Part 3, I have focused on these four methods, SimCLR, MoCo, PCL, and SwAV. We have been able to show high performance with each of them, but the big question is why the performance is so good, which is not clear.
So, to conclude Contrastive Learning, I'd like to explore this question in a paper titled "What makes", which begins with the title "Why is performance so good?
- What makes for good views for contrastive learning is a study that questioned the conditions for generating good views and led to performance improvements.
- In What makes instance discrimination good for transfer learning? we look at why the individual discrimination task improves the performance of the downstream task.
- On Mutual Information Maximization for Representation Learning, this paper introduces the fundamental question.
To read more,
Please register with AI-SCHOLAR.
ORCategories related to this article