Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

Contrastive Distillation Learning with Sparse Spatial Aggregation

View through CrossRef
Abstract Contrastive learning has advanced significantly and demonstrates excellent transfer learning capabilities. Knowledge distillation is one of the most effective methods of model compression for computer vision. When combined with contrastive learning, it can achieve even better results. Current knowledge distillation techniques based on contrastive learning struggle to efficiently utilize the information from both student and teacher models, often missing out on optimizing the contrastive framework. This results in a less effective knowledge transfer process, limiting the potential improvements in model performance and representation quality. To address this limitation, we propose a new contrastive distillation learning method by redesigning the contrastive learning framework and incorporating sparse spatial aggregation. This method introduces a novel integration of feature alignment and spatial aggregation mechanism to enhance the learning process. It ensures that the representations obtained by the model fully capture the semantics of the original input. Compared to traditional unsupervised learning methods, our approach demonstrates superior performance in both pre-training and transfer learning. It achieves 71.6 Acc@1, 57.6 AP, 75.8 mIoU, 39.8/34.8 AP on ImageNet linear classification, Pascal VOC object detection, Cityscapes semantic segmentation, MS-COCO object detection and instance segmentation. Moreover, our method exhibits stable training and does not require large pre-training batch-sizes or numerous epochs.
Springer Science and Business Media LLC
Title: Contrastive Distillation Learning with Sparse Spatial Aggregation
Description:
Abstract Contrastive learning has advanced significantly and demonstrates excellent transfer learning capabilities.
Knowledge distillation is one of the most effective methods of model compression for computer vision.
When combined with contrastive learning, it can achieve even better results.
Current knowledge distillation techniques based on contrastive learning struggle to efficiently utilize the information from both student and teacher models, often missing out on optimizing the contrastive framework.
This results in a less effective knowledge transfer process, limiting the potential improvements in model performance and representation quality.
To address this limitation, we propose a new contrastive distillation learning method by redesigning the contrastive learning framework and incorporating sparse spatial aggregation.
This method introduces a novel integration of feature alignment and spatial aggregation mechanism to enhance the learning process.
It ensures that the representations obtained by the model fully capture the semantics of the original input.
Compared to traditional unsupervised learning methods, our approach demonstrates superior performance in both pre-training and transfer learning.
It achieves 71.
6 Acc@1, 57.
6 AP, 75.
8 mIoU, 39.
8/34.
8 AP on ImageNet linear classification, Pascal VOC object detection, Cityscapes semantic segmentation, MS-COCO object detection and instance segmentation.
Moreover, our method exhibits stable training and does not require large pre-training batch-sizes or numerous epochs.

Related Results

A Comprehensive Review of Distillation in the Pharmaceutical Industry
A Comprehensive Review of Distillation in the Pharmaceutical Industry
Distillation processes play a pivotal role in the pharmaceutical industry for the purification of active pharmaceutical ingredients (APIs), intermediates, and solvent recovery. Thi...
Improving Neural Retrieval with Contrastive Learning
Improving Neural Retrieval with Contrastive Learning
In recent years, neural retrieval models have shown remarkable progress in improving the efficiency and accuracy of information retrieval systems. However, challenges remain in eff...
Steam Distillation Studies For The Kern River Field
Steam Distillation Studies For The Kern River Field
Abstract The interactions of heavy oil and injected steam in the mature steamflood at the Kern River Field have been extensively studied to gain insight into the ...
The effect of Angelica Dahurica extracts on platelet aggregation
The effect of Angelica Dahurica extracts on platelet aggregation
Platelet aggregation is one of the important mechanisms in hemostasis. Improper platelet function may lead to bleeding or atherothrombosis. Angelica dahurica (AD) has been used in ...
Contrastive Instruction-Trajectory Learning for Vision-Language Navigation
Contrastive Instruction-Trajectory Learning for Vision-Language Navigation
The vision-language navigation (VLN) task requires an agent to reach a target with the guidance of natural language instruction. Previous works learn to navigate step-by-step follo...
Sparse Friendly Distillation Using Feature Decoupling
Sparse Friendly Distillation Using Feature Decoupling
Abstract In our paper, we introduce the sparse-friendly distillation framework as an effective training strategy for knowledge distillation. While model sparsity techniques...
An Asymmetric Contrastive Loss for Handling Imbalanced Datasets
An Asymmetric Contrastive Loss for Handling Imbalanced Datasets
Contrastive learning is a representation learning method performed by contrasting a sample to other similar samples so that they are brought closely together, forming clusters in t...
Initial Experience with Pediatrics Online Learning for Nonclinical Medical Students During the COVID-19 Pandemic 
Initial Experience with Pediatrics Online Learning for Nonclinical Medical Students During the COVID-19 Pandemic 
Abstract Background: To minimize the risk of infection during the COVID-19 pandemic, the learning mode of universities in China has been adjusted, and the online learning o...

Back to Top