Javascript must be enabled to continue!
Label Ranker: Self-Aware Preference for Classification Label Position in Visual Masked Self-Supervised Pre-Trained Model
View through CrossRef
This paper investigates the impact of randomly initialized unique encoding of classification label position on the visual masked self-supervised pre-trained model when fine-tuning downstream classification tasks. Our findings indicate that different random initializations lead to significant variations in fine-tuned results, even when using the same allocation strategy for classification datasets. The accuracy gap between these results suggests that the visual masked self-supervised pre-trained model has an inherent preference for classification label positions. To investigate this, we compare it with the non-self-supervised visual pre-trained model and hypothesize that the masked self-supervised model exhibits a self-aware bias toward certain label positions. To mitigate the instability caused by random encoding, we propose a classification label position ranking algorithm, Label Ranker. It is based on 1-D dimensionality reduction of feature maps using Linear Discriminant Analysis and position-rank encoding of them by unsupervised feature clustering using the similarity property of Euclidean distance. This algorithm ensures that label position encoding align with the model is inherent preference. Extensive ablation experiments using ImageMAE and VideoMAE models on the CIFAR-100, UCF101, and HMDB51 classification datasets validate our approach. Results demonstrate that our method effectively stabilizes classification label position encoding, improving fine-tuned performance for visual masked self-supervised models.
Title: Label Ranker: Self-Aware Preference for Classification Label Position in Visual Masked Self-Supervised Pre-Trained Model
Description:
This paper investigates the impact of randomly initialized unique encoding of classification label position on the visual masked self-supervised pre-trained model when fine-tuning downstream classification tasks.
Our findings indicate that different random initializations lead to significant variations in fine-tuned results, even when using the same allocation strategy for classification datasets.
The accuracy gap between these results suggests that the visual masked self-supervised pre-trained model has an inherent preference for classification label positions.
To investigate this, we compare it with the non-self-supervised visual pre-trained model and hypothesize that the masked self-supervised model exhibits a self-aware bias toward certain label positions.
To mitigate the instability caused by random encoding, we propose a classification label position ranking algorithm, Label Ranker.
It is based on 1-D dimensionality reduction of feature maps using Linear Discriminant Analysis and position-rank encoding of them by unsupervised feature clustering using the similarity property of Euclidean distance.
This algorithm ensures that label position encoding align with the model is inherent preference.
Extensive ablation experiments using ImageMAE and VideoMAE models on the CIFAR-100, UCF101, and HMDB51 classification datasets validate our approach.
Results demonstrate that our method effectively stabilizes classification label position encoding, improving fine-tuned performance for visual masked self-supervised models.
Related Results
Persons and Their Private Personas: Living with Yourself
Persons and Their Private Personas: Living with Yourself
Public life is usually understood to be whatever we do or say in our formal and professional relationships. At the workplace, at the doctor’s office or at the café, we need to make...
The many faces of a text : applications and enhancements of multi-label text classification algorithms
The many faces of a text : applications and enhancements of multi-label text classification algorithms
Multi-Label Text Classification (MLTC) is a challenging yet vital component of analyzing large text collections. The aim of MLTC is to assign one or multiple labels to a text, whic...
Afaan Oromo Multi-Label News Text Classification Using Deep Learning Approach
Afaan Oromo Multi-Label News Text Classification Using Deep Learning Approach
Abstract
Classification is a technique for categorizing textual data into a form of predefined categories. Due to its major consequences in regard to critical tasks such as...
Hubungan Pengetahuan terkait Label Gizi dengan Kebiasaan Membaca Label Gizi pada Siswa SMA Al-Islam
Hubungan Pengetahuan terkait Label Gizi dengan Kebiasaan Membaca Label Gizi pada Siswa SMA Al-Islam
Latar Belakang: Masih sedikit konsumen yang dapat memahami dan menggunakan label gizi sesuai dengan fungsinya. Hal ini dikarenakan masih rendahnya kesadaran masyarakat terkait pent...
Dynamically Masked Audiograms With Machine Learning Audiometry
Dynamically Masked Audiograms With Machine Learning Audiometry
Objectives:
When one ear of an individual can hear significantly better than the other ear, evaluating the worse ear with loud probe tones may require delivering maskin...
A Semi-supervised Object Detection Learning Method under Queue Smoothing Pseudo-label Supervising and Embedding Consistency Constraint
A Semi-supervised Object Detection Learning Method under Queue Smoothing Pseudo-label Supervising and Embedding Consistency Constraint
Abstract
Semi-supervised object detection is an effective solution to balance the manual annotation cost and model performance in practical application. However, two major ...
Uncertainty-Aware Graph Neural Network for Semi-Supervised Diversified Recommendation
Uncertainty-Aware Graph Neural Network for Semi-Supervised Diversified Recommendation
Abstract
Graphs are a powerful tool for representing structured and relational data in various domains, including social networks, knowledge graphs, and molecular structure...
DIMENSIONS OF OCCUPATIONAL PREFERENCE
DIMENSIONS OF OCCUPATIONAL PREFERENCE
ABSTRACTThis study was undertaken to: (a) determine the dimensionality of the occupational preferences of college students, and (b) relate other measures of preference and ability ...

