Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

FacT: Factor-Tuning for Lightweight Adaptation on Vision Transformer

View through CrossRef
Recent work has explored the potential to adapt a pre-trained vision transformer (ViT) by updating only a few parameters so as to improve storage efficiency, called parameter-efficient transfer learning (PETL). Current PETL methods have shown that by tuning only 0.5% of the parameters, ViT can be adapted to downstream tasks with even better performance than full fine-tuning. In this paper, we aim to further promote the efficiency of PETL to meet the extreme storage constraint in real-world applications. To this end, we propose a tensorization-decomposition framework to store the weight increments, in which the weights of each ViT are tensorized into a single 3D tensor, and their increments are then decomposed into lightweight factors. In the fine-tuning process, only the factors need to be updated and stored, termed Factor-Tuning (FacT). On VTAB-1K benchmark, our method performs on par with NOAH, the state-of-the-art PETL method, while being 5x more parameter-efficient. We also present a tiny version that only uses 8K (0.01% of ViT's parameters) trainable parameters but outperforms full fine-tuning and many other PETL methods such as VPT and BitFit. In few-shot settings, FacT also beats all PETL baselines using the fewest parameters, demonstrating its strong capability in the low-data regime.
Association for the Advancement of Artificial Intelligence (AAAI)
Title: FacT: Factor-Tuning for Lightweight Adaptation on Vision Transformer
Description:
Recent work has explored the potential to adapt a pre-trained vision transformer (ViT) by updating only a few parameters so as to improve storage efficiency, called parameter-efficient transfer learning (PETL).
Current PETL methods have shown that by tuning only 0.
5% of the parameters, ViT can be adapted to downstream tasks with even better performance than full fine-tuning.
In this paper, we aim to further promote the efficiency of PETL to meet the extreme storage constraint in real-world applications.
To this end, we propose a tensorization-decomposition framework to store the weight increments, in which the weights of each ViT are tensorized into a single 3D tensor, and their increments are then decomposed into lightweight factors.
In the fine-tuning process, only the factors need to be updated and stored, termed Factor-Tuning (FacT).
On VTAB-1K benchmark, our method performs on par with NOAH, the state-of-the-art PETL method, while being 5x more parameter-efficient.
We also present a tiny version that only uses 8K (0.
01% of ViT's parameters) trainable parameters but outperforms full fine-tuning and many other PETL methods such as VPT and BitFit.
In few-shot settings, FacT also beats all PETL baselines using the fewest parameters, demonstrating its strong capability in the low-data regime.

Related Results

Automatic Load Sharing of Transformer
Automatic Load Sharing of Transformer
Transformer plays a major role in the power system. It works 24 hours a day and provides power to the load. The transformer is excessive full, its windings are overheated which lea...
ANALISIS PENGARUH MASA OPERASIONAL TERHADAP PENURUNAN KAPASITAS TRANSFORMATOR DISTRIBUSI DI PT PLN (PERSERO)
ANALISIS PENGARUH MASA OPERASIONAL TERHADAP PENURUNAN KAPASITAS TRANSFORMATOR DISTRIBUSI DI PT PLN (PERSERO)
One cause the interruption of transformer is loading that exceeds the capabilities of the transformer. The state of continuous overload will affect the age of the transformer and r...
LIFE CYCLE OF TRANSFORMER 110/X KV AND ITS VALUE
LIFE CYCLE OF TRANSFORMER 110/X KV AND ITS VALUE
In a deregulated environment, power companies are in the constant process of reducing the costs of operating power facilities, with the aim of optimally improving the quality of de...
PLC Based Load Sharing of Transformers
PLC Based Load Sharing of Transformers
The transformer is very expensive and bulky power system equipment. It runs and feed the load for 24 hours a day. Sometimes the load on the transformer unexpectedly rises above its...
Depth-aware salient object segmentation
Depth-aware salient object segmentation
Object segmentation is an important task which is widely employed in many computer vision applications such as object detection, tracking, recognition, and ret...
Simulation modeling study on short circuit ability of distribution transformer
Simulation modeling study on short circuit ability of distribution transformer
Abstract Under short circuit condition, the oil immersed distribution transformer will endure combined electro-thermal stress, eventually lead to the mechanical dama...
Vision-specific and psychosocial impacts of low vision among patients with low vision at the eastern regional Low Vision Centre
Vision-specific and psychosocial impacts of low vision among patients with low vision at the eastern regional Low Vision Centre
Purpose: To determine vision-specific and psychosocial implications of low vision among patients with low vision visiting the Low Vision Centre of the Eastern Regional Hospital in ...
Penetrability of lightweight aggregate concrete
Penetrability of lightweight aggregate concrete
In this paper, a study was conducted on the parameters affecting the sorption of water, chloride ingress and permeability of water into lightweight aggregate concrete. The paramete...

Back to Top