Javascript must be enabled to continue!
AdaPT: Adaptive Position Trigger for Improving Backdoors Attacks in Transfer Learning
View through CrossRef
Backdoor attacks in neural networks have emerged as one of the most critical and dangerous threats to AI security, attracting extensive research attention in recent years. Most existing backdoor attacks operate within an end-to-end learning framework. These attacks can achieve nearly 100% success rates on testing set while poisoning less than 10% of the training set in certain datasets. However, with the accumulation of public datasets and widespread adoption of pre-trained models, practical applications now commonly employ a transfer learning framework where features are first extracted by pre-trained models before training the classification head. Our investigation reveals that this transfer learning framework significantly degrades the effectiveness of traditional end-to-end backdoor attacks. To evaluate the security of such transfer learning networks using pre-trained models, new backdoor attack methods need to be designed. This paper analyze the failure mechanisms of traditional attacks from the perspective of the hidden feature discrepancy. Based on these insights, we propose AdaPT (Adaptive Position Trigger), a novel backdoor attack method that automatically searches the optimal trigger insertion position by maximizing the hidden feature discrepancy. Extensive experiments demonstrate that AdaPT not only enhances attack success rates in transfer learning and improves trigger learning efficiency in end-to-end learning, but also can resist typical backdoor defenses.
Title: AdaPT: Adaptive Position Trigger for Improving Backdoors Attacks in Transfer Learning
Description:
Backdoor attacks in neural networks have emerged as one of the most critical and dangerous threats to AI security, attracting extensive research attention in recent years.
Most existing backdoor attacks operate within an end-to-end learning framework.
These attacks can achieve nearly 100% success rates on testing set while poisoning less than 10% of the training set in certain datasets.
However, with the accumulation of public datasets and widespread adoption of pre-trained models, practical applications now commonly employ a transfer learning framework where features are first extracted by pre-trained models before training the classification head.
Our investigation reveals that this transfer learning framework significantly degrades the effectiveness of traditional end-to-end backdoor attacks.
To evaluate the security of such transfer learning networks using pre-trained models, new backdoor attack methods need to be designed.
This paper analyze the failure mechanisms of traditional attacks from the perspective of the hidden feature discrepancy.
Based on these insights, we propose AdaPT (Adaptive Position Trigger), a novel backdoor attack method that automatically searches the optimal trigger insertion position by maximizing the hidden feature discrepancy.
Extensive experiments demonstrate that AdaPT not only enhances attack success rates in transfer learning and improves trigger learning efficiency in end-to-end learning, but also can resist typical backdoor defenses.
Related Results
CSP beyond tractable constraint languages
CSP beyond tractable constraint languages
AbstractThe constraint satisfaction problem (CSP) is among the most studied computational problems. While NP-hard, many tractable subproblems have been identified (Bulatov 2017, Zh...
Deception-Based Security Framework for IoT: An Empirical Study
Deception-Based Security Framework for IoT: An Empirical Study
<p><b>A large number of Internet of Things (IoT) devices in use has provided a vast attack surface. The security in IoT devices is a significant challenge considering c...
Backdoors to Planning
Backdoors to Planning
Backdoors measure the distance to tractable fragments and have become an important tool to find fixed-parameter tractable (fpt) algorithms. Despite their success, backdoors have no...
Procedure for Western blot v1
Procedure for Western blot v1
Goal: This document has the objective of standardizing the protocol for Western blot. This technique allows the detection of specific proteins separated on polyacrylamide gel and t...
Manipulating Recommender Systems: A Survey of Poisoning Attacks and Countermeasures
Manipulating Recommender Systems: A Survey of Poisoning Attacks and Countermeasures
Recommender systems have become an integral part of online services due to their ability to help users locate specific information in a sea of data. However, existing studies show ...
Learning and adaptive management mechanisms in implementing development programs in Zaka district, Zimbabwe
Learning and adaptive management mechanisms in implementing development programs in Zaka district, Zimbabwe
The aim of the study was to investigate the application and importance of Learning and Adaptive Management by the Ministry of Women Affairs, Community, Small and Medium Enterprises...
Comprehensive Analysis of Cyber-Manufacturing Attacks Using a Cyber-Manufacturing Testbed
Comprehensive Analysis of Cyber-Manufacturing Attacks Using a Cyber-Manufacturing Testbed
Abstract
Cyber-Manufacturing Systems (CMS) are vulnerable to cyber-manufacturing attacks ironically because of its very beneficial advance: seamless integration with...
he prevalence and clinical presentation of fibularis myofascial trigger points in the assessment and treatment of inversion ankle sprains
he prevalence and clinical presentation of fibularis myofascial trigger points in the assessment and treatment of inversion ankle sprains
Ankle sprains account for 85% of all injuries to the ankle (Garrick, 1997). Inversion sprains result from a twisting of a weight-bearing foot into a plantarflexed and inverted posi...

