Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

The robustness of popular multiclass machine learning models against poisoning attacks: Lessons and insights

View through CrossRef
Despite the encouraging outcomes of machine learning and artificial intelligence applications, the safety of artificial intelligence–based systems is one of the most severe challenges that need further exploration. Data set poisoning is a severe problem that may lead to the corruption of machine learning models. The attacker injects data into the data set that are faulty or mislabeled by flipping the actual labels into the incorrect ones. The word “robustness” refers to a machine learning algorithm’s ability to cope with hostile situations. Here, instead of flipping the labels randomly, we use the clustering approach to choose the training samples for label changes to influence the classifiers’ performance and the distance-based anomaly detection capacity in quarantining the poisoned samples. According to our experiments on a benchmark data set, random label flipping may have a short-term negative impact on the classifier’s accuracy. Yet, an anomaly filter would discover on average 63% of them. On the contrary, the proposed clustering-based flipping might inject dormant poisoned samples until the number of poisoned samples is enough to influence the classifiers’ performance severely; on average, the same anomaly filter would discover 25% of them. We also highlight important lessons and observations during this experiment about the performance and robustness of popular multiclass learners against training data set–poisoning attacks that include: trade-offs, complexity, categories, poisoning resistance, and hyperparameter optimization.
Title: The robustness of popular multiclass machine learning models against poisoning attacks: Lessons and insights
Description:
Despite the encouraging outcomes of machine learning and artificial intelligence applications, the safety of artificial intelligence–based systems is one of the most severe challenges that need further exploration.
Data set poisoning is a severe problem that may lead to the corruption of machine learning models.
The attacker injects data into the data set that are faulty or mislabeled by flipping the actual labels into the incorrect ones.
The word “robustness” refers to a machine learning algorithm’s ability to cope with hostile situations.
Here, instead of flipping the labels randomly, we use the clustering approach to choose the training samples for label changes to influence the classifiers’ performance and the distance-based anomaly detection capacity in quarantining the poisoned samples.
According to our experiments on a benchmark data set, random label flipping may have a short-term negative impact on the classifier’s accuracy.
Yet, an anomaly filter would discover on average 63% of them.
On the contrary, the proposed clustering-based flipping might inject dormant poisoned samples until the number of poisoned samples is enough to influence the classifiers’ performance severely; on average, the same anomaly filter would discover 25% of them.
We also highlight important lessons and observations during this experiment about the performance and robustness of popular multiclass learners against training data set–poisoning attacks that include: trade-offs, complexity, categories, poisoning resistance, and hyperparameter optimization.

Related Results

Selection of Injectable Drug Product Composition using Machine Learning Models (Preprint)
Selection of Injectable Drug Product Composition using Machine Learning Models (Preprint)
BACKGROUND As of July 2020, a Web of Science search of “machine learning (ML)” nested within the search of “pharmacokinetics or pharmacodynamics” yielded over 100...
Manipulating Recommender Systems: A Survey of Poisoning Attacks and Countermeasures
Manipulating Recommender Systems: A Survey of Poisoning Attacks and Countermeasures
Recommender systems have become an integral part of online services due to their ability to help users locate specific information in a sea of data. However, existing studies show ...
Abnormal Brain Functional Network Dynamics in Acute CO Poisoning
Abnormal Brain Functional Network Dynamics in Acute CO Poisoning
Aims: Carbon monoxide poisoning is a common condition that can cause severe neurological sequelae. Previous studies have revealed that functional connectivity in carbon monoxide po...
Overview of acute Chinese medicine poisoning in Hong Kong
Overview of acute Chinese medicine poisoning in Hong Kong
Abstract Background Chinese medicine (CM) poisoning is relatively rare in Hong Kong. According to the Department of Healt...
Impacting Robustness in Deep Learning-Based NIDS through Poisoning Attacks
Impacting Robustness in Deep Learning-Based NIDS through Poisoning Attacks
The rapid expansion and pervasive reach of the internet in recent years have raised concerns about evolving and adaptable online threats, particularly with the extensive integratio...
Deception-Based Security Framework for IoT: An Empirical Study
Deception-Based Security Framework for IoT: An Empirical Study
<p><b>A large number of Internet of Things (IoT) devices in use has provided a vast attack surface. The security in IoT devices is a significant challenge considering c...
Using animal tracking for early detection of mass poisoning events
Using animal tracking for early detection of mass poisoning events
Abstract 1 Amidst the sixth mass extinction, some groups, such as vultures, the only obligate scavengers among vertebrates, are disappearing at an unprecedented rat...

Back to Top