Javascript must be enabled to continue!
Heteroscedastic-embedded Ensemble for Imbalanced Massive Data Classification
View through CrossRef
Abstract
The imbalanced learning methods aim to learn the unbiased models from massive class-imbalanced datasets. However, due to the uncertainty of data distributions affected by noise and borderline samples, the models based on heuristic assumptions or meta learning often suffer from the lack of stability and practicability for real-world classification tasks which often uses the skewed dataset of low-quality. In order to effectively cope with these issues, in this work, we propose a novel heteroscedastic-embedded ensemble (HEE) for imbalanced massive data classification. We first design an effective task sensing strategy with data-dependent heteroscedastic to adaptively guide the sampler’s focus on informative samples, which makes the learning method much more robust for noisy data. Then, similar to the learning style of human beings, i.e., easy to difficult, the HEE framework gradually builds a strong ensemble classifier through interactions between harmonizing data resampling and model-training. The simulation results and analysis on both synthetic and real-world tasks demonstrate the effectiveness, robustness, and transferability of the proposed method.
Title: Heteroscedastic-embedded Ensemble for Imbalanced Massive Data Classification
Description:
Abstract
The imbalanced learning methods aim to learn the unbiased models from massive class-imbalanced datasets.
However, due to the uncertainty of data distributions affected by noise and borderline samples, the models based on heuristic assumptions or meta learning often suffer from the lack of stability and practicability for real-world classification tasks which often uses the skewed dataset of low-quality.
In order to effectively cope with these issues, in this work, we propose a novel heteroscedastic-embedded ensemble (HEE) for imbalanced massive data classification.
We first design an effective task sensing strategy with data-dependent heteroscedastic to adaptively guide the sampler’s focus on informative samples, which makes the learning method much more robust for noisy data.
Then, similar to the learning style of human beings, i.
e.
, easy to difficult, the HEE framework gradually builds a strong ensemble classifier through interactions between harmonizing data resampling and model-training.
The simulation results and analysis on both synthetic and real-world tasks demonstrate the effectiveness, robustness, and transferability of the proposed method.
Related Results
Evaluasi Model Exponential Generelized Autoregressive Conditional Heteroscedastic (EGARCH)
Evaluasi Model Exponential Generelized Autoregressive Conditional Heteroscedastic (EGARCH)
Abstract. In time series data that has a fairly high volatility, it is possible to have an error variance that is not constant (Heteroscedasticity). This is reflected in the square...
Deep Neural Ensemble Classification for COVID-19 Dataset
Deep Neural Ensemble Classification for COVID-19 Dataset
The COVID-19 pandemic has necessitated the development of accurate and efficient classification models for diagnosis and prognosis. While deep learning has shown promising results ...
Selective Ensemble Learning Algorithm for Imbalanced Dataset
Selective Ensemble Learning Algorithm for Imbalanced Dataset
Abstract
Under the imbalanced dataset, the performance of the base-classifier, the computingmethod of weight of base-classifier and the selection method of the base-classif...
Improving Medical Document Classification via Feature Engineering
Improving Medical Document Classification via Feature Engineering
<p dir="ltr">Document classification (DC) is the task of assigning the predefined labels to unseen documents by utilizing the model trained on the available labeled documents...
Application of Machine Learning Techniques for Customer Churn Prediction in the Banking Sector
Application of Machine Learning Techniques for Customer Churn Prediction in the Banking Sector
Aim/Purpose: Previous studies have primarily focused on comparing predictive models without considering the impact of data preprocessing on model performance. Therefore, this study...
Informative prior on structural equation modelling with non-homogenous error structure
Informative prior on structural equation modelling with non-homogenous error structure
Introduction: This study investigates the impact of informative prior on Bayesian structural equation model (BSEM) with heteroscedastic error structure. A major drawback of homogen...
Informative prior on structural equation modelling with non-homogenous error structure
Informative prior on structural equation modelling with non-homogenous error structure
Introduction: This study investigates the impact of informative prior on Bayesian structural equation model (BSEM) with heteroscedastic error structure. A major drawback of homogen...
Weak tagging and imbalanced networks for online review sentiment classification
Weak tagging and imbalanced networks for online review sentiment classification
Sentiment classification aims to complete the automatic judgment task of text sentiment tendency. In the sentiment classification task of online reviews, traditional deep learning ...

