Classification

RaSEn R package

We propose a new model-free ensemble classification framework, RaSE algorithm, for the sparse classification problem. In RaSE algorithm, for each weak learner, some random subspaces are generated and the optimal one is chosen to train the model on the basis of some criterion. To be adapted to the problem, a novel criterion, ratio information criterion (RIC) is put up with based on Kullback-Leibler divergence. Besides minimizing RIC, multiple criteria can be applied, for instance, minimizing extended Bayesian information criterion (eBIC), minimizing training error, minimizing the validation error, minimizing the cross-validation error, minimizing leave-one-out error.

RaSE: Random Subspace Ensemble Classification

We propose a new model-free ensemble classification framework, Random Subspace Ensemble (RaSE), for sparse classification. In the RaSE algorithm, we aggregate many weak learners, where each weak learner is a base classifier trained in a subspace …

Imbalanced classification: an objective-oriented review

A common issue for classification in scientific research and industry is theexistence of imbalanced classes. When sample sizes of different classes areimbalanced in training data, naively implementing a classification method oftenleads to …

Neyman-Pearson classification: parametrics and sample size requirement

In contrast to the classical binary classification paradigm that minimizes the overall classification error, the Neyman-Pearson (NP) paradigm seeks classifiers with a minimal type II error while having a constrained type I error under a …

Neyman-Pearson classification algorithms and NP receiver operating characteristics

In many binary classification applications, such as disease diagnosis and spam detection, practitioners commonly face the need to limit type I error (that is, the conditional probability of misclassifying a class 0 observation as class 1) so that it …

JDINAC: joint density-based non-parametric differential interaction network analysis and classification using high-dimensional sparse omics data

Motivation A complex disease is usually driven by a number of genes interwoven into networks, rather than a single gene product. Network comparison or differential network analysis has become an important means of revealing the underlying mechanism …

A survey on Neyman-Pearson classification and suggestions for future research

In statistics and machine learning, classification studies how to automatically learn to make good qualitative predictions (i.e., assign class labels) based on past observations. Examples of classification problems include email spam filtering, fraud …

Feature Augmentation via Nonparametrics and Selection (FANS) in High Dimensional Classification

We propose a high-dimensional classification method that involves nonparametric feature augmentation. Knowing that marginal density ratios are the most powerful univariate classifiers, we use the ratio estimates to transform the original feature …

Neyman-Pearson Classification under High-Dimensional Settings

Most existing binary classification methods target on the optimization of the overall classification risk and may fail to serve some real-world applications such as cancer diagnosis, where users are more concerned with the risk of misclassifying one …

A road to classification in high dimensional space: the regularized optimal affine discriminant

For high dimensional classification, it is well known that naively performing the Fisher discriminant rule leads to poor results due to diverging spectra and accumulation of noise. Therefore, researchers proposed independence rules to circumvent the …