Contents

0.1 Installation

Typically, each feature selection method or classifier originates from a different R package, which ClassifyR provides a wrapper around. By default, only high-performance t-test/F-test and random forest are installed. If you intend to compare between numerous different modelling methods, you should install all suggested packages at once by using the command BiocManager::install("ClassifyR", dependencies = TRUE). This will take a few minutes, particularly on Linux, because each package will be compiled from source code.

0.2 Overview

ClassifyR provides a structured pipeline for cross-validated classification. Classification is viewed in terms of four stages, data transformation, feature selection, classifier training, and prediction. The driver functions crossValidate and runTests implements varieties of cross-validation. They are:

Driver functions can use parallel processing capabilities in R to speed up cross-validations when many CPUs are available. The output of the driver functions is a ClassifyResult object which can be directly used by the performance evaluation functions. The process of classification is summarised by a flowchart.

Importantly, ClassifyR implements a number of methods for classification using different kinds of changes in measurements between classes. Most classifiers work with features where the means are different. In addition to changes in means (DM), ClassifyR also allows for classification using differential variability (DV; changes in scale) and differential distribution (DD; changes in location and/or scale).

0.2.1 Case Study: Diagnosing Asthma

To demonstrate some key features of ClassifyR, a data set consisting of the 2000 most variably expressed genes and 190 people will be used to quickly obtain results. The journal article corresponding to the data set was published in Scientific Reports in 2018 and is titled A Nasal Brush-based Classifier of Asthma Identified by Machine Learning Analysis of Nasal RNA Sequence Data.

Load the package.

library(ClassifyR)

A glimpse at the RNA measurements and sample classes.

data(asthma) # Contains measurements and classes variables.
measurements[1:5, 1:5]
##            HBB BPIFA1  XIST FCGR3B HBA2
## Sample 1  9.72  14.06 12.28  11.42 7.83
## Sample 2 11.98  13.89  6.35  13.25 9.42
## Sample 3 12.15  17.44 10.21   7.87 9.68
## Sample 4 10.60  11.87  6.27  14.75 8.96
## Sample 5  8.18  15.01 11.21   6.77 6.43
head(classes)
## [1] No  No  No  No  Yes No 
## Levels: No Yes

The numeric matrix variable measurements stores the normalised values of the RNA gene abundances for each sample and the factor vector classes identifies which class the samples belong to. The measurements were normalised using DESeq2’s varianceStabilizingTransformation function, which produces \(log_2\)-like data.

For more complex data sets with multiple kinds of experiments (e.g. DNA methylation, copy number, gene expression on the same set of samples) a MultiAssayExperiment is recommended for data storage and supported by ClassifyR’s methods.

0.3 Quick Start: crossValidate Function

The crossValidate function offers a quick and simple way to start analysing a dataset in ClassifyR. It is a wrapper for runTests, the core model building and testing function of ClassifyR. crossValidate must be supplied with measurements, a simple tabular data container or a list-like structure of such related tabular data on common samples. The classes of it may be matrix, data.frame, DataFrame, MultiAssayExperiment or list of data.frames. For a dataset with \(n\) observations and \(p\) variables, the crossValidate function will accept inputs of the following shapes:

Data Type \(n \times p\) \(p \times n\)
matrix
data.frame
DataFrame
MultiAssayExperiment
list of data.frames

crossValidate must also be supplied with outcome, which represents the prediction to be made in a variety of possible ways.

ClassifyR enforces reproducible research. Before beginning any cross-validation, the random number seed needs to be set in the current R session. Any number is fine.

set.seed(9500)

The type of classifier used can be changed with the classifier argument. The default is a random forest, which seamlessly handles categorical and numerical data. A full list of classifiers can be seen by running ?crossValidate. A feature selection step can be performed before classification using nFeatures and selectionMethod, which is a t-test by default. Similarly, the number of folds and number of repeats for cross validation can be changed with the nFolds and nRepeats arguments. If wanted, nCores can be specified to run the cross validation in parallel. To perform 5-fold cross-validation of a Support Vector Machine with 2 repeats:

result <- crossValidate(measurements, classes, classifier = "SVM",
                        nFeatures = 20, nFolds = 5, nRepeats = 2, nCores = 1)
performancePlot(result)
## Warning in .local(results, ...): Balanced Accuracy not found in all elements of results. Calculating it now.

0.3.1 Data Integration with crossValidate

crossValidate also allows data from multiple sources to be integrated into a single model. The integration method can be specified with multiViewMethod argument. In this example, suppose the first 10 variables in the asthma data set are from a certain source and the remaining 1990 variables are from a second source. To integrate multiple data sets, each variable must be labeled with the data set it came from. This is done in a different manner depending on the data type of measurements.

If using Bioconductor’s DataFrame, this can be specified using mcols. In the column metadata, each feature must have an assay and a feature name.

measurementsDF <- DataFrame(measurements)
mcols(measurementsDF) <- data.frame(
  assay = rep(c("assay_1", "assay_2"), times = c(10, 1990)),
  feature = colnames(measurementsDF)
)

result <- crossValidate(measurementsDF, classes, classifier = "SVM", nFolds = 5,
                        nRepeats = 3, multiViewMethod = "merge")

performancePlot(result, characteristicsList = list(x = "Assay Name"))
## Warning in .local(results, ...): Balanced Accuracy not found in all elements of results. Calculating it now.

If using a list of data.frames, the name of each element in the list will be used as the assay name.

# Assigns first 10 variables to dataset_1, and the rest to dataset_2
measurementsList <- list(
  (measurements |> as.data.frame())[1:10],
  (measurements |> as.data.frame())[11:2000]
)
names(measurementsList) <- c("assay_1", "assay_2")

result <- crossValidate(measurementsList, classes, classifier = "SVM", nFolds = 5,
                        nRepeats = 3, multiViewMethod = "merge")

performancePlot(result, characteristicsList = list(x = "Assay Name"))
## Warning in .local(results, ...): Balanced Accuracy not found in all elements of results. Calculating it now.

0.4 A More Detailed Look at ClassifyR

In the following sections, some of the most useful functions provided in ClassifyR will be demonstrated. However, a user could wrap any feature selection, training, or prediction function to the classification framework, as long as it meets some simple rules about the input and return parameters. See the appendix section of this guide titled “Rules for New Functions” for a description of these.

0.4.1 Comparison to Existing Classification Frameworks

There are a few other frameworks for classification in R. The table below provides a comparison of which features they offer.

Package Run User-defined Classifiers Parallel Execution on any OS Parameter Tuning Intel DAAL Performance Metrics Ranking and Selection Plots Class Distribution Plot Sample-wise Error Heatmap Direct Support for MultiAssayExperiment Input
ClassifyR Yes Yes Yes Yes Yes Yes Yes Yes
caret Yes Yes Yes No No No No No
MLInterfaces Yes No No No No No No No
MCRestimate Yes No Yes No No No No No
CMA No No Yes No No No No No

0.4.2 Provided Functionality

Although being a cross-validation framework, a number of popular feature selection and classification functions are provided by the package which meet the requirements of functions to be used by it (see the last section).

0.4.2.1 Provided Methods for Feature Selection and Classification

In the following tables, a function that is used when no function is explicitly specified by the user is shown as functionName.

The functions below produce a ranking, of which different size subsets are tried and the classifier performance evaluated, to select a best subset of features, based on a criterion such as balanced accuracy rate, for example.

Function Description DM DV DD
differentMeansRanking t-test ranking if two classes, F-test ranking if three or more
limmaRanking Moderated t-test ranking using variance shrinkage
edgeRranking Likelihood ratio test for count data ranking
bartlettRanking Bartlett’s test non-robust ranking
leveneRanking Levene’s test robust ranking
DMDranking Difference in location (mean/median) and/or scale (SD, MAD, \(Q_n\))
likelihoodRatioRanking Likelihood ratio (normal distribution) ranking
KolmogorovSmirnovRanking Kolmogorov-Smirnov distance between distributions ranking
KullbackLeiblerRanking Kullback-Leibler distance between distributions ranking

Likewise, a variety of classifiers is also provided.

Function(s) Description DM DV DD
DLDAtrainInterface,
DLDApredictInterface
Wrappers for sparsediscrim’s functions dlda and predict.dlda functions
classifyInterface Wrapper for PoiClaClu’s Poisson LDA function classify
elasticNetGLMtrainInterface, elasticNetGLMpredictInterface Wrappers for glmnet’s elastic net GLM functions glmnet and predict.glmnet
NSCtrainInterface, NSCpredictInterface Wrappers for pamr’s Nearest Shrunken Centroid functions pamr.train and pamr.predict
fisherDiscriminant Implementation of Fisher’s LDA for departures from normality ✔*
mixModelsTrain, mixModelsPredict Feature-wise mixtures of normals and voting
naiveBayesKernel Feature-wise kernel density estimation and voting
randomForestTrainInterface, randomForestPredictInterface Wrapper for ranger’s functions ranger and predict
extremeGradientBoostingTrainInterface, extremeGradientBoostingPredictInterface Wrapper for xgboost’s functions xgboost and predict
kNNinterface Wrapper for class’s function knn
SVMtrainInterface, SVMpredictInterface Wrapper for e1071’s functions svm and predict.svm ✔ † ✔ †

* If ordinary numeric measurements have been transformed to absolute deviations using subtractFromLocation.
† If the value of kernel is not “linear”.

If a desired selection or classification method is not already implemented, rules for writing functions to work with ClassifyR are outlined in the wrapper vignette. Please visit it for more information.

0.4.2.2 Provided Meta-feature Methods

A number of methods are provided for users to enable classification in a feature-set-centric or interactor-centric way. The meta-feature creation functions should be used before cross-validation is done.

Function Description Before CV During CV
edgesToHubNetworks Takes a two-column matrix or DataFrame and finds all nodes with at least a minimum number of interactions
featureSetSummary Considers sets of features and calculates their mean or median
pairsDifferencesSelection Finds a set of pairs of features whose measurement inequalities can be used for predicting with
kTSPclassifier Voting classifier that uses inequalities between pairs of features to vote for one of two classes

0.4.3 Fine-grained Cross-validation and Modelling Using runTests

For more control over the finer aspects of cross-validation of a single data set, runTests may be employed in place of crossValidate. For the variety of cross-validation, the parameters are specified by a CrossValParams object. The default setting is for 100 permutations and five folds and parameter tuning is done by resubstitution. It is also recommended to specify a parallelParams setting. On Linux and MacOS operating systems, it should be MulticoreParam and on Windows computers it should be SnowParam. Note that each of these have an option RNGseed and this needs to be set by the user because some classifiers or feature selection functions will have some element of randomisation. One example that works on all operating systems, but is best-suited to Windows is:

CVparams <- CrossValParams(parallelParams = SnowParam(16, RNGseed = 123))
CVparams

For the actual operations to do to the data to build a model of it, each of the stages should be specified by an object of class ModellingParams. This controls how class imbalance is handled (default is to downsample to the smallest class), any transformation that needs to be done inside of cross-validation (i.e. involving a computed value from the training set), any feature selection and the training and prediction functions to be used. The default is to do an ordinary t-test (two groups) or ANOVA (three or more groups) and classification using diagonal LDA.

ModellingParams()
## An object of class "ModellingParams"
## Slot "balancing":
## [1] "downsample"
## 
## Slot "transformParams":
## NULL
## 
## Slot "selectParams":
## An object of class 'SelectParams'.
## Selection Name: Difference in Means.
## 
## Slot "trainParams":
## An object of class 'TrainParams'.
## Classifier Name: Diagonal LDA.
## 
## Slot "predictParams":
## An object of class 'PredictParams'.
## 
## Slot "doImportance":
## [1] FALSE

0.4.4 runTests Driver Function of Cross-validated Classification

runTests is the main function in ClassifyR which handles the sample splitting and parallelisation, if used, of cross-validation. To begin with, a simple classifier will be demonstrated. It uses a t-test or ANOVA ranking (depending on the number of classes) for feature ranking and DLDA for classification. This classifier relies on differences in means between classes. No parameters need to be specified, because this is the default classification of runTests. By default, the number of features is tuned by resubstitution on the training set.

crossValParams <- CrossValParams(permutations = 5)
DMresults <- runTests(measurements, classes, crossValParams, verbose = 1)

Here, 5 permutations (non-default) and 5 folds cross-validation (default) is specified. For computers with more than 1 CPU, the number of cores to use can be given to runTests by using the argument parallelParams. The parameter seed is important to set for result reproducibility when doing a cross-validation such as this, because it employs randomisation to partition the samples into folds. Also, RNGseed is highly recommended to be set to the back-end specified to BPPARAM if doing parallel processing. The first seed mentioned does not work for parallel processes. For more details about runTests and the parameter classes used by it, consult the help pages of such functions.

0.5 Evaluation of a Classification

The most frequently selected gene can be identified using the distribution function and its relative abundance values for all samples can be displayed visually by plotFeatureClasses.

selectionPercentages <- distribution(DMresults, plot = FALSE)
head(selectionPercentages)
sortedPercentages <- head(selectionPercentages[order(selectionPercentages, decreasing = TRUE)])
head(sortedPercentages)
mostChosen <- sortedPercentages[1]
bestGenePlot <- plotFeatureClasses(measurements, classes, names(mostChosen), dotBinWidth = 0.1,
                                   xAxisLabel = "Normalised Expression")
## allFeaturesText
##    SSBP4   ZDHHC1    CROCC C10orf95    CTXN1  TMEM190 
##     1.00     1.00     0.96     0.92     0.84     0.80 
## allFeaturesText
##    SSBP4   ZDHHC1    CROCC C10orf95    CTXN1  TMEM190 
##     1.00     1.00     0.96     0.92     0.84     0.80

The means of the abundance levels of SSBP4 are substantially different between the people with and without asthma. plotFeatureClasses can also plot categorical data, such as may be found in a clinical data table, as a bar chart.

Classification error rates, as well as many other prediction performance measures, can be calculated with calcCVperformance. Next, the balanced accuracy rate is calculated considering all samples, each of which was in the test set once. The balanced accuracy rate is defined as the average rate of the correct classifications of each class.

See the documentation of calcCVperformance for a list of performance metrics which may be calculated.

DMresults <- calcCVperformance(DMresults)
DMresults
## An object of class 'ClassifyResult'.
## Characteristics:
##    characteristic                   value
##    Selection Name     Difference in Means
##   Classifier Name            Diagonal LDA
##  Cross-validation 5 Permutations, 5 Folds
## Features: List of length 25 of feature identifiers.
## Predictions: A data frame of 950 rows.
## Performance Measures: Balanced Accuracy, AUC.
performance(DMresults)
## $`Balanced Accuracy`
##       1:1       1:2       1:3       1:4       1:5       2:1       2:2       2:3       2:4       2:5       3:1       3:2       3:3 
## 0.8230769 0.6876923 0.7676923 0.8415385 0.8541667 0.8615385 0.7076923 0.7846154 0.8415385 0.7321429 0.8630769 0.9200000 0.7476923 
##       3:4       3:5       4:1       4:2       4:3       4:4       4:5       5:1       5:2       5:3       5:4       5:5 
## 0.7430769 0.6904762 0.7092308 0.8800000 0.9000000 0.7461538 0.6964286 0.8046154 0.6861538 0.8815385 0.8815385 0.8035714 
## 
## $AUC
## sapply(strsplit(names(performance[["values"]]), ":"), "[", 1): 1
## [1] 0.451
## ------------------------------------------------------------------------------------------------- 
## sapply(strsplit(names(performance[["values"]]), ":"), "[", 1): 2
## [1] 0.458
## ------------------------------------------------------------------------------------------------- 
## sapply(strsplit(names(performance[["values"]]), ":"), "[", 1): 3
## [1] 0.45
## ------------------------------------------------------------------------------------------------- 
## sapply(strsplit(names(performance[["values"]]), ":"), "[", 1): 4
## [1] 0.449
## ------------------------------------------------------------------------------------------------- 
## sapply(strsplit(names(performance[["values"]]), ":"), "[", 1): 5
## [1] 0.447

The error rate is about 20%. If only a vector of predictions and a vector of actual classes is available, such as from an old study which did not use ClassifyR for cross-validation, then calcExternalPerformance can be used on a pair of factor vectors which have the same length.

0.5.1 Comparison of Different Classifications

The samplesMetricMap function allows the visual comparison of sample-wise error rate or accuracy measures from different ClassifyResult objects. Firstly, a classifier will be run that uses Kullback-Leibler divergence ranking and resubstitution error as a feature selection heuristic and a naive Bayes classifier for classification. This classification will use features that have either a change in location or in scale between classes.

modellingParamsDD <- ModellingParams(selectParams = SelectParams("KL"),
                                     trainParams = TrainParams("naiveBayes"),
                                     predictParams = NULL)
DDresults <- runTests(measurements, classes, crossValParams, modellingParamsDD, verbose = 1)
DDresults
## An object of class 'ClassifyResult'.
## Characteristics:
##    characteristic                       value
##    Selection Name Kullback-Leibler Divergence
##   Classifier Name          Naive Bayes Kernel
##  Cross-validation     5 Permutations, 5 Folds
## Features: List of length 25 of feature identifiers.
## Predictions: A data frame of 950 rows.
## Performance Measures: None calculated yet.

The naive Bayes kernel classifier by default uses the vertical distance between class densities but it can instead use the horizontal distance to the nearest non-zero density cross-over point to confidently classify samples in the tails of the densities.

The per-sample classification accuracy is automatically calculated for both the differential means and differential distribution classifiers and plotted with samplesMetricMap.

resultsList <- list(Abundance = DMresults, Distribution = DDresults)
samplesMetricMap(resultsList, showXtickLabels = FALSE)
## Warning in .local(results, ...): Sample Accuracy not found in all elements of results. Calculating it now.
## Warning: Removed 2 rows containing missing values or values outside the scale range (`geom_tile()`).

## TableGrob (2 x 1) "arrange": 2 grobs
##   z     cells    name                grob
## 1 1 (2-2,1-1) arrange      gtable[layout]
## 2 2 (1-1,1-1) arrange text[GRID.text.581]

The benefit of this plot is that it allows the easy identification of samples which are hard to classify and could be explained by considering additional information about them. Differential distribution class prediction appears to be biased to the majority class (No Asthma).

More traditionally, the distribution of performance values of each complete cross-validation can be visualised by performancePlot by providing them as a list to the function. The default is to draw box plots, but violin plots could also be made. The default performance metric to plot is balanced accuracy. If it’s not already calculated for all classifications, as in this case for DD, it will be done automatically.

performancePlot(resultsList)
## Warning in .local(results, ...): Balanced Accuracy not found in all elements of results. Calculating it now.

We can observe that the spread of balanced accuracy rates is small, but slightly wider for the differential distribution classifier.

The features being ranked and selected in the feature selection stage can be compared within and between classifiers by the plotting functions rankingPlot and selectionPlot. Consider the task of visually representing how consistent the feature rankings of the top 100 different features were for the differential distribution classifier for all 5 folds in the 5 cross-validations.

rankingPlot(DDresults, topRanked = 1:100, xLabelPositions = c(1, seq(10, 100, 10)))

The top-ranked features are fairly similar between all pairs of the 20 cross-validations.

For a large cross-validation scheme, such as leave-2-out cross-validation, or when results contains many classifications, there are many feature set comparisons to make. Note that rankingPlot and selectionPlot have a parallelParams options which allows for the calculation of feature set overlaps to be done on multiple processors.

0.5.2 Generating a ROC Plot

Some classifiers can output scores or probabilities representing how likely a sample is to be from one of the classes, instead of, or as well as, class labels. This enables different score thresholds to be tried, to generate pairs of false positive and false negative rates. The naive Bayes classifier used previously by default has its returnType parameter set to “both”, so class predictions and scores are both stored in the classification result. So does diagonal LDA. In this case, a data frame with class predictions and scores for each class is returned by the classifier to the cross-validation framework. Setting returnType to “score” for a classifier which has such an option is also sufficient to generate a ROC plot. Many existing classifiers in other R packages also have an option that allows a score or probability to be calculated.

By default, scores from different iterations of prediction are merged and one line is drawn per classification. Alternatively, setting mode = “average” will consider each iteration of prediction separately, average them and also calculate and draw confidence intervals. The default interval is a 95% interval and is customisable by setting interval.

ROCplot(resultsList, fontSizes = c(24, 12, 12, 12, 12))

This ROC plot shows the classifiability of the asthma data set is high. Some examples of functions which output scores are fisherDiscriminant, DLDApredictInterface, and SVMpredictInterface.

0.6 Other Use Cases

Apart from cross-validation of one data set, ClassifyR can be used in a couple of other ways.

0.6.1 Using an Independent Test Set

Sometimes, cross-validation is unnecessary. This happens when studies have large sample sizes and are designed such that a large number of samples is prespecified to form a test set. The classifier is only trained on the training sample set, and makes predictions only on the test sample set. This can be achieved by using the function runTest directly. See its documentation for required inputs.

0.6.2 Cross-validating Selected Features on a Different Data Set

Once a cross-validated classification is complete, the usefulness of the features selected may be explored in another dataset. previousSelection is a function which takes an existing ClassifyResult object and returns the features selected at the equivalent iteration which is currently being processed. This is necessary, because the models trained on one data set are not directly transferrable to a new dataset; the classifier training (e.g. choosing thresholds, fitting model coefficients) is redone. Of course, the features in the new dataset should have the same naming system as the ones in the old dataset.

0.6.3 Parameter Tuning

Some feature ranking methods or classifiers allow the choosing of tuning parameters, which controls some aspect of their model learning. An example of doing parameter tuning with a linear SVM is presented. This particular SVM has a single tuning parameter, the cost. Higher values of this parameter penalise misclassifications more. Moreover, feature selection happens by using a feature ranking function and then trying a range of top-ranked features to see which gives the best performance, the range being specified by a list element named nFeatures and the performance type (e.g. Balanced Accuracy) specified by a list element named performanceType. Therefore, some kind of parameter tuning always happens, even if the feature ranking or classifier function does not have any explicit tuning parameters.

Tuning is achieved in ClassifyR by providing a variable called tuneParams to the SelectParams or TrainParams constructor. tuneParams is a named list, with the names being the names of the tuning variables, except for one which is named “performanceType” and specifies the performance metric to use for picking the parameter values. Any of the non-sample-specific performance metrics which calcCVperformance calculates can be optimised.

tuneList <- list(cost = c(0.01, 0.1, 1, 10))
SVMparams <- ModellingParams(trainParams = TrainParams("SVM", kernel = "linear", tuneParams = tuneList),
                             predictParams = PredictParams("SVM"))
SVMresults <- runTests(measurements, classes, crossValParams, SVMparams)

The index of chosen of the parameters, as well as all combinations of parameters and their associated performance metric, are stored for every validation, and can be accessed with the tunedParameters function.

length(tunedParameters(SVMresults))
## [1] 25
tunedParameters(SVMresults)[1:5]
## [[1]]
## [[1]]$tuneCombinations
##    topN  cost Balanced Accuracy
## 1    10  0.01         0.8022680
## 2    20  0.01         0.8167524
## 3    30  0.01         0.8268534
## 4    40  0.01         0.8218029
## 5    50  0.01         0.8268534
## 6    60  0.01         0.8268534
## 7    70  0.01         0.8507719
## 8    80  0.01         0.8507719
## 9    90  0.01         0.8507719
## 10  100  0.01         0.8507719
## 11   10  0.10         0.7985516
## 12   20  0.10         0.8514389
## 13   30  0.10         0.8746903
## 14   40  0.10         0.8898418
## 15   50  0.10         0.8905089
## 16   60  0.10         0.9144273
## 17   70  0.10         0.9093768
## 18   80  0.10         0.9282447
## 19   90  0.10         0.9238613
## 20  100  0.10         0.9194778
## 21   10  1.00         0.7854012
## 22   20  1.00         0.8709739
## 23   30  1.00         0.8804079
## 24   40  1.00         0.8861254
## 25   50  1.00         0.9049933
## 26   60  1.00         0.9383457
## 27   70  1.00         0.9710311
## 28   80  1.00         0.9811321
## 29   90  1.00         0.9811321
## 30  100  1.00         0.9811321
## 31   10 10.00         0.7753002
## 32   20 10.00         0.8817419
## 33   30 10.00         0.8113208
## 34   40 10.00         0.8106537
## 35   50 10.00         0.8490566
## 36   60 10.00         1.0000000
## 37   70 10.00         1.0000000
## 38   80 10.00         1.0000000
## 39   90 10.00         1.0000000
## 40  100 10.00         1.0000000
## 
## [[1]]$bestIndex
## [1] 36
## 
## 
## [[2]]
## [[2]]$tuneCombinations
##    topN  cost Balanced Accuracy
## 1    10  0.01         0.8022680
## 2    20  0.01         0.7972175
## 3    30  0.01         0.8160854
## 4    40  0.01         0.8261864
## 5    50  0.01         0.8211359
## 6    60  0.01         0.8218029
## 7    70  0.01         0.8218029
## 8    80  0.01         0.8356204
## 9    90  0.01         0.8645893
## 10  100  0.01         0.8551553
## 11   10  0.10         0.7941681
## 12   20  0.10         0.8369545
## 13   30  0.10         0.8514389
## 14   40  0.10         0.9087097
## 15   50  0.10         0.9188107
## 16   60  0.10         0.9188107
## 17   70  0.10         0.9137602
## 18   80  0.10         0.9188107
## 19   90  0.10         0.9572136
## 20  100  0.10         0.9477797
## 21   10  1.00         0.8362874
## 22   20  1.00         0.8608729
## 23   30  1.00         0.8766914
## 24   40  1.00         0.9471126
## 25   50  1.00         0.9332952
## 26   60  1.00         0.9477797
## 27   70  1.00         0.9332952
## 28   80  1.00         0.9666476
## 29   90  1.00         0.9905660
## 30  100  1.00         1.0000000
## 31   10 10.00         0.8558224
## 32   20 10.00         0.8659234
## 33   30 10.00         0.9376787
## 34   40 10.00         0.9804650
## 35   50 10.00         0.9905660
## 36   60 10.00         1.0000000
## 37   70 10.00         1.0000000
## 38   80 10.00         1.0000000
## 39   90 10.00         1.0000000
## 40  100 10.00         1.0000000
## 
## [[2]]$bestIndex
## [1] 30
## 
## 
## [[3]]
## [[3]]$tuneCombinations
##    topN  cost Balanced Accuracy
## 1    10  0.01         0.8305698
## 2    20  0.01         0.8167524
## 3    30  0.01         0.8123690
## 4    40  0.01         0.8079855
## 5    50  0.01         0.8551553
## 6    60  0.01         0.8551553
## 7    70  0.01         0.8602058
## 8    80  0.01         0.8746903
## 9    90  0.01         0.8797408
## 10  100  0.01         0.8942253
## 11   10  0.10         0.8703068
## 12   20  0.10         0.8652563
## 13   30  0.10         0.8942253
## 14   40  0.10         0.8753573
## 15   50  0.10         0.8847913
## 16   60  0.10         0.8992758
## 17   70  0.10         0.8992758
## 18   80  0.10         0.9332952
## 19   90  0.10         0.9332952
## 20  100  0.10         0.9282447
## 21   10  1.00         0.8382885
## 22   20  1.00         0.8797408
## 23   30  1.00         0.8804079
## 24   40  1.00         0.9238613
## 25   50  1.00         0.9420621
## 26   60  1.00         0.9383457
## 27   70  1.00         0.9622642
## 28   80  1.00         0.9811321
## 29   90  1.00         1.0000000
## 30  100  1.00         1.0000000
## 31   10 10.00         0.8760244
## 32   20 10.00         0.8810749
## 33   30 10.00         0.8955594
## 34   40 10.00         0.9716981
## 35   50 10.00         0.9905660
## 36   60 10.00         1.0000000
## 37   70 10.00         1.0000000
## 38   80 10.00         1.0000000
## 39   90 10.00         1.0000000
## 40  100 10.00         1.0000000
## 
## [[3]]$bestIndex
## [1] 29
## 
## 
## [[4]]
## [[4]]$tuneCombinations
##    topN  cost Balanced Accuracy
## 1    10  0.01         0.8356204
## 2    20  0.01         0.8167524
## 3    30  0.01         0.8740233
## 4    40  0.01         0.8885077
## 5    50  0.01         0.8979417
## 6    60  0.01         0.8885077
## 7    70  0.01         0.8986087
## 8    80  0.01         0.8986087
## 9    90  0.01         0.9225272
## 10  100  0.01         0.9130932
## 11   10  0.10         0.8878407
## 12   20  0.10         0.8878407
## 13   30  0.10         0.9269106
## 14   40  0.10         0.9124261
## 15   50  0.10         0.9225272
## 16   60  0.10         0.9130932
## 17   70  0.10         0.9174767
## 18   80  0.10         0.9332952
## 19   90  0.10         0.9332952
## 20  100  0.10         0.9282447
## 21   10  1.00         0.8935582
## 22   20  1.00         0.8935582
## 23   30  1.00         0.9457785
## 24   40  1.00         0.9420621
## 25   50  1.00         0.9457785
## 26   60  1.00         0.9703640
## 27   70  1.00         0.9659806
## 28   80  1.00         0.9811321
## 29   90  1.00         1.0000000
## 30  100  1.00         1.0000000
## 31   10 10.00         0.8942253
## 32   20 10.00         0.9319611
## 33   30 10.00         0.9797980
## 34   40 10.00         0.9898990
## 35   50 10.00         1.0000000
## 36   60 10.00         1.0000000
## 37   70 10.00         1.0000000
## 38   80 10.00         1.0000000
## 39   90 10.00         1.0000000
## 40  100 10.00         1.0000000
## 
## [[4]]$bestIndex
## [1] 29
## 
## 
## [[5]]
## [[5]]$tuneCombinations
##    topN  cost Balanced Accuracy
## 1    10  0.01         0.8250000
## 2    20  0.01         0.8107692
## 3    30  0.01         0.8542308
## 4    40  0.01         0.8926923
## 5    50  0.01         0.8834615
## 6    60  0.01         0.8784615
## 7    70  0.01         0.9073077
## 8    80  0.01         0.9126923
## 9    90  0.01         0.9223077
## 10  100  0.01         0.9223077
## 11   10  0.10         0.8446154
## 12   20  0.10         0.8688462
## 13   30  0.10         0.9076923
## 14   40  0.10         0.9319231
## 15   50  0.10         0.9369231
## 16   60  0.10         0.9469231
## 17   70  0.10         0.9469231
## 18   80  0.10         0.9515385
## 19   90  0.10         0.9515385
## 20  100  0.10         0.9519231
## 21   10  1.00         0.8734615
## 22   20  1.00         0.9030769
## 23   30  1.00         0.9561538
## 24   40  1.00         0.9807692
## 25   50  1.00         0.9807692
## 26   60  1.00         0.9903846
## 27   70  1.00         0.9903846
## 28   80  1.00         1.0000000
## 29   90  1.00         1.0000000
## 30  100  1.00         1.0000000
## 31   10 10.00         0.8638462
## 32   20 10.00         0.9323077
## 33   30 10.00         0.9853846
## 34   40 10.00         1.0000000
## 35   50 10.00         1.0000000
## 36   60 10.00         1.0000000
## 37   70 10.00         1.0000000
## 38   80 10.00         1.0000000
## 39   90 10.00         1.0000000
## 40  100 10.00         1.0000000
## 
## [[5]]$bestIndex
## [1] 28

The cost value of 1 or 10 appears to often be chosen.

0.7 Summary

ClassifyR is a framework for cross-validated classification that provides a variety of unique functions for performance evaluation. It provides wrappers for many popular classifiers but is designed to be extensible if other classifiers are desired.

0.8 References

Strbenac D., Yang, J., Mann, G.J. and Ormerod, J. T. (2015) ClassifyR: an R package for performance assessment of classification with applications to transcriptomics, Bioinformatics, 31(11):1851-1853
Strbenac D., Mann, G.J., Yang, J. and Ormerod, J. T. (2016) Differential distribution improves gene selection stability and has competitive classification performance for patient survival, Nucleic Acids Research, 44(13):e119