Skip to main content

Table 1 Machine learning methods used in the experiments with the optional abbreviations used in further work

From: The influence of the inactives subset generation on the performance of machine learning methods

Classifier

Classification scheme

Parameters

Naïve Bayes (NB)

bayes

-

Sequential Minimal Optimization (SMO)

functions

The complexity parameter was set at 1, the epsilon for a round-off error was 1.0 E-12, and an option of normalizing training data was chosen.

Kernels:

 1) The normalized polynomial kernel,

 2) The polynomial kernel

 3) The RBF kernel

Instance-Based Learning (Ibk)

lazy

The brute force search algorithm for nearest neighbour search with Euclidean distance function.

The number of neighbours used:

 1) 1

 2) 5

 3) 10

 4) 20

Decorate

meta

One artificial example used during training, number of member classifiers in the Decorate ensemble: 10, the maximum number of iterations: 10.

Base classifiers:

 1) NaïveBayes

 2) J48

Hyperpipes

misc

-

J48

trees

 1) With reduced-error pruning

 2) With C.4.5 pruning

Random Forest (RF)

trees

Trees with unlimited depth, seed number: 1.

Number of generated trees:

 1) 5

 2) 10

 3) 50

 4) 100

  1. Bolded parameters correspond with the one providing the best results for particular machine learning method (see Results section & Additional file 1: Figure S1).