knitr::opts_chunk$set(cache = FALSE, autodep = TRUE, collapse = TRUE, comment = "#>" # , fig.height = 3, fig.width = 3 )
The bnclassify package implements state-of-the-art algorithms for learning discrete Bayesian network classifiers from data, as well as functions for using these classifiers for prediction, assessing their predictive performance, and inspecting and analyzing their properties. This vignette gives \invindesc/. Other resources provide additional information:
vignette("overview", package="bnclassify") provides \overvindesc/. ?bnclassify provides a \pkghelp/.vignette("methods", package="bnclassify") provides \tecvindesc/. Throughout the vignette we will use the car evaluation data set. It has six discrete features, describing car properties such as buying price or the number of doors, and 1728 instances assigned to four different classes (unacc, acc, good, vgood). See ?car for more details.
library(bnclassify) data(car) dim(car) head(car)
Using bnclassify generally consists of four steps:
In between those steps, you may also want to inspect the model's properties.
Below is an example of the four steps done in four lines.
nb <- nb('class', car) # Learn a naive Bayes structure nb <- lp(nb, car, smooth = 1) # Learn parameters cv(nb, car, k = 10) # 10-fold Cross-validation estimate of accuracy head(predict(nb, car)) # Classify the entire data set
While there are multiple alternatives to nb for the first step, you are most likely to use lp, cv, and predict for steps 2-4. We will elaborate on all four steps throughout the rest of the vignette.
\texttt{bnlassify} provides one function per each structure learning algorithm that it implements. Grouped according to algorithm type (see vignette("bnclassify-technical")), these are:
Naive Bayes:
nbCL ODE:
tan_clGreedy wrapper:
tan_hctan_hcspfssjbsejThey all receive the name of the class variable and the data set as their first two arguments, followed by optional arguments.
The following learns three different structures with three different algorithms.
# Naive Bayes nb <- nb('class', car) # ODE Chow-Liu with AIC score (penalized log-likelihood) ode_cl_aic <- tan_cl('class', car, score = 'aic') # Semi-naive Bayes with forward sequential selection and joining (FSSJ) and # 5-fold cross-validation fssj <- fssj('class', car, k = 5, epsilon = 0)
For details on the learning algorithms, see the corresponding functions (e.g., ?tan_cl) and vignette("bnclassify-technical").
\label{interpreting-the-models}
The above nb, ode_cl_aic, and fssj are objects of class bnc_dag. There are a number of functions that you can perform on such objects.
Printing the object to console outputs basic information on structure:
ode_cl_aic
The above tells that the ode_cl_aic object is a network structure without any parameters, the name of the class variables is "class", it has six feature nodes and nine arcs, and it was learned with the tan_cl function.
Plotting network structure can reveal probabilistic relationships among the variables:
plot(ode_cl_aic)
If the network is not displaying properly, e.g., with node names overlapping in large networks, you may try different layout types and font sizes (see ?plot.bnc_dag).
plot(ode_cl_aic, layoutType = 'twopi', fontsize = 15)
An alternative to plotting, useful when the graph is large, is to query for the families that compose the structure (a family of a node is itself plus its parents in the graph).
families(ode_cl_aic)
narcs gives the number of arcs in a structure.
narcs(nb)
Functions such as is_ode, is_nb, or id_semi query the type of structure. For example:
is_ode(ode_cl_aic) is_semi_naive(ode_cl_aic)
For more functions to query a network structure, see ?inspect_bnc_dag.
bnclassify provides three parameter estimation methods, all implemented with the lp function.
lp which takes the network structure and the dataset from which to learn parameters as its first two arguments.
To get Bayesian parameter estimates assuming a Dirichlet prior, provide a positive \texttt{smooth} argument to \texttt{lp}.
nb <- lp(nb, car, smooth = 0.01)
For AWNB or MANB parameter estimation, provide the appropriate arguments to \texttt{lp}, in addition to smooth.
awnb <- lp(nb, car, smooth = 0.01, awnb_trees = 10, awnb_bootstrap = 0.5) manb <- lp(nb, car, smooth = 0.01, manb_prior = 0.5)
The bnc function is shorthand for learning both structure and parameters in a single step. Provide the name of the structure learning algorithm, as a character, and its optional arguments in dag_args.
ode_cl_aic <- bnc('tan_cl', 'class', car, smooth = 1, dag_args = list(score = 'aic'))
lp and bnc return objects of class bnc_bn, which are fully specified Bayesian network classifiers (i.e., with both structure and parameters).
Printing the ode_cl_aic object now also shows how many free parameters there are in the model (131).
ode_cl_aic
params lets you access the conditional probability tables (CPTs). For example, the CPT of the buying feature in nb is:
params(nb)$buying
params(awnb)$buying params(manb)$buying
nparams gives the number of parameters of the classifier.
nparams(nb)
For more functions for querying a bnc_bn object, see ?inspect_bnc_bn
bnlearn, gRain, and graphYou can convert a bnc_bn object to bnlearn [@scutari2009learning], gRain [@Hojsgaard2012] and graph [@Gentleman2015] objects to leverage functionalities from those packages, such as Bayesian network querying or inference.
Use
as_igraph for graphas_grain for gRainFor bnlearn, first convert to gRain and then convert the gRain object to a bnlearn one (see bnlearn docs for how to do this).
The following uses gRain to ge the marginal probability of the buying feature: (NOTE: not currently working due to recent changes in the gRain package)
a <- lp(nb('class', car), car, smooth = 1) g <- as_grain(a) gRain::querygrain(g)$buying
\label{sec:fss}
Some structure and parameter learning methods perform feature selection:
fssj and bsej : embedded wrapperfssj and bsej perform feature selection while learning structure. On the car evaluation data they both select all features.
length(features(fssj)) suppressWarnings(RNGversion("3.5.0")) set.seed(0) bsej <- bsej('class', car, k = 5, epsilon = 0) length(features(bsej))
MANB has computed zero posterior probability for the arc from class to doors and 100\% probability for arcs to the other features.
manb_arc_posterior(manb)
This means that it has effectively omitted doors from the model, rendering it independent from the class.
params(manb)$doors
It has left the other features' parameters unaltered.
all.equal(params(manb)$buying, params(nb)$buying)
The AWNB method has decreased the effect of each feature on the class posterior, especially doors, lug_boot, and maint, also modifying their local distributions towards independence from the class.
awnb_weights(awnb)
\label{sec:efss}
You can use R packages such as mlr [@Bischl2015] or caret [@Kuhn2008] to select features prior to learning a classifier with bnclassify. See Section \ref{sec:mlr} for how to do it with mlr.
The are three functions for computing penalized log-likelihood network scores of bnc_bn objects.
logLikAICBICIn addition to the model, provide them the dataset on which to compute the score.
logLik(ode_cl_aic, car) AIC(ode_cl_aic, car) BIC(ode_cl_aic, car)
accuracy lets you compute the classifier's predictive accuracy on a given data set. You need to provide the vectors of predicted and true labels.
p <- predict(nb, car) accuracy(p, car$class)
cv estimates predictive accuracy with stratified cross-validation. Indicate the desired number of folds with k.
suppressWarnings(RNGversion("3.5.0")) set.seed(0) cv(ode_cl_aic, car, k = 10)
Each bnc_bn object records the structure and parameter learning methods that were used to produce it. cv just reruns these methods. Hence, the above is the accuracy estimate for tan_cl with the AIC score and Bayesian parameter estimation with smooth = 0.01.
To keep the structure fixed and evaluate just the parameter learning method, set dag = FALSE:
suppressWarnings(RNGversion("3.5.0")) set.seed(0) cv(ode_cl_aic, car, k = 10, dag = FALSE)
To get the accuracy for each of the folds, instead of the mean accuracy, set mean = FALSE.
suppressWarnings(RNGversion("3.5.0")) set.seed(0) cv(ode_cl_aic, car, k = 10, dag = FALSE, mean = FALSE)
Finally, to cross-validate multiple classifiers at once pass a list of bnc_bn objects to cv.
suppressWarnings(RNGversion("3.5.0")) set.seed(0) accu <- cv(list(nb = nb, ode_cl_aic = ode_cl_aic), car, k = 5, dag = TRUE) accu
General-purpose machine learning packages such as mlr or caret provide additional options for evaluating a model, including bootstrap resampling and performance measures such as the area under the curve. See Section \ref{sec:mlr} for how that could be done with mlr.
We can use a bnc_bn object to classify data instances, with predict.
Here we use the naive Bayes to predict the class for our entire data set.
p <- predict(nb, car) # We use head() to print the first elements of vector p head(p)
You can also get the class posterior probabilities.
pp <- predict(nb, car, prob = TRUE) head(pp)
You can compute the (conditional) mutual information between two variables with cmi. Mutual information of maint and buying:
cmi('maint', 'buying', car)
Mutual information of maint and buying conditioned to class:
cmi('maint', 'buying', car, 'class')
bnclassify with mlr\label{sec:mlr}
General-purpose machine learning packages, such as mlr and caret, provide many options for feature selection and model evaluation. For example, the provide resampling methods other than cross-validation and performance measures other than accuracy. Here we use mlr to:
tan_cltan_cl and random forestTo use a bnc_bn object with mlr, call the as_mlr function.
library(mlr) ode_cl_aic_mlr <- as_mlr(ode_cl_aic, dag = TRUE, id = "ode_cl_aic")
The obtained ode_cl_aic_mlr behaves like any other classifier supported by mlr.
Set up sequential forward search with 2-fold cross validation and ode_cl_aic_mlr as the classifier.
# 5-fold cross-validation rdesc = makeResampleDesc("CV", iters = 2) # sequential floating forward search ctrl = makeFeatSelControlSequential(method = "sfs", alpha = 0) # Wrap ode_cl_aic_mlr with feature selection ode_cl_aic_mlr_fs = makeFeatSelWrapper(ode_cl_aic_mlr, resampling = rdesc, control = ctrl, show.info = FALSE) t <- makeClassifTask(id = "car", data = car, target = 'class', fixup.data = "no", check.data = FALSE)
Select features:
suppressWarnings(RNGversion("3.5.0")) set.seed(0) # Select features mod <- train(ode_cl_aic_mlr_fs, task = t) sfeats <- getFeatSelResult(mod) sfeats
mlr makes it easy to evaluate the predictive performance of the combination of feature selection plus classifier learning. The following estimates accuracy with 2-fold cross-validation:
suppressWarnings(RNGversion("3.5.0")) set.seed(0) r = resample(learner = ode_cl_aic_mlr_fs, task = t, resampling = rdesc, show.info = FALSE, measure = mlr::acc)
With mlr you can compare the predictive performance of bnclassify models to those of different classification paradigms, such as random forests.
rf <- makeLearner("classif.randomForest", id = "rf") classifiers <- list(ode_cl_aic_mlr, rf) suppressWarnings(RNGversion("3.5.0")) set.seed(0) benchmark(classifiers, t, rdesc, show.info = FALSE, measures = mlr::acc)
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.