Description Usage Arguments Details Value See Also Examples
Learn a network (structure and parameters) of a BN from a BNDataset (see the Details
section).
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 | learn.network(x, ...)
## S4 method for signature 'BN'
learn.network(
x,
y = NULL,
algo = "mmhc",
scoring.func = "BDeu",
initial.network = NULL,
alpha = 0.05,
ess = 1,
bootstrap = FALSE,
layering = c(),
max.fanin = num.variables(y) - 1,
max.fanin.layers = NULL,
max.parents = num.variables(y) - 1,
max.parents.layers = NULL,
layer.struct = NULL,
cont.nodes = c(),
use.imputed.data = FALSE,
use.cpc = TRUE,
mandatory.edges = NULL,
...
)
## S4 method for signature 'BNDataset'
learn.network(
x,
algo = "mmhc",
scoring.func = "BDeu",
initial.network = NULL,
alpha = 0.05,
ess = 1,
bootstrap = FALSE,
layering = c(),
max.fanin = num.variables(x) - 1,
max.fanin.layers = NULL,
max.parents = num.variables(x) - 1,
max.parents.layers = NULL,
layer.struct = NULL,
cont.nodes = c(),
use.imputed.data = FALSE,
use.cpc = TRUE,
mandatory.edges = NULL,
...
)
|
x |
can be a |
... |
potential further arguments for methods. |
y |
a |
algo |
the algorithm to use. Currently, one among |
scoring.func |
the scoring function to use. Currently, one among
|
initial.network |
network structure to be used as starting point for structure search.
Can take different values:
a |
alpha |
confidence threshold (only for |
ess |
Equivalent Sample Size value. |
bootstrap |
|
layering |
vector containing the layers each node belongs to. |
max.fanin |
maximum number of parents for each node (only for |
max.fanin.layers |
matrix of available parents in each layer (only for |
max.parents |
maximum number of parents for each node (for |
max.parents.layers |
matrix of available parents in each layer (only for |
layer.struct |
|
cont.nodes |
vector containing the index of continuous variables. |
use.imputed.data |
|
use.cpc |
(when using |
mandatory.edges |
binary matrix, where a |
Learn the structure (the directed acyclic graph) of a BN
object according to a BNDataset
.
We provide five algorithms for learning the structure of the network, that can be chosen with the algo
parameter.
The first one is the Silander-Myllym\"aki (sm
)
exact search-and-score algorithm, that performs a complete evaluation of the search space in order to discover
the best network; this algorithm may take a very long time, and can be inapplicable when discovering networks
with more than 25–30 nodes. Even for small networks, users are strongly encouraged to provide
meaningful parameters such as the layering of the nodes, or the maximum number of parents – refer to the
documentation in package manual for more details on the method parameters.
The second method is the constraint-based Max-Min Parents-and-Children (mmpc
), that returns the skeleton of the network.
Given the possible presence of loops, due to the non-directionality of the edges discovered, no parameter learning
is possible using this algorithm. Also note that in the case of a very dense network and lots of obsevations, the statistical evaluation
of the search space may take a long time. Also for this algorithm there are parameters that may need to be tuned,
mainly the confidence threshold of the statistical pruning. Please refer to the rest of this documentation for their explanation.
The third algorithm is another heuristic, the Hill-Climbing (hc
). It can start from the complete space of possibilities
(default) or from a reduced subset of possible edges, using the cpc
argument.
The fourth algorithm (and the default one) is the Max-Min Hill-Climbing heuristic (mmhc
), that performs a statistical
sieving of the search space followed by a greedy evaluation, by combining the MMPC and the HC algorithms.
It is considerably faster than the complete method, at the cost of a (likely)
lower quality. As for MMPC, the computational time depends on the density of the network, the number of observations and
the tuning of the parameters.
The fifth method is the Structural Expectation-Maximization (sem
) algorithm,
for learning a network from a dataset with missing values. It iterates a sequence of Expectation-Maximization (in order to “fill in”
the holes in the dataset) and structure learning from the guessed dataset, until convergence. The structure learning used inside SEM,
due to computational reasons, is MMHC. Convergence of SEM can be controlled with the parameters struct.threshold
and param.threshold
, for the structure and the parameter convergence, respectively.
Search-and-score methods also need a scoring function to compute an estimated measure of each configuration of nodes.
We provide three of the most popular scoring functions, BDeu
(Bayesian-Dirichlet equivalent uniform, default),
AIC
(Akaike Information Criterion) and BIC
(Bayesian Information Criterion). The scoring function
can be chosen using the scoring.func
parameter.
Structure learning sets the dag
field of the BN
under study, unless bootstrap or the mmpc
algorithm
are employed. In these cases, given the possible presence of loops, the wpdag
field is set.
In case of missing data, the default behaviour (with no other indication from the user)
is to learn the structure using mmhc
starting from the raw dataset, using only the
available cases with no imputation.
In case of learning from a dataset containing observations of a dynamic system, learn.dynamic.network
will be employed.
Then, the parameters of the network are learnt using MAP (Maximum A Posteriori) estimation (when not using bootstrap or mmpc
).
See documentation for learn.structure
and learn.params
for more informations.
new BN
object with structure (DAG) and conditional probabilities
as learnt from the given dataset.
learn.structure learn.params learn.dynamic.network
1 2 3 4 5 6 7 8 9 10 11 | ## Not run:
mydataset <- BNDataset("data.file", "header.file")
# starting from a BN
net <- BN(mydataset)
net <- learn.network(net, mydataset)
# start directly from the dataset
net <- learn.network(mydataset)
## End(Not run)
|
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.