Score of the Bayesian network
Compute the score of the Bayesian network.
1 2 3 4 5 6 7 8
an object of class
a data frame containing the data the Bayesian network was learned from.
a character string, the label of a network score. If none is
specified, the default score is the Bayesian Information Criterion
for both discrete and continuous data sets. See
a boolean value. If
extra arguments from the generic method (for the
a numeric value, the penalty per parameter to be used; the default
Additional parameters of the
iss: the imaginary sample size, used by the Bayesian Dirichlet equivalent score (both the
mbde) and the Bayesian Gaussian score (
bge). It is also known as “equivalent sample size”. The default value is equal to
10for both the
exp: a list of indexes of experimental observations (those that have been artificially manipulated). Each element of the list must be named after one of the nodes, and must contain a numeric vector with indexes of the observations whose value has been manipulated for that node.
k: the penalty per parameter to be used by the AIC and BIC scores. The default value is
1for AIC and
phi: the prior phi matrix formula to use in the Bayesian Gaussian equivalent (
bge) score. Possible values are
bottcher(the one used by default in the deal package.)
prior: the prior distribution to be used with the various Bayesian Dirichlet scores (
bds) and the Bayesian Gaussian score (
bge). Possible values are
vsp(the Bayesian variable selection prior, which puts a probability of inclusion on parents) and
cs(the Castelo & Siebes prior, which puts an independent prior probability on each arc and direction).
beta: the parameter associated with
betais ignored. If
betais the probability of inclusion of an additional parent (the default is
betais a data frame with columns
probspecifying the prior probability for a set of arcs. A uniform probability distribution is assumed for the remaining arcs.
A numeric value, the score of the Bayesian network.
AIC and BIC are computed as
logLik(x) - k * nparams(x), that is, the
classic definition rescaled by -2. Therefore higher values are better, and
for large sample sizes BIC converges to log(BDe).
When using the Castelo & Siebes prior in structure learning, the prior
probabilties associated with an arc are bound away from zero and one by
shrinking them towards the uniform distribution as per Hausser and Strimmer
(2009) with a lambda equal to
3 * sqrt(.Machine$double.eps). This
dramatically improves structure learning, which is less likely to get stuck
when starting from an empty graph. As an alternative to prior probabilities,
a blacklist can be used to prevent arcs from being included in the network,
and a whitelist can be used to force the inclusion of particular arcs.
beta is not modified when the prior is used from functions other than
those implementing score-based and hybrid structure learning.
Castelo R, Siebes A (2000). "Priors on Network Structures. Biasing the Search for Bayesian Networks". International Journal of Approximate Reasoning, 24(1), 39-57.
Chickering DM (1995). "A Transformational Characterization of Equivalent Bayesian Network Structures". In "UAI '95: Proceedings of the Eleventh Annual Conference on Uncertainty in Artificial Intelligence", pp. 87-98. Morgan Kaufmann.
Cooper GF, Yoo C (1999). "Causal Discovery from a Mixture of Experimental and Observational Data". In "UAI '99: Proceedings of the Fifteenth Annual Conference on Uncertainty in Artificial Intelligence", pp. 116-125. Morgann Kaufmann.
Geiger D, Heckerman D (1994). "Learning Gaussian Networks". In "UAI '94: Proceedings of the Tenth Annual Conference on Uncertainty in Artificial Intelligence", pp. 235-243. Morgann Kaufmann. Available as Technical Report MSR-TR-94-10.
Hausser J, Strimmer K (2009). "Entropy inference and the James-Stein estimator, with application to nonlinear gene association networks". Statistical Applications in Genetics and Molecular Biology, 10, 1469-1484.
Heckerman D, Geiger D, Chickering DM (1995). "Learning Bayesian Networks: The Combination of Knowledge and Statistical Data". Machine Learning, 20(3), 197-243. Available as Technical Report MSR-TR-94-09.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22
data(learning.test) res = set.arc(gs(learning.test), "A", "B") score(res, learning.test, type = "bde") ## let's see score equivalence in action! res2 = set.arc(gs(learning.test), "B", "A") score(res2, learning.test, type = "bde") ## K2 score on the other hand is not score equivalent. score(res, learning.test, type = "k2") score(res2, learning.test, type = "k2") ## BDe with a prior. beta = data.frame(from = c("A", "D"), to = c("B", "F"), prob = c(0.2, 0.5), stringsAsFactors = FALSE) score(res, learning.test, type = "bde", prior = "cs", beta = beta) ## equivalent to logLik(res, learning.test) score(res, learning.test, type = "loglik") ## equivalent to AIC(res, learning.test) score(res, learning.test, type = "aic")
Want to suggest features or report bugs for rdrr.io? Use the GitHub issue tracker.