Nothing
# Copyright (C) 2012 - 2018 Paul Fink
#
# This file is part of imptree.
#
# imptree is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 2 of the License, or
# (at your option) any later version.
#
# imptree is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with imptree. If not, see <https://www.gnu.org/licenses/>.
#' @name imptree-package
#'
#' @title imptree: Classification Trees with Imprecise Probabilities
#'
#' @description The \code{imptree} package implements the creation of
#' imprecise classification trees based on algorithm developed by
#' Abellan and Moral.
#' The credal sets of the classification variable within each node
#' are estimated by either the imprecise Dirichlet model (IDM) or the
#' nonparametric predictive inference (NPI).
#' As split possible split criteria serve the 'information gain',
#' based on the maximal entropy distribution, and the adaptable
#' entropy-range based criterion propsed by Fink and Crossman.
#' It also implements different correction terms for the entropy.
#'
#' The performance of the tree can be evaluated with respect to the
#' common criteria in the context of imprecise classification trees.
#'
#' It also provides the functionality for estimating credal sets via
#' IDM or NPI and obtain their minimal/maximal entropy (distribution)
#' to be used outside the tree growing process.
#'
#'
#' @references Abell\ifelse{latex}{\out{\'{a}}}{\ifelse{html}{\out{á}}{a}}n,
#' J. and Moral, S. (2005), Upper entropy of credal sets. Applications to
#' credal classification, \emph{International Journal of Approximate Reasoning}
#' \bold{39}, pp. 235--255.
#' @references Baker, R. M. (2010), \emph{Multinomial Nonparametric Predictive Inference:
#' Selection, Classification and Subcategory Data}, PhD thesis. Durham University, GB.
#' @references Strobl, C. (2005), Variable Selection in Classification Trees Based on
#' Imprecise Probabilities, \emph{ISIPTA '05: Proceedings of the Fourth
#' International Symposium on Imprecise Probabilities and Their Applications},
#' 339--348.
#' @references Fink, P. and Crossman, R.J. (2013), Entropy based classification trees,
#' \emph{ISIPTA '13: Proceedings of the Eighth International Symposium on Imprecise
#' Probability: Theories and Applications}, pp. 139--147.
#'
#' @seealso
#' \code{\link{imptree}} for tree creation, \code{\link{probInterval}} for the credal set
#' and entropy estimation functionality
#'
#' @examples
#' data("carEvaluation")
#'
#' ## create a tree with IDM (s=1) to full size
#' ## carEvaluation, leaving the first 10 observations out
#' ip <- imptree(acceptance~., data = carEvaluation[-(1:10),],
#' method="IDM", method.param = list(splitmetric = "globalmax", s = 1),
#' control = list(depth = NULL, minbucket = 1))
#'
#' ## summarize the tree and show performance on training data
#' summary(ip)
#'
#' ## predict the first 10 observations
#' ## Note: The result of the prediction is return invisibly
#' pp <- predict(ip, dominance = "max", data = carEvaluation[(1:10),])
#' ## print the general evaluation statistics
#' print(pp)
#' ## display the predicted class labels
#' pp$classes
#'
#' @keywords tree
#'
#' @docType package
#'
#' @useDynLib imptree
#' @importFrom Rcpp sourceCpp
#' @importFrom stats na.fail
NULL
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.