LRMoE contains all the parameters of a Logistic Regularized Mixture-of-Experts.
X
The matrix data for the input.
Y
Vector of the response variable.
d
Numeric. Number of explanatory variables (including the intercept variable).
n
Numeric. Length of the response/output vector Y
.
R
Numeric. Maximum value of Y
.
K
Number of expert classes.
Lambda
Penalty value for the expert part.
Gamma
Penalty value for the gating network.
wk
Parameters of the gating network. Matrix of dimension (K - 1,
d), with d
the number of explanatory variables (including the
intercept).
eta
Values of the regression coefficients for each level r = 1,...,R. Array of dimension (K, R-1, d).
loglik
Numeric. Observed-data log-likelihood of the LRMoE model.
storedloglik
Numeric vector. Stored values of the log-likelihood at each EM iteration.
BIC
Numeric. Value of BIC (Bayesian Information Criterion).
zerocoeff
Matrix. Proportion of zero coefficients obtained during each
iteration of the EM. First column gives the number of zero coefficients for
wk
and the second column for eta
.
Cluster
Numeric vector. Clustering label for each observation.
plot(what = c("loglik", "zerocoefficients"))
Plot method.
what
The type of graph requested:
"loglik" =
Value of the log-likelihood for
each iteration.
"zerocoefficients" =
Proportion of zero
coefficients for each iteration.
By default, all the above graphs are produced.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.