GLmix | R Documentation |
Kiefer Wolfowitz Nonparametric MLE for Gaussian Location Mixtures
GLmix(x, v = 300, sigma = 1, hist = FALSE, histm = 300, weights = NULL, ...)
x |
Data: Sample Observations |
v |
Undata: Grid Values defaults equal spacing of with v bins, when v is a scalar |
sigma |
scale parameter of the Gaussian noise, may take vector values of length(x) |
hist |
If TRUE then aggregate x to histogram bins, when sigma is vector valued this option is inappropriate unless there are only a small number of distinct sigma values. |
histm |
histogram bin boundaries, equally spacing with |
weights |
replicate weights for x obervations, should sum to 1 |
... |
other parameters to pass to KWDual to control optimization |
Kiefer Wolfowitz MLE as proposed by Jiang and Zhang for
the Gaussian compound decision problem. The histogram option is intended
for large problems, say n > 1000, where reducing the sample size dimension
is desirable. When sigma
is heterogeneous and hist = TRUE
the
procedure tries to do separate histogram binning for distinct values of
sigma
, however this is only feasible when there are only a small
number of distinct sigma
. By default the grid for the binning is
equally spaced on the support of the data. This function does the normal
convolution problem, for gamma mixtures of variances see GVmix
, or
for mixtures of both means and variances TLVmix
.
The predict method for GLmix
objects will compute means, medians or
modes of the posterior according to whether the Loss
argument is 2, 1
or 0, or posterior quantiles if Loss
is in (0,1).
An object of class density with components:
x |
points of evaluation on the domain of the density |
y |
estimated function values at the points v, the mixing density |
g |
the estimated mixture density function values at x |
logLik |
Log likelihood value at the proposed solution |
dy |
prediction of mean parameters for each observed x value via Bayes Rule |
status |
exit code from the optimizer |
Roger Koenker
Kiefer, J. and J. Wolfowitz Consistency of the Maximum Likelihood Estimator in the Presence of Infinitely Many Incidental Parameters Ann. Math. Statist. Volume 27, Number 4 (1956), 887-906.
Jiang, Wenhua and Cun-Hui Zhang General maximum likelihood empirical Bayes estimation of normal means Ann. Statist., Volume 37, Number 4 (2009), 1647-1684.
Koenker, R and I. Mizera, (2013) “Convex Optimization, Shape Constraints, Compound Decisions, and Empirical Bayes Rules,” JASA, 109, 674–685.
Koenker, R. and J. Gu, (2017) REBayes: An R Package for Empirical Bayes Mixture Methods, Journal of Statistical Software, 82, 1–26.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.