$$ P(Y_{i}) = \begin{cases} \theta \ 1 - \theta \end{cases} $$
$$ E\left[Y_{i}\right] = \ $$
Given data (Y_{1}, Y_{2}, ., Y_{n},) write the Bernoulli model as a set of joint data densities, i.e. like our general formulation of the statistical model, (\mathcal{M}=\left{\mathrm{f}{\psi}\left(y{1}, y_{2}, ., y_{n}\right), \psi \in \Psi\right}, \Psi \subseteq \mathbb{R}^{k} .) State a condition that the model is correctly specified. This will be assumed in the following.
Show furthermore that these joint data densities (i.e. the elements of (\mathcal{M}) ) can be written in terms of the sample mean, (\bar{y}) Hintt: use that, for numbers (a_{i}) and (b_{i}) it holds that, (\prod_{i=1}^{n} a_{i} b_{i}=\prod_{i=1}^{n} a_{i} \Pi_{i}^{n} \Pi_{i=1}^{n} b_{i}) Hint2: use eq. 1.3 .1 in Hint 3: use the definition of the sample meanl.
State the log-likelihood function for the Bernoulli model and go through the derivations leading to the (\mathrm{MLE}, \widehat{\theta})
Show that (\widehat{\theta}) is unbiased and that the variance of it is (\frac{V a r\left[Y_{1}\right]}{n} .) What is the standard error of (\widehat{\theta} .)
Consider the distributional statement (\left.\widehat{\theta} \stackrel{P}{\rightarrow} \theta_{0}, \text { where } \theta_{0} \in\right] 0 ; 1[) is the population value/true value. Explain what it means and what assumptions are needed.
What does it mean that (\widehat{\theta}) is asymptotically normally (Gaussian) distributed? Explain why it holds.
In 5% of the cases, hypothaticcaly repeated samples are out in the ends, CI will not include theta-node
In 95% of the cases, CI will include the true value.
CI is random, not the true value.
consistent
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.