Chapter 7 Estimation
7.1 Statistical Inference
Nothing on this section.
7.2 Prior and Posterior Distributions
Suppose that the proportion \(\theta\) of defective items in a lot is either \(0.1\) or \(0.2\) and the prior probability function is \(\xi(0.1)=0.6\) and \(\xi(0.2)=0.4\). Suppose that nine items were selected at random, and exactly two of them were defective. Determine the posterior distribution of \(\theta\).
- Suppose that the number of mice per hectare follows a \(X\sim \textrm{Poisson}(\lambda)\) distribution, and our prior distribution for \(\lambda\) is Gamma(\(\alpha=2, \beta=1\)) distribution. Suppose that we collect one observation \(x\).
- Show that \[\xi(\lambda | x) \propto g(x)h(x,\lambda) = \frac{1}{(x-1)!}\,e^{-2\lambda}\lambda^{x+1}\]
- Recognize \(h(x,\lambda)\) as the kernel of a Gamma distribution. What are the parameters of the posterior distribution?
7.3 Conjugate Prior Distributions
Suppose that the proportion \(\theta\) of defective items in a large shipment is unknown and that the prior distribution of \(\theta\) is \(\textrm{Beta}(\alpha=2, \beta=200)\). If 100 itiems are selected at random, and three are found to be defective, what is the posterior distribution of \(\theta\)?
Suppose that the number of defects in a roll of aluminum has a Poisson distribution with rate parameter \(\theta\). Suppose that we use a prior distribution \(\theta \sim \textrm{Gamma}(3, 1)\). Determine the posterior distribution of \(\theta\).
Suppose that \(X \sim \textrm{NegBinomial}(r,p)\) and \(r\) is known, but we are interested in inference on \(p\). Show that \(p \sim \textrm{Beta}(\alpha,\beta)\) is conjugate prior by deriving the posterior distribution.
- Suppose that the continuous random variable \(X_i \stackrel{iid}{\sim} \textrm{Uniform}(0, \theta)\). The conjugate prior distribution for \(\theta\) is the Pareto distribution, which has pdf \[f(x|x_0, \alpha) = \frac{\alpha x_0^\alpha}{x^{\alpha+1}}\;I(x > x_0)\]
- Graph the Pareto distribution with \(x_0=1\), and \(\alpha=2\) and again with \(\alpha=3\).
- Show that the joint distribution of \(X_1,\dots,X_n\) can be written in terms of the sample maximum, which you may denote \(X_{(n)}\). Make sure to use an indicator function to denote the support of the distribution.
- Consider the prior distribtion on \(\theta \sim \textrm{Pareto}(\theta_0, \alpha)\). Write down the prior pdf, again making sure to include your indicator function.
- If \(x_{(n)} \ge \theta_0\), what is the posterior distribution of \(\theta\)?
7.4 Bayes Estimators
- Supposet that a random sample of size \(n\) is taken from a Bernoulli distribution with unknown probability of success \(\theta\). As usual, we will assign the conjugate prior \(\textrm{Beta}(\alpha, \beta)\). Denote the mean of the prior distribution \(\mu_0\).
- Show that the Bayes estimator under squared error loss will be a weighted average of the sample mean and \(\mu_0\). That is, show that \[\hat{\theta}_n = \gamma_n \bar{X}_n + (1-\gamma_n)\mu_0\]
- Show that \(\hat{\theta}_n\) is a consistent estimator of \(\theta\).
- Suppose that a random sample of size \(n\) is taken from a Poisson distribution for which the value of the mean \(\theta\) is unknown, and the prior distribution of \(\theta\) is a gamma distribution for which the mean is \(\mu_0\).
- Show that \(\hat{\theta}\), the mean of the posterior distribution of \(\theta\), will be a weighted average of sample mean \(\bar{X}\) and the prior mean \(\mu_0\).
- Show that \(\hat{\theta}_n\) is a consistent estimator of \(\theta\).
- Suppose that \(X_i \stackrel{iid}{\sim} \textrm{Uniform}(0,\theta)\) and that the prior distribution for \(\theta\) is \(\textrm{Pareto}(\theta_0, \alpha)\) where \(\alpha > 1\). Suppose that we observe \(\max(x_i) > \theta_0\). What is the Bayes estimator of \(\theta\) under the squared error loss function?
7.5 Maximum Likelihood Estimators
- Suppose that \(X_1, \dots, X_n\) form a random sample from a Poisson(\(\theta\)) distribution for which the mean \(\theta\) is unknown and \(\theta > 0\).
- Determine the MLE of \(\theta\), assuming that at least one of the observed values is different than zero.
- Show that the MLE of \(\theta\) does not exist if every observed value is zero.
Suppose that \(X_1, \dots, X_n\) form a random sample from the normal distribution for which the mean \(\mu\) is known, but the variance, \(\sigma^2\), is unknown. Find the MLE of \(\sigma^2\). Hint, it is NOT the sample variance!
- Suppose that \(X_1, \dots, X_n\) form a random sample from a distribution with pdf \[f(x|\theta) = \theta x^{\theta-1} \, I( 0 < x < 1 )\] where \(\theta > 0\).
- Graph the distribution for various values of \(\theta\).
- Find the MLE of \(\theta\).
7.6 Properties of Maximum Likelihood Estimators
- Suppose that \(X_i \stackrel{iid}{\sim} \textrm{Exponential}(\theta)\) for a sample of size \(n\).
- Find the MLE of \(\theta\).
- Find the MLE of the median of the distribution.
Suppose that \(X_i \stackrel{iid}{\sim} N( \mu, \sigma^2 )\). Find the MLE of the 0.95 quantile of the distribution, that is, of the point \(\theta\) such that \(Pr( X \le \theta) = 0.95\).
Suppose that \(X_i \stackrel{iid}{\sim} \textrm{Gamma}(\alpha, \beta)\) where \(\alpha\) is known. Find the MLE of \(\beta\) and then also find the MLE of \(\theta = \alpha/\beta\).
Supppose that \(X_i \stackrel{iid}{\sim} \textrm{Beta}( \alpha, \beta )\). Find the Method of Moment estimators for \(\alpha\) and \(\beta\). Hint: This sort of calculation can be done easily in Mathematica or Wolfram Alpha. It really pays to get comfortable with some software packages.
7.7 Sufficient Statistics
Suppose that \(X_i \stackrel{iid}{\sim} \textrm{NegBinom}(r,p)\) where \(r\) is known. Show that \(T=\sum X_i\) is sufficient for \(p\).
Suppose that \(X_i \stackrel{iid}{\sim} \textrm{Gamma}(\alpha, \beta)\) where \(\alpha\) is known. Show that \(T=\sum X_i\) is sufficient for \(\beta\).
- Suppose that \(X_i \stackrel{iid}{\sim} \textrm{Gamma}(\alpha, \beta)\) where \(\beta\) is known.
- Show that \(T=\prod X_i\) is sufficient for \(\alpha\).
- Show that \(T=\sum \log(X_i)\) is also sufficient for \(\alpha\).