Math 152 - Statistical Theory - Homework 9
Important Note:
You should work to turn in assignments that are clear, communicative, and concise. Part of what you need to do is not print pages and pages of output. Additionally, you should remove these exact sentences and the information about HW scoring below.
Click on the Knit to PDF icon at the top of R Studio to run the R code and create a PDF document simultaneously. [PDF will only work if either (1) you are using R on the network, or (2) you have LaTeX installed on your computer. Lightweight LaTeX installation here: https://yihui.name/tinytex/]
Either use the college’s RStudio server (https://rstudio.pomona.edu/) or install R and R Studio on to your personal computer. See: https://m152-stat-theory.netlify.app/syllabus.html for resources.
Assignment
Goals:
In this assignment, the fun will include:
- calculating Fisher information
- setting up hypotheses
- calculationg power and error rates.
Book problems
- Feel free to do the book problems with a pencil or in LaTeX (RMarkdown supports writing mathematics using LaTeX).
- If you use a pencil, you can take a picture of the problem(s), and include the image(s) using (remove the tick marks to make it work):

Note that myimage.jpeg needs to live in the same folder as the relevant .Rmd file (maybe you called the folder “math 152 hw” and put it on your desktop?)
Saving as jpg, jpeg, png, or pdf should work, but make sure to specify the exact name of the file.
If you have the 3rd edition of the book, the problems will be the same unless they don’t exist – that is, the 4th edition added problems but didn’t change the order of them. Ask me if you want to see the 4th edition problems.
Assignment
1: Community Q
Describe one thing you learned (not from lecture, maybe from working in pairs during class) from a member of the class (student, mentor, professor) – it could be: content, logistical help, background material, R information, etc. 1-3 sentences.
2: 8.8.4
Suppose that a random variable has the normal distribution with mean 0 and unknown standard deviation
3: 8.8.5
Suppose that a random variable X has the normal distribution with mean 0 and unknown variance
Also, show that the unbiased estimator of
Note to self: there is no simple way to get from problem 8.8.4 to 8.8.5. That is, we know the MLE of a function is that function of the MLE. Not true for information.
4: 8.8.16
Suppose that
Note to self: this is a Bayesian problem. See the connection between Fisher Information and asymptotic / approximate Bayesian distributions in the text.
5: 8.9.15
Suppose that
where the value of
6: 9.1.1
Let
Consider the test procedure
- Determine the power function of the test.
- Compute the size of the test.
7: 9.1.6
Suppose that a single observation
Construct a test procedure
8: 9.1.9
Assume that
Find a test statistic
9: R -
We are going to study almost, but not exactly, the same model as from the exam. The model for this problem is normal with mean and standard deviation both
The results from class about properties of MLEs are asymptotic. What happens in small samples?
The estimators of
- the sample median
- the sample mean
- the sample standard deviation times the sign of the sample mean
- the MLE
The MLE of
is . Show (with pencil / LaTeX) that the Fisher Information in is .Use a simulation to compare the four estimators above with respect to bias, variance, and MSE. Answer the following questions in your comparison:
- Which estimator is (empirically) least biased?
- Which estimator has lowest empirical variability? Do any of the estimators reach the CRLB (assume unbiasedness)?
- Which estimator has lowest empirical MSE?
- Are you comfortable with the idea of using a normal distribution to describe the sampling distribution for any/all of the estimators? Explain.
- Which estimator would you recommend for the given setting? Explain.
# Use sample size n = 15. keep this set-up in the first 3 lines
<- 15
n.obs <- 10000
n.samps <- exp(1)
theta
<- numeric(n.samps)
means <- numeric(n.samps)
medians <- numeric(n.samps)
sds <- numeric(n.samps)
MLEs
for (i in 1:n.samps){
# generate some data from the given model
# means[i] <- mean(the data you generated)
# etc.
}
You can write alternative code for calculating relevant characteristics of the distribution and displaying it, but I choose to put it together in a tidy framework like this:
Note: in the code below I’ve created smoothed histograms (called density plots) so as to plot the empirical distributions on top of one another.
library(tidyverse)
est.info <- data.frame(value = c(means , ___, ...),
type = c(rep("mean", n.samps), ____, ...) )
est.info %>%
group_by(type) %>%
summarize(est.bias = mean(value) - theta, est.mean = mean(value),
est.var = var(value), est.sd = sd(value)) %>%
mutate(est.mse = est.var + est.bias^2)
est.info %>%
ggplot(aes(x=value, color = type)) + geom_density() +
geom_vline(xintercept = exp(1))