1. (20 points) Let X1....X be a random sample from a uniform distribution over [0,0]. (a)...
5. Let X1,.. ., Xn be a random sample from Uniform(0,0) with an unknown endpoint θ > 0, we want to estimate the parameter θ (a) Find the method of moments estimator (MME) of θ. (b) Find the MLE θ of θ (c) (R) Set the sample size as 25, do a simulation in R to compare these two esti- mators in terms of their bias and variance. Include a side-by-side boxplot that compares their sampling distributions
Let X1,X2Xn be a random sample from a uniform distribution on the interval (0,0) (a) Show that the density function of Xcp-minXXXn) is given by n-1 72 0 otherwise (b) Use (a) to calculate E[Xcu]. Calculate the bias, B(6). Find a function of Xo) that is an unbiased estimator of 0
Let X1, . . . , Xn be a random sample from a population with density 8. Let Xi,... ,Xn be a random sample from a population with density 17 J 2.rg2 , if 0<、〈릉 0 , if otherwise ( a) Find the maximum likelihood estimator (MLE) of θ . (b) Find a sufficient statistic for θ (c) Is the above MLE a minimal sufficient statistic? Explain fully.
1. Let X1, X2,...,x. be a random sample from the unif(0,0) distribution (a) Find an unbiased estimatior of O based on the sample mean X (b) Find an unbiased estimator of based on the sample maximum X (c) Which estimator is better in terms of variance?
2. Let X1,..., Xn be i.i.d. according to a normal distribution N(u,02). (a) Get a sufficient statistic for u. Show your work. (b) Find the maximum likelihood estimator for u. (c) Show that the MLE in part (b) is an unbiased estimator for u. (d) Using Basu's theorem, prove that your MLE from before and sº, the sample variance, are independent. (Hint: use W; = X1-0 and (n-1)32)
a) Consider a random sample {X1, X2, ... Xn} of X from a uniform distribution over [0,0], where 0 <0 < co and e is unknown. Is п Х1 п an unbiased estimator for 0? Please justify your answer. b) Consider a random sample {X1,X2, ...Xn] of X from N(u, o2), where u and o2 are unknown. Show that X2 + S2 is an unbiased estimator for 2 a2, where п п Xi and S (X4 - X)2. =- п...
2 Let X1, X2, ...,X, be independent continuous random variables from the following distribution: f(3) = ox-(0-1) where : > 1 and a > 1 You may use the fact: E[X]- .- 2.1 Show that the maximum likelihood estimator of a is ômle = Ei log Xi 2.3 Derive a sufficient statistic for a. What theorem are you using to determine sufficiency? 2.4 Show that the fisher information in the whole sample is: 1(a)= 2.5 What Cramer Rao lower bound...
Exercise: Let Yİ,Y2, ,, be a random sample from a Gamma distribution with parameters and β. Assume α > 0 is known. a. Find the Maximum Likelihood Estimator for β. b. Show that the MLE is consistent for β. c. Find a sufficient statistic for β. d. Find a minimum variance unbiased estimator of β. e. Find a uniformly most powerful test for HO : β-2 vs. HA : β > 2. (Assume P(Type!Error)- 0.05, n 10 and a -...
7. Let X1,....Xn random sample from a Bernoulli distribution with parameter p. A random variable X with Bernoulli distribution has a probability mass function (pmf) of with E(X) = p and Var(X) = p(1-p). (a) Find the method of moments (MOM) estimator of p. (b) Find a sufficient statistic for p. (Hint: Be careful when you write the joint pmf. Don't forget to sum the whole power of each term, that is, for the second term you will have (1...
1. Let X1, ..., Xn be a random sample from a distribution with the pdf le-x/0, x > 0, N = (0,00). (a) Find the maximum likelihood estimator of 0. (b) Find the method of moments estimator of 0. (c) Are the estimators in a) and b) unbiased? (d) What is the variance of the estimators in a) and b)? (e) Suppose the observed sample is 2.26, 0.31, 3.75, 6.92, 9.10, 7.57, 4.79, 1.41, 2.49, 0.59. Find the maximum likelihood...