Please Do Q2 only.......... 1 = .*.vk-1e-ta 1. Suppose that X is a random variable with...
Please solve this by
explaining everything.
1 1. Suppose that X is a random variable with a Gamma-(k, 1) distribution where k > 0 is known, but > 0 is unknown Vx € (0,1), we have fx(x) 1k2k-1e-x T(k) Let us use 0 = 1/1 which is the standard approach, for example in Hogg and Tanis. Calculate the Fisher information I(0).
Problem 2: Let (X1,... Xn) denote a random variable from X having density fx(x) = 1/ β,0 < x < β where β > 0 is an unknown param eter. Explain why the Cramer Rao Theorem cannot be applied to show that an unbiased estimator of β is MVU. (Hint: see slides. Condition (A) of Cramer Rao Theorem)
3. Suppose X1, X2, , Xn are iid based on the random variable modeled by 2,0-1 (1-2)a-1 where 0 < x < 1 and α > 0 a. Find an equation that the MLE for a must satisfy. Note: You will not be able to explicitly solve for the MLE as in other problems b. If you are told E(X) = 2 and Var(X) = 8a14, example where someone might prefer the MME over the MLE find the MME for...
Let X1, . . . , Xn be a random sample from a population X with p.d.f fθ(x) = θ xθ−1 , for 0 < x < 1 0, otherwise, where θ > 1 is parameter. Find the MLE of 1/θ. If it is an unbiased estimator of 1/θ, compare its variance with the Cramer-Rao lower bound.
7. Let X1,....Xn random sample from a Bernoulli distribution with parameter p. A random variable X with Bernoulli distribution has a probability mass function (pmf) of with E(X) = p and Var(X) = p(1-p). (a) Find the method of moments (MOM) estimator of p. (b) Find a sufficient statistic for p. (Hint: Be careful when you write the joint pmf. Don't forget to sum the whole power of each term, that is, for the second term you will have (1...
Suppose that X1,X2,. X are iid random variables with pdf ,220 (a) Find the maximum likelihood estimate of the parameter a (b) Find the Fisher Information of X1,X2,.. ., Xn and use it to estimate a 95% confidence interval on the MLE of a (c) Explain how the central limit theorem relates to (b).
Suppose X1, X2, . . . , Xn are iid based on the random variable modeled by where 0 ≤ x ≤ 1 and α > 0. a. Find an equation that the MLE for α must satisfy. Note: You will not be able to explicitly solve for the MLE as in other problems. b. If you are told E(X) = 1/2 and Var(X) = 1/(8α + 4), find the MME for α. This problem is a nice example where...
1. (40) Suppose that X1, X2, .. , Xn, forms an normal distribution with mean /u and variance o2, both unknown: independent and identically distributed sample from 2. 1 f(ru,02) x < 00, -00 < u < 00, o20 - 00 27TO2 (a) Derive the sample variance, S2, for this random sample (b) Derive the maximum likelihood estimator (MLE) of u and o2, denoted fi and o2, respectively (c) Find the MLE of 2 (d) Derive the method of moment...
2 Let X1, X2, ..., X, be independent continuous random variables from the following distribution: f(x) = or-(-) where x 2 1 and a > 1 You may use the fact: E[X] - - 2.4 Show that the fisher information in the whole sample is: In(a)= 2.5 What Cramer Rao lower bound for unbiased estimators of a? 2.7 Consider estimating the unknown quantity: g(a) = 0 - 4+.. Determine the MLE of gla). What property are you using to justify...
Suppose X1, X2, ..., Xn is an iid sample from fx(r ja-θ(1-z)0-11(0 1), where x θ>0. (a) Find the method of moments (MOM) estimator of θ. (b) Find the maximum likelihood estimator (MLE) of θ (c) Find the MLE of Po(X 1/2) d) Is there a function of θ, say T 0), for which there exists an unbiased estimator whose variance attains the Cramér-Rao Lower Bound? If so, find it and identify the corresponding estimator. If not, show why not.