Question

Consider the simple linear regression model yi-Xißl + β0 + Ei, where the errors €1, €n are i.i.d. random variables with E[c]-0, var(G) σ2, i 1, , n. Solve either one of the questions below. 1. Let h be the least squares estimator for β1- Show that Bi is the best linear unbiased estimator for β1. (Note: you can read the proof in wikipedia, but you cannot use the matrix notation in this proof.) 2. Consider a new loss function Lx(βίΆ) TL where λ > 0 is a positive constant. Let (βιΑνβαλ) be the minimizer for (2). Show that var(B)var(1).

#2
can you please go through the steps of how you got this

0 0
Add a comment Improve this question Transcribed image text
Know the answer?
Add Answer to:
#2 can you please go through the steps of how you got this Consider the simple...
Your Answer:

Post as a guest

Your Name:

What's your source?

Earn Coins

Coins can be redeemed for fabulous gifts.

Not the answer you're looking for? Ask your own homework help question. Our experts will answer your question WITHIN MINUTES for Free.
Similar Homework Help Questions
  • Consider the simple linear regression model y - e, where the errors €1, ,en are iid....

    Consider the simple linear regression model y - e, where the errors €1, ,en are iid. random variables with Eki-0, var(G)-σ2, i-1, .. . ,n. Solve either one of the questions below. 1. Let Bi be the least squares estimator for B. Show that B is the best linear unbiased estimator for B1. (Note: you can read the proof in wikipedia, but you cannot use the matrix notation in this proof.) 2. Consider a new loss function Lx(A,%) 71 where...

  • 4. (24 marks) Suppose that the random variables Yi,..., Yn satisfy Y-B BX,+ Ei, 1-1, , n, where βο and βι are parameters, X1, ,X, are con- stants, and e1,... ,en are independent and identically distr...

    4. (24 marks) Suppose that the random variables Yi,..., Yn satisfy Y-B BX,+ Ei, 1-1, , n, where βο and βι are parameters, X1, ,X, are con- stants, and e1,... ,en are independent and identically distributed ran- dom variables with Ei ~ N (0,02), where σ2 is a third unknown pa- rameter. This is the familiar form for a simple linear regression model, where the parameters A, β, and σ2 explain the relationship between a dependent (or response) variable Y...

  • 1. Consider the simple linear regression model: Ү, — Во + B а; + Ei, where 1, . . , En are i.i.d. N(0,02), for i1,2,......

    1. Consider the simple linear regression model: Ү, — Во + B а; + Ei, where 1, . . , En are i.i.d. N(0,02), for i1,2,... ,n. Let b1 = s^y/8r and bo = Y - b1 t be the least squared estimators of B1 and Bo, respectively. We showed in class, that N(B; 02/) Y~N(BoB1 T;o2/n) and bi ~ are uncorrelated, i.e. o{Y;b} We also showed in class that bi and Y 0. = (a) Show that bo is...

  • 3. Consider the linear model: Yİ , n where E(Ei)-0. Further α +Ari + Ei for...

    3. Consider the linear model: Yİ , n where E(Ei)-0. Further α +Ari + Ei for i 1, assume that Σ.r.-0 and Σ r-n. (a) Show that the least square estimates (LSEs) of α and ß are given by à--Ỹ and (b) Show that the LSEs in (a) are unbiased. (c) Assume that E(e-σ2 Yi and E(49)-0 for all i where σ2 > 0. Show that V(β)--and (d) Use (b) and (c) above to show that the LSEs are consistent...

  • Exercise 1 Consider the regression model through the origin y.-β1zi-ci, where Ei ~ N(0,o). It is ...

    Exercise 2b please! Exercise 1 Consider the regression model through the origin y.-β1zi-ci, where Ei ~ N(0,o). It is assumed that the regression line passes through the origin (0, 0) that for this model a: T N, is an unbiased estimator of o2. a. Show d. Show that (n-D2 ~X2-1, where se is the unbiased estimator of σ2 from question (a). Exercise2 Refer to exercise 1 a. Show that is BLUE (best linear unbiased estimator) b. Show that +1 has...

  • Consider a simple linear regression model with nonstochastic regressor: Yi = β1 + β2Xi + ui....

    Consider a simple linear regression model with nonstochastic regressor: Yi = β1 + β2Xi + ui. 1. [3 points] What are the assumptions of this model so that the OLS estimators are BLUE (best linear unbiased estimates)? 2. [4 points] Let βˆ and βˆ be the OLS estimators of β and β . Derive βˆ and βˆ. 12 1212 3. [2 points] Show that βˆ is an unbiased estimator of β .22

  • 2. Consider a simple linear regression model for a response variable Yi, a single predictor variable...

    2. Consider a simple linear regression model for a response variable Yi, a single predictor variable ri, i-1,... , n, and having Gaussian (i.e. normally distributed) errors Ý,-BzitEj, Ejį.i.d. N(0, σ2) This model is often called "regression through the origin" since E(Yi) 0 if xi 0 (a) Write down the likelihood function for the parameters β and σ2 (b) Find the MLEs for β and σ2, explicitly showing that they are unique maximizers of the likelihood function. (Hint: The function...

  • S y, and that yi-μ +Ei. You can assume that Ele]-0 for all i, Ele: -σ2...

    S y, and that yi-μ +Ei. You can assume that Ele]-0 for all i, Ele: -σ2 for all i, and Ele#3-0 for all i j You want to estimate a sample mean, and your friend tells you to use the following estimator: uppose that vou have collected n observations on where w is a known sample weight for observation i (this means w; is non-random) (a) Find E( (b) Under what conditions, if any, is p an unbiased estimator? Under...

  • 2. Consider a simple linear regression i ion model for a response variable Y, a single...

    2. Consider a simple linear regression i ion model for a response variable Y, a single predictor variable ,i1.., n, and having Gaussian (i.e. normally distributed) errors: This model is often called "regression through the origin" since E(X) = 0 if xi = 0 (a) Write down the likelihood function for the parameters β and σ2 (b) Find the MLEs for β and σ2, explicitly showing that they are unique maximizers of the likelihood function Hint: The function g(x)log(x) +1-x...

  • 5) Consider the simple linear regression model N(0, o2) i = 1,...,n Let g be the...

    5) Consider the simple linear regression model N(0, o2) i = 1,...,n Let g be the mean of the yi, and let â and ß be the MLES of a and B, respectively. Let yi = â-+ Bxi be the fitted values, and let e; = yi -yi be the residuals a) What is Cov(j, B) b) What is Cov(â, ß) c) Show that 1 ei = 0 d) Show that _1 x;e; = 0 e) Show that 1iei =...

ADVERTISEMENT
Free Homework Help App
Download From Google Play
Scan Your Homework
to Get Instant Free Answers
Need Online Homework Help?
Ask a Question
Get Answers For Free
Most questions answered within 3 hours.
ADVERTISEMENT
ADVERTISEMENT
ADVERTISEMENT