1 / 24

Econometrics

Chengyuan Yin School of Mathematics. Econometrics. Econometrics. 8. Hypothesis Testing in the Linear Regression Model. Classical Hypothesis Testing.

mcgheer
Download Presentation

Econometrics

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Chengyuan Yin School of Mathematics Econometrics

  2. Econometrics 8. Hypothesis Testing in the Linear Regression Model

  3. Classical Hypothesis Testing We are interested in using the linear regression to establish or cast doubt on the validity of a theory about the real world counterpart to our statistical model. The model is used to test hypotheses about the underlying data generating process.

  4. Inference in the Linear Model Hypothesis testing: Formulating hypotheses: linear restrictions as a general framework Substantive restrictions: What is a "testable hypothesis?" Nested vs. nonnested models Methodological issues Classical (likelihood based approach): Are the data consistent with the hypothesis? Bayesian approach: How do the data affect our prior odds?  The posterior odds ratio.

  5. Testing a Hypothesis About a Parameter: Confidence Interval bk = the point estimate Std.Dev[bk] = sqr{[σ2(X’X)-1]kk} = vk Assume normality of ε for now: • bk ~ N[βk,vk2] for the true βk. • (bk-βk)/vk ~ N[0,1] Consider a range of plausible values of βk given the point estimate bk. bk +/- sampling error. • Measured in standard error units, • |(bk – βk)/vk| < z* • Larger z*  greater probability (“confidence”) • Given normality, e.g., z* = 1.96  95%, z*=1.64590% • Plausible range for βk then is bk ± z* vk

  6. Estimating the Confidence Interval Assume normality of ε for now: • bk ~ N[βk,vk2] for the true βk. • (bk-βk)/vk ~ N[0,1] vk = [σ2(X’X)-1]kk is not known because σ2 must be estimated. Using s2 instead of σ2, (bk-βk)/est.(vk) ~ t[n-K]. (Proof: ratio of normal to sqr(chi-squared)/df is pursued in your text.) Use critical values from t distribution instead of standard normal.

  7. Testing a Hypothesis Using a Confidence Interval Given the range of plausible values The confidence interval approach. Testing the hypothesis that a coefficient equals zero or some other particular value: Is the hypothesized value in the confidence interval? Is the hypothesized value within the range of plausible values

  8. Wald Distance Measure Testing more generally about a single parameter. Sample estimate is bk Hypothesized value is βk How far is βk from bk? If too far, the hypothesis is inconsistent with the sample evidence. Measure distance in standard error units t = (bk - βk)/Estimated vk. If t is “large” (larger than critical value), reject the hypothesis.

  9. The Wald Statistic

  10. Robust Tests • The Wald test generally will (when properly constructed) be more robust to failures of the narrow model assumptions than the t or F • Reason: Based on “robust” variance estimators and asymptotic results that hold in a wide range of circumstances. • Analysis: Later in the course – after developing asymptotics.

  11. The General Linear Hypothesis: H0:R - q = 0 A unifying departure point: Regardless of the hypothesis, least squares is unbiased. Two approaches (1) Is Rb - q close to 0? Basing the test on the discrepancy vector: m = Rb - q. Using the Wald criterion: m(Var[m])-1m has a chi-squared distribution with J degrees of freedom But, Var[m] = R[2(X’X)-1]R. If we use our estimate of 2, we get an F[J,n-K], instead. (Note, this is based on using ee/(n-K) to estimate 2.) (2) We know that imposing the restrictions leads to a loss of fit. R2 must go down. Does it go down “a lot?” (I.e., significantly?). R2 = unrestricted model, R*2 = restricted model fit. F = { (R2 - R*2)/J } / [(1 - R2)/(n-K)] = F[J,n-K]. These are the same in the linear model

  12. t and F statistics An important relationship between t and F For a single restriction, m = r’b - q. The variance is r’(Var[b])r The distance measure is m / standard error of m. The t-ratio is the square root of the F ratio.

  13. Lagrange Multiplier Statistics Specific to the classical model: Recall the Lagrange multipliers:  = [R(XX)-1R]-1m. Suppose we just test H0:  = 0, using the Wald criterion. The resulting test statistic is just JF where F is the F statistic above. This is to be taken as a chi-squared statistic. (Note, again, using ee/(n-K) to estimate 2. If ee/n, instead, the more formal, likelihood based statistic results.)

  14. Application Time series regression, LogG = 1 + 2logY + 3logPG + 4logPNC + 5logPUC + 6logPPT + 7logPN + 8logPD + 9logPS +  Period = 1960 - 1995. A significant event occurs in October 1973. We will be interested to know if the model 1960 to 1973 is the same as from 1974 to 1995. Note that all coefficients in the model are elasticities.

  15. Full Model

  16. Test about one Parameter Is the price of public transportation really relevant? H0 : 6 = 0. Confidence interval: b6 t(.95,27)  Standard error = .1279  2.052(.0791) = .1279  .1623 = (-.0344 ,.2902) Contains 0.0. Do not reject hypothesis Distance measure: (b6 - 0) / sb6 = (.1279 - 0) / .0791 = 1.617 < 2.052. Regression fit if drop? Without LPPT, R-squared= .988215 Compare R2, was .989256, F(1,27) = [(.989256 - .988215)/1]/[(1-.989256)/(36-9)] = 2.616 = 1.61722 (!)

  17. Sum of Coefficients

  18. Imposing the Restrictions

  19. Joint Hypotheses

  20. Using the Wald Statistic

  21. Details • Which restriction is the problem? • After building the restrictions into the model and computing restricted and unrestricted regressions: Based on R2s, F = ((.989256 - .997237)/2)/((1-.989256)/(36-9)) = -.10.02824832465 (!) What's wrong?

  22. Chow Test

  23. Algebra for the Chow Test

More Related