




版權說明:本文檔由用戶提供并上傳,收益歸屬內容提供方,若內容存在侵權,請進行舉報或認領
文檔簡介
1、Econometrics IProfessor William GreeneStern School of BusinessDepartment of EconomicsEconometrics IPart 6 Finite Sample Properties of Least SquaresTerms of ArtEstimates and estimatorsProperties of an estimator - the sampling distribution“Finite sample” properties as opposed to “asymptotic” or “l(fā)arge
2、 sample” propertiesApplication: Health Care Panel DataGerman Health Care Usage Data, 7,293 Individuals, Varying Numbers of PeriodsData downloaded from Journal of Applied Econometrics Archive. There are altogether 27,326 observations. The number of observations ranges from 1 to 7. (Frequencies are: 1
3、=1525, 2=2158, 3=825, 4=926, 5=1051, 6=1000, 7=987).Variables in the file are DOCVIS = number of doctor visits in last three months HOSPVIS = number of hospital visits in last calendar year DOCTOR = 1(Number of doctor visits 0) HOSPITAL= 1(Number of hospital visits 0) HSAT = health satisfaction, cod
4、ed 0 (low) - 10 (high) PUBLIC = insured in public health insurance = 1; otherwise = 0 ADDON = insured by add-on insurance = 1; otherswise = 0 HHNINC = household nominal monthly net e in German marks / 10000. (4 observations with e=0 were dropped) HHKIDS = children under age 16 in the household = 1;
5、otherwise = 0 EDUC = years of schooling AGE = age in years MARRIED = marital statusFor now, treat this sample as if it were a cross section, and as if it were the full population.Population RegressionSampling DistributionA sampling experiment: Draw 25 observations at random from the population. Comp
6、ute the regression. Repeat 100 times. Display estimates.matrix ; beduc=init(100,1,0)$proc$draw ; n=25 $regress; quietly ; lhs=hhninc ; rhs = one,educ $matrix ; beduc(i)=b(2) $sample;all$endproc$execute ; i=1,100 $histogram;rhs=beduc $How should we interpret this variation in the regression slope?The
7、 Statistical Context of Least Squares EstimationThe sample of data from the population: Data generating process is y = x + The stochastic specification of the regression model: Assumptions about the random .Endowment of the stochastic properties of the model upon the least squares estimator. The est
8、imator is a function of the observed (realized) data. Least SquaresDeriving the Propertiesb = a parameter vector + a linear combination of the disturbances, each times a vector.Therefore, b is a vector of random variables. We analyze it as such.The assumption of nonstochastic regressors. How it is u
9、sed at this point.We do the analysis conditional on an X, then show that results do not depend on the particular X in hand, so the result must be general i.e., independent of X. Properties of the LS Estimator: b is unbiasedExpected value and the property of unbiasedness.Eb|X = E + (XX)-1X|X = + (XX)
10、-1XE|X = + 0Eb = EXEb|X = Eb.(The law of iterated expectations.)Sampling ExperimentMeans of Repetitions b|xPartitioned RegressionA Crucial Result About Specification: y = X11 + X22 + Two sets of variables. What if the regression is computed without the second set of variables?What is the expectation
11、 of the short regression estimator? Eb1|(y = X11 + X22 + ) b1 = (X1X1)-1X1yThe Left Out Variable Formula“Short” regression means we regress y on X1 when y = X11 + X22 + and 2 is not 0(This is a VVIR!) b1 = (X1X1)-1X1y = (X1X1)-1X1(X11 + X22 + ) = (X1X1)-1X1X11 + (X1X1)-1X1 X22 + (X1X1)-1X1) Eb1 = 1
12、+ (X1X1)-1X1X22ApplicationThe (truly) short regression estimator is biased.Application: Quantity = 1Price + e + If you regress Quantity on Price and leave out e. What do you get?Application: Left out VariableLeave out e. What do you get? In time series data, 1 0 (usually)CovPrice, e 0 in time series
13、 data.So, the short regression will overestimate the price coefficient. It will be pulled toward and even past zero.Simple Regression of G on a constant and PGPrice Coefficient should be negative.Estimated Demand EquationShouldnt the Price Coefficient be Negative?Multiple Regression of G on Y and PG
14、. The Theory Works!-Ordinary least squares regression .LHS=G Mean = 226.09444 Standard deviation = 50.59182 Number of observs. = 36Model size Parameters = 3 Degrees of freedom = 33Residuals Sum of squares = 1472.79834 Standard error of e = 6.68059Fit R-squared = .98356 Adjusted R-squared = .98256Mod
15、el test F 2, 33 (prob) = 987.1(.0000)-+-Variable| Coefficient Standard Error t-ratio P|T|t Mean of X-+-Constant| -79.7535* 8.67255 -9.196 .0000 Y| .03692* .00132 28.022 .0000 9232.86 PG| -15.1224* 1.88034 -8.042 .0000 2.31661-+-The Extra Variable FormulaA Second Crucial Result About Specification: y
16、 = X11 + X22 + but 2 really is 0.Two sets of variables. One is superfluous. What if the regression is computed with it anyway?The Extra Variable Formula: (This is a VIR!) Eb1.2| 2 = 0 = 1The long regression estimator in a short regression is unbiased.)Extra variables in a model do not induce biases.
17、 Why not just include them? Variance of bAssumption about disturbances:i has zero mean and is uncorrelated with every other j Vari|X = 2. The variance of i does not depend on any data in the sample. Variance of the Least Squares EstimatorVariance of the Least Squares EstimatorSpecification Errors-1O
18、mitting relevant variables: Suppose the correct model is y = X11 + X22 + . I.e., two sets of variables. Compute least squares omitting X2. Some easily proved results:Varb1 is smaller than Varb1.2. (The latter is the northwest submatrix of the full covariance matrix. The proof uses the residual maker
19、 (again!). I.e., you get a smaller variance when you omit X2. (One interpretation: Omitting X2 amounts to using extra information (2 = 0). Even if the information is wrong (see the next result), it reduces the variance. (This is an important result.)Omitted Variables(No free lunch) Eb1 = 1 + (X1X1)-
20、1X1X22 1. So, b1 is biased.(!) The bias can be huge. Can reverse the sign of a price coefficient in a “demand equation.” b1 may be more “precise.” Precision = Mean squared error = variance + squared bias. Smaller variance but positive bias. If bias is small, may still favor the short regression.(Fre
21、e lunch?) Suppose X1X2 = 0. Then the bias goes away. Interpretation, the information is not “right,” it is irrelevant. b1 is the same as b1.2.Specification Errors-2 Including superfluous variables: Just reverse the results.Including superfluous variables increases variance. (The cost of not using in
22、formation.)Does not cause a bias, because if the variables in X2 are truly superfluous, then 2 = 0, so Eb1.2 = 1. Linear RestrictionsContext: How do linear restrictions affect the properties of the least squares estimator? Model: y = X + Theory (information) R - q = 0Restricted least squares estimat
23、or: b* = b - (XX)-1RR(XX)-1R-1(Rb - q)Expected value: Eb* = - (XX)-1RR(XX)-1R-1(Rb - q)Variance: 2(XX)-1 - 2 (XX)-1RR(XX)-1R-1 R(XX)-1 = Varb a nonnegative definite matrix VarbImplication: (As before) nonsample information reduces the variance of the estimator.InterpretationCase 1: Theory is correct
24、: R - q = 0 (the restrictions do hold). b* is unbiased Varb* is smaller than Varb How do we know this?Case 2: Theory is incorrect: R - q 0 (the restrictions do not hold). b* is biased what does this mean? Varb* is still smaller than VarbRestrictions and InformationHow do we interpret this important
25、result? The theory is information Bad information leads us away from the truth Any information, good or bad, makes us more certain of our answer. In this context, any information reduces variance.What about ignoring the information? Not using the correct information does not lead us away from the tr
26、uth Not using the information foregoes the variance reduction - i.e., does not use the ability to reduce uncertainty.Gauss-Markov TheoremA theorem of Gauss and Markov: Least Squares is the minimum variance linear unbiased estimator (MVLUE) 1. Linear estimator2. Unbiased: Eb|X = Theorem: Varb*|X Varb
27、|X is nonnegative definite for any other linear and unbiased estimator b* that is not equal to b.Definition: b is efficient in this class of estimators.Implications of Gauss-MarkovTheorem: Varb*|X Varb|X is nonnegative definite for any other linear and unbiased estimator b* that is not equal to b. Implies:bk = the kth particular element of b.Varbk|X = the kth diagonal element of Varb|XVarbk|X Varbk*|X for each coefficient.cb = any linear combination of
溫馨提示
- 1. 本站所有資源如無特殊說明,都需要本地電腦安裝OFFICE2007和PDF閱讀器。圖紙軟件為CAD,CAXA,PROE,UG,SolidWorks等.壓縮文件請下載最新的WinRAR軟件解壓。
- 2. 本站的文檔不包含任何第三方提供的附件圖紙等,如果需要附件,請聯系上傳者。文件的所有權益歸上傳用戶所有。
- 3. 本站RAR壓縮包中若帶圖紙,網頁內容里面會有圖紙預覽,若沒有圖紙預覽就沒有圖紙。
- 4. 未經權益所有人同意不得將文件中的內容挪作商業(yè)或盈利用途。
- 5. 人人文庫網僅提供信息存儲空間,僅對用戶上傳內容的表現方式做保護處理,對用戶上傳分享的文檔內容本身不做任何修改或編輯,并不能對任何下載內容負責。
- 6. 下載文件中如有侵權或不適當內容,請與我們聯系,我們立即糾正。
- 7. 本站不保證下載資源的準確性、安全性和完整性, 同時也不承擔用戶因使用這些下載資源對自己和他人造成任何形式的傷害或損失。
最新文檔
- T/CNCA 048-2023礦用防爆永磁同步伺服電動機通用技術條件
- 文安消費廣場景觀設計方案
- 書籍承印合同樣本6篇
- 公司委托合同書范本5篇
- 健康促進知識課件
- 2025二手安置房買賣合同5篇
- 車間衛(wèi)生標準化管理體系
- 2025遼寧開放大學輔導員考試試題及答案
- T/ZHCA 010-2020染發(fā)類化妝品皮膚變態(tài)反應體外測試方法人源細胞系激活試驗法
- 2025焦作職工醫(yī)學院輔導員考試試題及答案
- 2025年船舶駕駛員考試試卷及答案
- 制造部生產效率提升計劃
- 寵物丟失諒解協議書
- 幼兒園中班科學活動公開課《飛機本領大》課件
- 體育競彩考試題及答案
- 中國日用器皿行業(yè)市場前景預測及投資價值評估分析報告
- 2025年天津市西青區(qū)九年級二模數學試題(含部分答案)
- 2025年企業(yè)人力資源管理師考試真題及答案
- 2025-2030年中國玻璃容器行業(yè)市場發(fā)展趨勢與前景展望戰(zhàn)略分析報告
- 山東省濟南市2025屆高三三模化學試卷(含答案)
- 延長縣黑家堡共大灘防洪工程環(huán)境影響評價報告書
評論
0/150
提交評論