Linear Regression Model and the OLS Estimator
Effect of Heteroskedasticity
Topic of the module
Understand the effect of heteroskedasticity on the sampling distribution of the OLS estimator.
Data generating process (DGP)
Consider \(n\) observations are generated from the simple regression model,
$$ \begin{align} Y_i = \beta_{0} + \beta_{1} X_{i} + u_{i}, \end{align} $$where \(\beta_{0}=1\) is the intercept and \(\beta_{1}\) is the slope parameter.
Furthermore, suppose \(X_{i}\) and \(u_{i}\) are independently normally distributed, i.e.,
$$ \begin{align} X_{i} \sim N\left(0, \sigma_{X}^{2}\right), \;\;\;\;\; u_{i} \sim N\left(0, \sigma_{ui}^{2}\right), \end{align} $$where \(\sigma_{X}^{2}\) and \(\sigma_{ui}^{2}\) is the variance of \(X_{i}\) and \(u_{i}\).
Conditions for and Effects of heteroskedasticity
Heteroskedasticity occurs when \(\sigma_{ui}^{2}\) is not constant but a function of \(X_{i}\), e.g.,
$$ \begin{align} \sigma_{ui}^{2} = \text{exp}\left(\gamma_{0} + \gamma_{1} X_{i}\right), \end{align} $$which is also known as multiplicative heteroskedasticity where \(\gamma_{0}\) and \(\gamma_{1}\) are parameters which determine the shape of the heteroskedasticity.
Thus, for positive/negative values of \(\gamma_{1}\) the variance of \(u_{i}\) is positively/negatively related with the value of \(X_{i}\).
Heteroskedasticity effects the standard error of the OLS estimator \(\widehat{\beta}_{1}\) and thus the construction of the standardized OLS estimator \(z_{\widehat{\beta}_{1}}\).
For the construction of the standardized OLS estimator \(z_{\widehat{\beta}_{1}}\), the variance of \(\widehat{\beta}_{1}\), i.e., \(\sigma_{\widehat{\beta}_{1}}^{2}\), has to be estimated.
The variance of \(\widehat{\beta}_1\), i.e., \(\sigma_{\widehat{\beta}_{1}}^{2}\), can be robustly estimated by,
$$ \begin{align} \widehat{\sigma}_{\widehat{\beta}_{1}}^{2} = \frac{1}{n} \times \frac{\frac{1}{n-2}\sum_{i=1}^{n}\left(X_{i} - \overline{X}\right)^{2}\widehat{u}_{i}^{2}}{\left[\frac{1}{n}\sum_{i=1}^{n}\left(X_{i} - \overline{X}\right)^{2}\right]^{2}}, \end{align} $$where \(\widehat{u}_{i}\) are the residuals of the estimate regression line.
Note, the estimator for \(\sigma_{\widehat{\beta}_{1}}^{2}\) above is robust w.r.t. to heteroskedasticity, i.e., it does not rely on the assumption of homoskedasticity.
Instead, some statistic software report estimates \(\sigma_{\widehat{\beta}_{1}}^{2}\), based on the assumption of homoskedasticity.
The so called homoskedasticity-only estimator of \(\sigma_{\widehat{\beta}_{1}}^{2}\), is given by,
$$ \begin{align} \widetilde{\sigma}_{\widehat{\beta}_{1}}^{2} = \frac{\frac{1}{n-2}\sum_{i=1}^{n}\widehat{u}_{i}^{2}}{\sum_{i=1}^{n}\left(X_{i} - \overline{X}\right)^{2}}. \end{align} $$Illustration
Change the parameters and see the effect on the properties of the OLS estimator \(\widehat{\beta}_{1}\) as estimator for \(\beta_{1}\).
Parameters
Sample Size \(n\)
Variance of \(X_{i}\)
Variance of \(u_{i}\)
Heteroskedasticity
Scatter plot (realizations) |
The red fitted regression line is based on the regression of,
$$ \begin{align} Y_{i} \;\;\;\;\; \text{on} \;\;\;\;\; X_{i}. \end{align} $$The scatter plots and the fitted regression lines represent the result for only one simulation. The shaded areas illustrate the range of all fitted regression lines across all simulation outcomes.
Scatter plot (fitted residuals) |
The fitted unobserved residuals are constructed as,
$$ \begin{align} \widehat{u}_{i} = Y_{i} - \widehat{\beta}_{1} X_{i}, \end{align} $$for only one simulation where \(\widehat{\beta}_{1}\) is the respective OLS estimate.
Histogram of the OLS estimates \(\widehat{\beta}_{1}\) |
As the sample size \(N\) grows the OLS estimator \(\widehat{\beta}_{1}\) gets closer to \(\beta\), i.e.,
$$ \begin{align} \widehat{\beta}_{1} \overset{p}{\to} \beta. \end{align} $$Histogram of the standardized OLS estimates \(z_{\widehat{\beta}_{1}}\) |
In the case of heteroskedasticity,
$$ \begin{align} z_{\overline{Y}} &= \frac{\overline{Y} - \beta}{\widehat{\sigma}_{\widehat{\beta}_{1}}}, \end{align} $$i.e., based on robust standard errors \(\widehat{\sigma}_{\widehat{\beta}_{1}}\) gets closer to the standard normal distribution \(N\left(0, 1\right)\) (see green histogram).
Instead,
$$ \begin{align} z_{\overline{Y}} &= \frac{\overline{Y} - \beta}{\widetilde{\sigma}_{\widehat{\beta}_{1}}}, \end{align} $$i.e., based on ordinary standard errors \(\widetilde{\sigma}_{\widehat{\beta}_{1}}\) does not get closer to the standard normal distribution \(N\left(0, 1\right)\) (see red histogram).
More Details
|
This module is part of the DeLLFi project of the University of Hohenheim and funded by the ![]() |
|