This notion of OVB is confusing to me.
In order for OVB to be present, the regression equation must omit a variable that is correlated with a variable that is included in the regression equation and also, the dependent variable and omitted variable must be dependent, essentially.
So, I interpret these conditions to suggest that there must be an omitted variable for which both the regressor and regressand are dependent.
But the assumptions of MLR clearly state that the regressors are assumed to be independent, which would imply a correlation of zero among them, so I do not understand (I cannot reconcile) how it is possible to require i.i.d. regressors while also requiring that there not be any omitted variables that are correlated with the included regressors.
For example, assume a regression equation includes independent variables x, y, and z and that the dependent variable is w. Also, assume that variable v is correlated with variable x and that w is also dependent on v. Then if we do not include v, we have omitted variable bias yet if we include v, we have violated the i.i.d. multiple linear regression condition.
Further, page 201 of S&W states, in assumption #2, that the regressors are i.i.d. but on page 204, S&W states that the OLS estimators "... are correlated; this correlation arises from the correlation between the regressors."
Fundamentally, I am assuming i.i.d. random variables are uncorrelated since independence implies a covariance of zero, and in kind, a correlation of zero.
(I recall that a correlation of zero does not imply independence, but I believe independence does imply zer0 correlation.)
Anyone have any thoughts?
Thanks,
Brian
In order for OVB to be present, the regression equation must omit a variable that is correlated with a variable that is included in the regression equation and also, the dependent variable and omitted variable must be dependent, essentially.
So, I interpret these conditions to suggest that there must be an omitted variable for which both the regressor and regressand are dependent.
But the assumptions of MLR clearly state that the regressors are assumed to be independent, which would imply a correlation of zero among them, so I do not understand (I cannot reconcile) how it is possible to require i.i.d. regressors while also requiring that there not be any omitted variables that are correlated with the included regressors.
For example, assume a regression equation includes independent variables x, y, and z and that the dependent variable is w. Also, assume that variable v is correlated with variable x and that w is also dependent on v. Then if we do not include v, we have omitted variable bias yet if we include v, we have violated the i.i.d. multiple linear regression condition.
Further, page 201 of S&W states, in assumption #2, that the regressors are i.i.d. but on page 204, S&W states that the OLS estimators "... are correlated; this correlation arises from the correlation between the regressors."
Fundamentally, I am assuming i.i.d. random variables are uncorrelated since independence implies a covariance of zero, and in kind, a correlation of zero.
(I recall that a correlation of zero does not imply independence, but I believe independence does imply zer0 correlation.)
Anyone have any thoughts?
Thanks,
Brian
Last edited: