Differences between revisions 6 and 23 (spanning 17 versions)
Revision 6 as of 2023-10-28 07:04:18
Size: 2049
Comment:
Revision 23 as of 2025-01-10 14:33:11
Size: 2224
Comment: Killing Econometrics page
Deletions are marked like this. Additions are marked like this.
Line 1: Line 1:
## page was renamed from Econometrics/OrdinaryLeastSquares
Line 13: Line 14:
The regression line passes through two points: Given one independent variable and one dependent (outcome) variable, the OLS model is specified as:
Line 15: Line 16:
{{attachment:regression1.svg}} {{attachment:model.svg}}
Line 17: Line 18:
and It is estimated as:
Line 19: Line 20:
{{attachment:regression2.svg}} {{attachment:estimate.svg}}
Line 21: Line 22:
Take the generic equation form of a line: This model describes (1) the mean observation and (2) the marginal changes to the outcome per unit changes in the independent variable.
Line 23: Line 24:
{{attachment:b01.svg}}

Insert the first point into this form.

{{attachment:b02.svg}}

This can be trivially rewritten to solve for ''a'' in terms of ''b'':

{{attachment:b03.svg}}

Insert the second point into the original form.

{{attachment:b04.svg}}

Now additionally insert the solution for ''a'' in terms of ''b''.

{{attachment:b05.svg}}

Expand all terms to produce:

{{attachment:b06.svg}}

This can now be eliminated into:

{{attachment:b07.svg}}

Giving a solution for ''b'':

{{attachment:b08.svg}}

This solution is trivially rewritten as:

{{attachment:b09.svg}}

Expand the formula for correlation as:

{{attachment:b10.svg}}

This can now be eliminated into:

{{attachment:b11.svg}}

Finally, ''b'' can be eloquently written as:

{{attachment:b12.svg}}

Giving a generic formula for the regression line:

{{attachment:b13.svg}}
The derivation can be seen [[Econometrics/OrdinaryLeastSquares/Univariate|here]].
Line 77: Line 30:
== Linear Model == == Multivariate ==
Line 79: Line 32:
The linear model can be expressed as: Given ''k'' independent variables, the OLS model is specified as:
Line 81: Line 34:
{{attachment:model1.svg}} {{attachment:mmodel.svg}}
Line 83: Line 36:
If these assumptions can be made: It is estimated as:

{{attachment:mestimate.svg}}

More conventionally, this is estimated with [[LinearAlgebra|linear algebra]] as:

{{attachment:matrix.svg}}

The derivation can be seen [[Econometrics/OrdinaryLeastSquares/Multivariate|here]].

----



== Estimated Coefficients ==

The '''Gauss-Markov theorem''' demonstrates that (with some assumptions) the OLS estimations are the '''best linear unbiased estimators''' ('''BLUE''') for the regression coefficients. The assumptions are:
Line 86: Line 55:
 2. Exogeneity  2. [[Econometrics/Exogeneity|Exogeneity]]
 3. Random sampling
 4. No perfect [[LinearAlgebra/Basis|multicolinearity]]
 5. [[Econometrics/Homoskedasticity|Homoskedasticity]]
Line 88: Line 60:
{{attachment:model2.svg}} The variances for each coefficient are:
Line 90: Line 62:
 3.#3 Random sampling
 4. No perfect multicolinearity
 5. Heteroskedasticity
{{attachment:homo1.svg}}
Line 94: Line 64:
Then OLS is the best linear unbiased estimator ('''BLUE''') for these coefficients. Note that the standard deviation of the population's parameter is unknown, so it's estimated like:
Line 96: Line 66:
Using the computation above, the coefficients are estimated to produce: {{attachment:homo2.svg}}
Line 98: Line 68:
{{attachment:model3.svg}} If the homoskedasticity assumption does not hold, then the estimators for each coefficient are actually:
Line 100: Line 70:
The variance for each coefficient is estimated as: {{attachment:hetero1.svg}}
Line 102: Line 72:
{{attachment:model4.svg}} Wherein, for example, ''r,,1j,,'' is the residual from regressing ''x,,1,,'' onto ''x,,2,,'', ... ''x,,k,,''.
Line 104: Line 74:
Where R^2^ is calculated as: The variances for each coefficient can be estimated with the Eicker-White formula:
Line 106: Line 76:
{{attachment:model5.svg}} {{attachment:hetero2.svg}}
Line 108: Line 78:
Note also that the standard deviation of the population's parameter is unknown, so it's estimated like:

{{attachment:model6.svg}}
See [[https://www.youtube.com/@kuminoff|Nicolai Kuminoff's]] video lectures for the derivation of the robust estimators.

Ordinary Least Squares

Ordinary Least Squares (OLS) is a linear regression method. It minimizes root mean square errors.


Univariate

Given one independent variable and one dependent (outcome) variable, the OLS model is specified as:

model.svg

It is estimated as:

estimate.svg

This model describes (1) the mean observation and (2) the marginal changes to the outcome per unit changes in the independent variable.

The derivation can be seen here.


Multivariate

Given k independent variables, the OLS model is specified as:

mmodel.svg

It is estimated as:

mestimate.svg

More conventionally, this is estimated with linear algebra as:

matrix.svg

The derivation can be seen here.


Estimated Coefficients

The Gauss-Markov theorem demonstrates that (with some assumptions) the OLS estimations are the best linear unbiased estimators (BLUE) for the regression coefficients. The assumptions are:

  1. Linearity
  2. Exogeneity

  3. Random sampling
  4. No perfect multicolinearity

  5. Homoskedasticity

The variances for each coefficient are:

[ATTACH]

Note that the standard deviation of the population's parameter is unknown, so it's estimated like:

[ATTACH]

If the homoskedasticity assumption does not hold, then the estimators for each coefficient are actually:

[ATTACH]

Wherein, for example, r1j is the residual from regressing x1 onto x2, ... xk.

The variances for each coefficient can be estimated with the Eicker-White formula:

[ATTACH]

See Nicolai Kuminoff's video lectures for the derivation of the robust estimators.


CategoryRicottone

Statistics/OrdinaryLeastSquares (last edited 2025-09-03 02:08:40 by DominicRicottone)