Differences between revisions 18 and 22 (spanning 4 versions)
Revision 18 as of 2024-06-05 21:29:24
Size: 1866
Comment: Rewrite
Revision 22 as of 2024-06-07 14:49:00
Size: 2164
Comment: Rewrite of coefficients section
Deletions are marked like this. Additions are marked like this.
Line 23: Line 23:
The proof can be seen [[Econometrics/OrdinaryLeastSquares/UnivariateProof|here]]. The derivation can be seen [[Econometrics/OrdinaryLeastSquares/Univariate|here]].
Line 31: Line 31:
Given ''k'' independent variables, the OLS model is specified as:

{{attachment:mmodel.svg}}

It is estimated as:

{{attachment:mestimate.svg}}

More conventionally, this is estimated with [[LinearAlgebra|linear algebra]] as:

{{attachment:matrix.svg}}

The derivation can be seen [[Econometrics/OrdinaryLeastSquares/Multivariate|here]].
Line 35: Line 49:
== Linear Model == == Estimated Coefficients ==
Line 37: Line 51:
The linear model can be expressed as:

{{attachment:model1.svg}}

If these assumptions can be made:
The '''Gauss-Markov theorem''' demonstrates that (with some assumptions) the OLS estimations are the '''best linear unbiased estimators''' ('''BLUE''') for the regression coefficients. The assumptions are:
Line 46: Line 56:
 4. No perfect multicolinearity  4. No perfect [[LinearAlgebra/Basis|multicolinearity]]
Line 48: Line 58:

Then OLS is the best linear unbiased estimator ('''BLUE''') for these coefficients.

Using the computation above, the coefficients are estimated to produce:

{{attachment:model2.svg}}

Ordinary Least Squares

Ordinary Least Squares (OLS) is a linear regression method. It minimizes root mean square errors.


Univariate

Given one independent variable and one dependent (outcome) variable, the OLS model is specified as:

model.svg

It is estimated as:

estimate.svg

This model describes (1) the mean observation and (2) the marginal changes to the outcome per unit changes in the independent variable.

The derivation can be seen here.


Multivariate

Given k independent variables, the OLS model is specified as:

mmodel.svg

It is estimated as:

mestimate.svg

More conventionally, this is estimated with linear algebra as:

matrix.svg

The derivation can be seen here.


Estimated Coefficients

The Gauss-Markov theorem demonstrates that (with some assumptions) the OLS estimations are the best linear unbiased estimators (BLUE) for the regression coefficients. The assumptions are:

  1. Linearity
  2. Exogeneity

  3. Random sampling
  4. No perfect multicolinearity

  5. Homoskedasticity

The variances for each coefficient are:

[ATTACH]

Note that the standard deviation of the population's parameter is unknown, so it's estimated like:

[ATTACH]

If the homoskedasticity assumption does not hold, then the estimators for each coefficient are actually:

[ATTACH]

Wherein, for example, r1j is the residual from regressing x1 onto x2, ... xk.

The variances for each coefficient can be estimated with the Eicker-White formula:

[ATTACH]

See Nicolai Kuminoff's video lectures for the derivation of the robust estimators.


CategoryRicottone

Statistics/OrdinaryLeastSquares (last edited 2025-09-03 02:08:40 by DominicRicottone)