Differences between revisions 1 and 7 (spanning 6 versions)
Revision 1 as of 2023-10-28 05:18:15
Size: 1390
Comment:
Revision 7 as of 2023-10-28 16:21:41
Size: 1381
Comment:
Deletions are marked like this. Additions are marked like this.
Line 1: Line 1:
= Linear Regression = = Ordinary Least Squares =
Line 3: Line 3:
A linear regression expresses the linear relation of a treatment variable to an outcome variable. '''Ordinary Least Squares''' ('''OLS''') is a linear regression method. It minimizes root mean square errors.
Line 11: Line 11:
== Regression Line ==

A regression line can be especially useful on a scatter plot.
== Univariate ==
Line 23: Line 21:
These points, with the generic equation for a line, can [[Econometrics/OrdinaryLeastSquares/UnivariateProof|prove]] that the slope of the regression line is equal to:

{{attachment:b12.svg}}

The generic formula for the regression line is:

{{attachment:b13.svg}}
Line 27: Line 33:
== Regression Computation == == Linear Model ==
Line 29: Line 35:
Take the generic equation form of a line: The linear model can be expressed as:
Line 31: Line 37:
{{attachment:b01.svg}} {{attachment:model1.svg}}
Line 33: Line 39:
Insert the first point into this form. If these assumptions can be made:
Line 35: Line 41:
{{attachment:b02.svg}}  1. Linearity
 2. Exogeneity
Line 37: Line 44:
This can be trivially rewritten to solve for ''a'' in terms of ''b'': {{attachment:model2.svg}}
Line 39: Line 46:
{{attachment:b03.svg}}  3.#3 Random sampling
 4. No perfect multicolinearity
 5. Heteroskedasticity
Line 41: Line 50:
Insert the second point into the original form. Then OLS is the best linear unbiased estimator ('''BLUE''') for these coefficients.
Line 43: Line 52:
{{attachment:b04.svg}} Using the computation above, the coefficients are estimated to produce:
Line 45: Line 54:
Now additionally insert the solution for ''a'' in terms of ''b''. {{attachment:model3.svg}}
Line 47: Line 56:
{{attachment:b05.svg}} The variance for each coefficient is estimated as:
Line 49: Line 58:
Expand all terms to produce: {{attachment:model4.svg}}
Line 51: Line 60:
{{attachment:b06.svg}} Where R^2^ is calculated as:
Line 53: Line 62:
This can now be eliminated into: {{attachment:model5.svg}}
Line 55: Line 64:
{{attachment:b07.svg}} Note also that the standard deviation of the population's parameter is unknown, so it's estimated like:
Line 57: Line 66:
Giving a solution for ''b'':

{{attachment:b08.svg}}

This solution is trivially rewritten as:

{{attachment:b09.svg}}

Expand the formula for correlation as:

{{attachment:b10.svg}}

This can now be eliminated into:

{{attachment:b11.svg}}

Finally, ''b'' can be eloquently written as:

{{attachment:b12.svg}}

Giving a generic formula for the regression line:

{{attachment:b13.svg}}
{{attachment:model6.svg}}

Ordinary Least Squares

Ordinary Least Squares (OLS) is a linear regression method. It minimizes root mean square errors.


Univariate

The regression line passes through two points:

[ATTACH]

and

[ATTACH]

These points, with the generic equation for a line, can prove that the slope of the regression line is equal to:

[ATTACH]

The generic formula for the regression line is:

[ATTACH]


Linear Model

The linear model can be expressed as:

model1.svg

If these assumptions can be made:

  1. Linearity
  2. Exogeneity

model2.svg

  1. Random sampling
  2. No perfect multicolinearity
  3. Heteroskedasticity

Then OLS is the best linear unbiased estimator (BLUE) for these coefficients.

Using the computation above, the coefficients are estimated to produce:

[ATTACH]

The variance for each coefficient is estimated as:

[ATTACH]

Where R2 is calculated as:

[ATTACH]

Note also that the standard deviation of the population's parameter is unknown, so it's estimated like:

[ATTACH]


CategoryRicottone

Statistics/OrdinaryLeastSquares (last edited 2025-01-10 14:33:38 by DominicRicottone)