Differences between revisions 8 and 9
Revision 8 as of 2020-10-22 20:20:17
Size: 2330
Comment:
Revision 9 as of 2020-11-05 18:20:36
Size: 3297
Comment:
Deletions are marked like this. Additions are marked like this.
Line 26: Line 26:
The base weight is the inverse of the probability of being sampled. Think ''desired over actual''. As such, the sum of base weights should equal the population size. The base weight is the inverse of the probability of being sampled. ''Think '''desired over actual'''.''
Line 28: Line 28:
For a SRS design, this is calculated as a simple rate. Given a population of 20,000 and a sample size of 667, the propbability of being sampled is 20,000/667 = '''29.99'''.  * For a '''census''', all respondents have a weight of 1.
 * For a '''
SRS design''', this is calculated as a simple rate. Given a population of 20,000 and a sample size of 667, the propbability of being sampled is 20,000/667 = '''29.99'''.
 * For a '''STSRS design''', the same process is applied per stratum.
Line 30: Line 32:
For a STSRS design, the same process is applied per stratum. Note that, in each, the sum of base weights should equal the population size.
Line 38: Line 40:
Survey weights can adjust for non-response bias. The core concept is to use auxiliary frame data (i.e. descriptives known for ''both'' respondents and non-respondents) that is correlated with key measures or responsivity. There are two main methods for adjusting weights:
Line 40: Line 42:
'''Weighting class adjustments''' divides the sample into weighting classes and applies a class-specific adjustment factor to every case.  1. '''Weighting class adjustments''' involve dividing the sample into discrete classes and applying an adjustment factor by class.
 2. '''Propensity score adjustments''' involve calculating the inverse of the estimated probability to respond and applying that as a secondary weight.
Line 42: Line 45:
'''Propensity score adjustments''' calculates the inverse of the estimated probability to respond and applies that as a secondary weight.

=== Reapportioning Weight ===

The collected measures should reflect the sample (and therefore the population), but incomplete data complicates this. It is common to break the sample into weighting classes based on responsivity, and then reapportion the weight of non-respondents to respondents.

Consider a simple design without eligibility.

||'''Class''' ||'''Count'''||
||Respondent ||800 ||
||Non-respondent||200 ||

To re-apportion the weight of non-respondents, the respondents' weight factors would be adjusted by a factor of (800+200)/800 or 1.25. The non-respondents would then be dropped, or assigned weight factors of 0. ''This is, again, a calculation of '''desired over actual'''.''



=== Non-response Bias ===

Responsivity is commonly related to the key measures of a survey, and therefore introduces non-response bias. Weighting can account for this error. The core concept is to use auxiliary frame data (i.e. descriptives known for both respondents ''and'' non-respondents).

Survey Weights

Survey weights account for the design of a survey sample and other biases/errors introduced by a survey instrument.


The Basic Process

  1. Set survey dispositions
  2. Calculate base weights
  3. Apply non-response adjustments to base weights
  4. Calibrate the weights

See here for details about survey dispositions.


Calculating Weights

The base weight is the inverse of the probability of being sampled. Think desired over actual.

  • For a census, all respondents have a weight of 1.

  • For a SRS design, this is calculated as a simple rate. Given a population of 20,000 and a sample size of 667, the propbability of being sampled is 20,000/667 = 29.99.

  • For a STSRS design, the same process is applied per stratum.

Note that, in each, the sum of base weights should equal the population size.


Non-Response Adjustments

There are two main methods for adjusting weights:

  1. Weighting class adjustments involve dividing the sample into discrete classes and applying an adjustment factor by class.

  2. Propensity score adjustments involve calculating the inverse of the estimated probability to respond and applying that as a secondary weight.

Reapportioning Weight

The collected measures should reflect the sample (and therefore the population), but incomplete data complicates this. It is common to break the sample into weighting classes based on responsivity, and then reapportion the weight of non-respondents to respondents.

Consider a simple design without eligibility.

Class

Count

Respondent

800

Non-respondent

200

To re-apportion the weight of non-respondents, the respondents' weight factors would be adjusted by a factor of (800+200)/800 or 1.25. The non-respondents would then be dropped, or assigned weight factors of 0. This is, again, a calculation of desired over actual.

Non-response Bias

Responsivity is commonly related to the key measures of a survey, and therefore introduces non-response bias. Weighting can account for this error. The core concept is to use auxiliary frame data (i.e. descriptives known for both respondents and non-respondents).

Adjustments are applied in phases. Cases with unknown eligibility often cannot be adjusted through these methods, and need to be removed. Ineligible cases often are undesirable in analysis datasets, so weights are further adjusted to account for their removal.


Calibration Adjustments

Survey weights can be adjusted to ensure that known population descriptives are reflected in the estimates.

Methods include:

  • post-stratification (i.e. desired over actual)

  • raking
  • linear calibration (GREG)

Raking

Raking, or RIM weighting, involves applying post-stratification by each dimension iteratively, until the weights converge. Convergence is defined as the root mean square (RMS) falling below a threshold, typically 0.000005.

Raked weights generally should not be applied if their efficiency falls below 70%.


CategoryRicottone

SurveyWeights (last edited 2025-01-10 16:10:44 by DominicRicottone)