Analysis of Survey Data phần 6 pps

38 270 0
Analysis of Survey Data phần 6 pps

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

Thông tin tài liệu

The sample distribution is different also from the familiar px distribution, defined as the combined distribution over all possible realizations of the finite population measurements (the population x distribution) and all possible sample values for a given population (the randomization p distribution). The px distribution is often used for comparing the performance of design-based estimators in situations where direct comparisons of randomization variances or mean square errors are not feasible. The obvious difference between the sample distribution and the px distribution is that the former conditions on the selected sample (and values of auxiliary variables measured for units in the sample), whereas the latter accounts for all possible sample selections. Finally, rather than conditioning on the selected sample when constructing the sample distribution (and hence the sample likelihood), one could compute instead the joint distribution of the selected sample and the corresponding sample measurements. Denote by y s  {y t , t P s} the outcome variable values measured for the sample units and by x s  {x t , t P s} and x ~ s  {x t , t TP s} the values of the auxiliary variables corresponding to the sampled and nonsampled units. Assuming independence of the population measurements and independ- ent sampling of the population units (Poisson sampling), the joint pdf of (s, y s )j(x s , x ~ s ) can be written as f (s, y s jx s , x ~ s )   tPs [p( y t , x t )f U ( y t jx t )ap(x t )]  tPs p(x t )  tTPs [1 À p(x t )] (12X6) where p( y t , x t )  E U (p t jy t , x t ) and p(x t )  E U (p t jx t ). Note that the product of the terms in the first set of square brackets on the right hand side of (12.6) is the joint sample pdf, f s ( y s jx s , s), for units in the sample as obtained from (12.3). The use of (12.6) for likelihood-based inference has the theoretical advantage of employing the information on the sample selection probabilities for units outside the sample, but it requires knowledge of the expectations p(x t )  E U (p t jx t ) for all t P U and hence the values x ~ s . This information is not needed when inference is based on the sample pdf, f s ( y s jx s , s). When the values x ~ s are unknown, it is possible in theory to regard the values {x t , t TP s} as random realizations from some pdf g ~ s (x t ) and replace the expectations p(x t ) for units t TP s by the uncondi- tional expectations p(t)   p(x t )g ~ s (x t )dx t . See, for example, Rotnitzky and Robins (1997) for a similar analysis in a different context. However, modeling the distribution of the auxiliary variables might be formidable and the resulting likelihood f (s, y s jx s , x ~ s ) could be very cumbersome. 12.3. INFERENCE UNDER INFORMATIVE PROBABILITY SAMPLING inference underinformative probability sampling 12.3.1. Estimating equations with application to the GLM In this section we consider four different approaches for defining estimating equations under informative probability sampling. We compare the various approaches empirically in Section 12.5. INFERENCE UNDER INFORMATIVE PROBABILITY SAMPLING 179 Suppose that the population measurements ( y U , x U )  {( y t , x t ), t  1 F F F N} can be regarded as N independent realizations from some pdf f y, x . Denote by f U ( yjx; b) the conditional pdf of y t given x t . The true value of the vector parameter b  ( b 0 , b 1 , F F F , b k ) 0 is defined as the unique solution of the equa- tions W U ( b)   N t1 E U [d Ut jx t ]  0 (12X7) where d Ut  (d Ut, 0 d Ut, 1 F F F d Ut, k ) 0  ] log f U ( y t jx t ; b)a]b is the tth score func- tion. We refer to (12.7) as the `parameter equations' since they define the vector parameter b. For the GLM defined in the introduction with the distribution of y t belonging to the exponential family, f U ( y; y, f)  exp {[ yy À b(y)]a a(f)  c( y, f)} where a(X) b 0, b(X) and c(.) are known functions, and f is known. It follows that m(y)  E( y)  ]b(y)a]y so that if y  h(xb) for some function h(.) with derivative g(.), d Ut, j  {y t À m(h(x 0 t b))}[g(x 0 t b)]x t,j X (12X8) The `census parameter' (Binder, 1983) corresponding to (12.7) is defined as the solution B U of the equations W Ã U ( b)   N t1 d Ut  0X (12X9) Note that (12.9) defines the maximum likelihood estimating equations based on all the population values. Let s  {1 F F F n} denote the sample, assumed to be selected by some probabil- ity sampling scheme with first-order selection probabilities p t  Pr(t P s). (The sample size n can be random.) The first approach that we consider for specifying the estimating equations involves redefining the parameter equations with respect to the sample distribution f s ( y t jx t ) (Equation (12.3)), rather than the population distribution as in (12.7). Assuming that the form of the conditional expectations E U (p t jy t , x t ) is known (see Section 12.3.2) and that the expect- ations E U (p t jx t )   y E U (p t jy, x t )f U ( yjx t ; b)dy are differentiable with respect to b, the parameter equations corresponding to the sample distribution are W 1s ( b)   s E s {[] log f s ( y t jx t ; b)a]b]jx t }   s E s {[d Ut  ] log E s (w t jx t )a]b]jx t }  0X (12X10) (The second equation follows from (12.3) and (12.5).) The parameters b are estimated under this approach by solving the equations W 1s,e ( b)   s [d Ut  ] log E s (w t jx t )a]b]  0X (12X11) Note that (12.11) defines the sample likelihood equations. 180 FITTING GLMS UNDER INFORMATIVE SAMPLING The second approach uses the relationship (12.4) in order to convert the population expectations in (12.7) into sample expectations. Assuming a random sample of size n from the sample distribution, the parameter equations then have the form W 2s ( b)   s E s (q t d Ut jx t )  0 (12X12) where q t  w t aE s (w t jx t ). The vector b is estimated under this approach by solving the equations W 2s, e ( b)   s q t d Ut  0X (12X13) The third approach is based on the property that if b solves Equations (12.7), then it solves also the equations ~ W U ( b)   N t1 E U (d Ut )  E x  N t1 E U (d Ut jx t ) 4 5  0 where the expectation E x is over the population distribution of the x t . Applica- tion of (12.4) to each of the terms E U (d Ut ) (without conditioning on x t ) yields the following parameter equations for a random sample of size n from the sample distribution: W 3s ( b)   s E s (w t d Ut )aE s (w t )  0X (12X14) The corresponding estimating equations are W 3s,e ( b)   s w t d Ut  0X (12X15) An interesting feature of the equations in (12.15) is that they coincide with the pseudo-likelihood equations as obtained when estimating the census equations (12.9) by the Horvitz±Thompson estimators. (For the concept and uses of pseudo-likelihood see Binder, 1983; SHS; Godambe and Thompson, 1986; Pfeffermann, 1993; and Chapter 2.) Comparing (12.13) with (12.15) shows that the former equations use the adjusted weights, q t  w t aE s (w t jx t ) instead of the standard weights w t used in (12.15). As discussed in PS, the weights q t account for the net sampling effects on the target conditional distribution of y t jx t , whereas the weights w t account also for the sampling effects on the marginal distribution of x t . In particular, when w is a deterministic function of x so that the sampling process is noninformative, q t  1 and Equations (12.13) reduce to the ordinary likelihood equations (see (12.16) below). The use of (12.15) on the other hand may yield highly variable estimators in such cases, depending on the variability of the x t . The three separate sets of estimating equations defined by (12.11), (12.13), and (12.15) all account for the sampling effects. On the other hand, ignoring the INFERENCE UNDER INFORMATIVE PROBABILITY SAMPLING 181 sampling process results in the use of the ordinary (face value) likelihood equations W 4s,e ( b)   s d Ut  0X (12X16) We consider the solution to (12.16) as a benchmark for the assessment of the performance of the other estimators. Comment The estimating equations proposed in this section employ the scores d Ut  ] log f U ( y t jx t ; b)a]b. However, similar equations can be obtained for other functions d Ut ; see Bickel et al. (1993) for examples of alternative definitions. 12.3.2. Estimation of E s (w t jx t ) The estimating equations defined by (12.11) and (12.13) contain the expect- ations E s (w t jx t ) that depend on the unknown parameters b. When the w t are continuous as in probability proportional to size (PPS) sampling with a continuous size variable, the form of these expectations can be identified from the sample data by the following three-step procedure that utilizes (12.5): 1. Regress w t against (y t , x t ) to obtain an estimate of E s (w t jy t , x t ). 2. Integrate  y E U (p t jy, x t )f U ( yjx t ; b)dy   y [1aE s (w t jy, x t )]f U ( yjx t ; b)dy to obtain an estimate of E U (p t jx t ) as a function of b. 3. Compute E s (w t jx t )  1aE U (p t jx t ). (The computations in steps 2 and 3 use the estimates obtained in the previous step.) The articles by PKR and PS contain several plausible models for E U (p t jy t , x t ) and examples for which the integral in step 2 can be carried out analytically. In practice, however, the specific form of the expectation E U (p t jy t , x t ) will usually be unknown but the expectation E s (w t jy t , x t ) can be identified and estimated in this case from the sample. (The expectation depends on unknown parameters that are estimated in step 1.) Comment 1 Rather than estimating the coefficients indexing the expectation E s (w t jy t , x t ) from the sample (step 1), these coefficients can be considered as additional unknown parameters, with the estimating equations extended ac- cordingly. This, however, may complicate the solution of the estimating equa- tions and also result in identifiability problems under certain models. See PKR for examples and discussion. Comment 2 For the estimating equations (12.13) that use the weights q t  w t aE s (w t jx t ), the estimation of the expectation E s (w t jx t ) can be carried out by simply regressing w t against x t , thus avoiding steps 2 and 3. This is so because in this case there is no need to express the expectation as a function of the parameters b indexing the population distribution. 182 FITTING GLMS UNDER INFORMATIVE SAMPLING The discussion so far focuses on the case where the sample selection prob- abilities are continuous. The evaluation of the expectation E s (w t jx t ) in the case of discrete selection probabilities is simpler. For example, in the empirical study of this chapter we consider the case of logistic regression with a discrete independent variable x and three possible values for the dependent variable y. For this case the expectation E s (w t jx t  k) is estimated as E s (w t jx t  k) 1aE U (p t jx t  k), E U (p t jx t  k)   3 a1 Pr U ( y t  ajx t  k)E U (p t jy t  a, x t  k)   3 a1 Pr U ( y t  ajx t  k)aE s (w t jy t  a, x t  k) ^  3 a1 Pr U ( y t  ajx t  k)aw ak (12X17) where w ak  [  s w t I( y t  a, x t  k)]a[  s I( y t  a, x t  k)]. Here I(A) is the indicator function for the event A. Substituting the logistic function for Pr( y t  ajx t  k) in the last expression of (12.17) yields the required specifica- tion. The estimators w ak are considered as fixed numbers when solving the estimating equations. For the estimating equations (12.13), the expectations E s (w t jx t  k) in (12.13) in this example are estimated by E s (w t jx t  k) ^w k   s w t I(x t  k) 4 5  s I(x t  k) 4 5D (12X18) rather than using (12.17) that depends on the unknown logistic coefficients (see Comment 2 above). For an example of the evaluation of the expectation E s (w t jx t ) with discrete selection probabilities but continuous outcome and explanatory variables, see PS (section 5.2). 12.3.3. Testing the informativeness of the sampling process The estimating equations developed in Section 12.3.1 for the case of informa- tive sampling involve the use of the sampling weights in various degrees of complexity. It is clear therefore that when the sampling process is in fact noninformative, the use of these equations yields more variable estimators than the use of the ordinary score function defined by (12.16). See Tables 12.2 and 12.4 below for illustrations. For the complex sampling schemes in common use, the sample selection probabilities are often determined by the values of several design variables, in which case the informativeness of the selection process is not always apparent. This raises the need for test procedures as a further indication of whether the sampling process is ignorable or not. Several tests have been proposed in the past for this problem. The common INFERENCE UNDER INFORMATIVE PROBABILITY SAMPLING 183 feature of these tests is that they compare the probability-weighted estimators of the target parameters to the ordinary (unweighted) estimators that ignore the sampling process, see Pfeffermann (1993) for review and discussion. For the classical linear regression model, PS propose a set of test statistics that compare the moments of the sample distribution of the regression residuals to the corresponding moments of the population distribution. The use of these tests is equivalent to testing that the correlations under the sample distribution between powers of the regression residuals and the sampling weights are all zero. In Chapter 11 the tests developed by PS are extended to situations where the moments of the model residuals are functions of the regressor variables, as under many of the GLMs in common use. A drawback of these test procedures is that they involve the use of a series of tests with dependent test statistics, such that the interpretation of the results of these tests is not always clear-cut. For this reason, we propose below a single alternative test that compares the estimating equations that ignore the sampling process to estimating equations that account for it. As mentioned before, the question arising in practice is whether to use the estimating equations (12.16) that ignore the sample selection or one of the estimating equations (12.11), (12.13), or (12.15) that account for it, so that basing the test on these equations is very natural. In what follows we restrict attention to the comparison of the estimating equations (12.13) and (12.16) (see Comment below). From a theoretical point of view, the sampling process can be ignored for inference if the corresponding parameter equations are equivalent or  s E s (d Ut jx t )   s E s (q t d Ut jx t ). Denot- ing R(x t )  E s (d Ut jx t ) À E s (q t d Ut jx t ), the null hypothesis is therefore H 0 X R n  n À1  s R(x t )  0X (12X19) Note that dim(R n )  k  1  dim( b). If b were known, the hypothesis could be tested by use of the Hotelling test statistic, H(R)  n À (k  1) k  1 ^ R 0 n S À1 n ^ R n $ H 0 F k1, nÀ(k1) (12X20) where ^ R n  n À1  s ^ R(x t ); ^ R(x t )  (d Ut À q t d Ut ) and S n  n À1  s ( ^ R(x t ) À ^ R n )( ^ R(x t ) À ^ R n ) 0 X In practice, b is unknown and the score d Ut in ^ R(x t ) has to be evaluated at a sample estimate of b. In principle, any of the estimates defined in Section 12.3.1 could be used for this purpose since under H 0 all the estimators are consistent for b, but we find that the use of the solution of (12.16) that ignores the sampling process is the simplest and yields the best results. 184 FITTING GLMS UNDER INFORMATIVE SAMPLING Let ^ d Ut define the value of d Ut evaluated at ^ b ± the solution of (12.16) ± and let ~ R(x t ), ~ R n , and ~ S n be the corresponding values of ^ R(x t ), ^ R n , and S n obtained after substituting ^ b for b in (12.20). The test statistic is therefore ~ H(R)  n À (k  1) k  1 ~ R 0 n ~ S À1 n ~ R n % H 0 F k1, nÀ(k1) X (12X21) Note that  s ^ d Ut  0 by virtue of (12.16), so ~ R n  Àn À1  s q t ^ d Ut . The random variables q t ^ d Ut are no longer independent since  s ^ d Ut  0, but utiliz- ing the property that E s (q t jx t )  1 implies that under the null hypothesis var s [  s q t ^ d Ut ]  var s [  s (q t ^ d Ut À ^ d Ut )]   s var s ( ^ d Ut À q t ^ d Ut ), thus justifying the use of ~ S n a(n À 1) as an estimator of var( ~ R n ) in the construction of the test statistic in (12.21). Comment The Hotelling test statistic uses the estimating equations (12.13) for the comparison with (12.16) and here again, one could use instead the equa- tions defined by (12.11) or (12.15): that is, replace q t d Ut in the definition of ^ R(x t ) by d Ut  ] log [E s (w t jx t )]a]b, or by w t d Ut respectively. The use of (12.11) is more complicated since it requires evaluation of the expectation E s (w t jx t ) as a function of b (see Section 12.3.2). The use of (12.15) is the simplest but it yields inferior results to the use of (12.13) in our simulation study. 12.4. VARIANCE ESTIMATION variance estimation Having estimated the model parameters by any of the solutions of the estimat- ing equations in Section 12.3.1, the question arising is how to estimate the variances of these estimators. Unless stated otherwise, the true (estimated) variances are with respect to the sample distribution for a given sample of units, that is, the variance under the pdf obtained by the product of the sample pdfs (12.3). Note also that since the estimating equations are only for the b-parameters, with the coefficients indexing the expectations E s (w t jy t , x t ) held fixed at their estimators of these values, the first four variance estimators below do not account for the variability of the estimated coefficients. For the estimator ^ b 1s defined by the solution to the estimating equations (12.11), that is, the maximum likelihood estimator under the sample distribu- tion, a variance estimator can be obtained from the inverse of the information matrix evaluated at this estimator. Thus, ^ V( ^ b 1s )  { ÀE s []W 1s,e ( b)a]b 0 ] b ^ b 1s } À1 X (12X22) For the estimators ^ b 2s solving (12.13), we use a result from Bickel et al. (1993). By this result, if for the true vector parameter b 0 , the left hand side of an estimating equation W n ( b)  0 can be approximated as W n ( b 0 )  n À1  s j( y t , x t ; b 0 )  O p (n À1a2 ) for some function j satisfying E(j)  0 and E(j 2 ) ` I, then under some additional regularity conditions on the order of convergence of certain functions, VARIANCE ESTIMATION 185 n 1a2 ( ^ b n À b 0 )  n À1a2  s [  W( b 0 )] À1 j( y t , x t ; b 0 )  o p (1) (12X23) where ^ b n is the solution of W n ( b)  0,  W( b 0 )  []W( b)a]b 0 ] bb 0 and W( b)  0 is the parameter equation with  W( b 0 ) assumed to be nonsingular. For the estimating equations (12.13), j( y t , x t ; b)  q t d Ut , implying that the variance of ^ b 2s solving (12.13) can be estimated as ^ V s ( ^ b 2s )  [  W 2s,e ( ^ b 2s )] À1  s [q t d Ut ( ^ b 2s )] 2 @ A [  W 2s,e ( ^ b 2s )] À1 (12X24) where  W 2s,e ( ^ b 2s )  []W 2s,e ( b)ab 0 ] b ^ b 2s and d Ut ( ^ b 2s ) is the value of d Ut evaluated at ^ b 2s . Note that since E s (q t d Ut jx t )  0 (and also  s q t d Ut ( ^ b 2s )  0), the estimator (12.24) estimates the conditional variance V s ( ^ b 2s j{x t , t P s}); that is, the variance with respect to the conditional sample distribution of the outcome y. The estimating equations (12.15) have been derived in Section 12.3.1 by two different approaches, implying therefore two separate variance estimators. Under the first approach, these equations estimate the parameter equations (12.14), which are defined in terms of the unconditional sample expectation (the expectation under the sample distribution of {( y t , x t ), t P s}). Application of the result from Bickel et al. (1993) mentioned before yields the following variance estimator (compare with (12.24)): ^ V s ( ^ b 3s )  [  W 3s,e ( ^ b 3s )] À1  s [w t d Ut ( ^ b 3s )] 2 @ A [  W 3s,e ( ^ b 3s )] À1 (12X25) where  W 3s,e ( ^ b 3s )  []W 3s, e ( b)a]b 0 ] b ^ b 3s . For this case E s (w t d Ut )  0 (and also  s w t d Ut ( ^ b 3s )  0) so that (12.25) estimates the unconditional variance over the joint sample distribution of {( y t , x t ), t P s}. Under the second approach, the estimating equations (12.15) are the ran- domization unbiased estimators of the census equations (12.9). As such, the variance of ^ b 3s can be evaluated with respect to the randomization distribution over all possible sample selections, with the population values held fixed. Following Binder (1983), the randomization variance is estimated as ^ V R ( ^ b 3s )  [ ^  W Ã U ( ^ b 3s )] À1 ^ V R  s w t d Ut ( ^ b 3s ) 4 5 [ ^  W Ã U ( ^ b 3s )] À1 (12X26) where ^  W Ã U ( ^ b 3s ) is design (randomization) consistent for []W Ã U a]b 0 ] bb 0 and ^ V R [  s w t d Ut ( ^ b 3s )] is an estimator of the randomization variance of  s w t d Ut ( b), evaluated at ^ b 3s . In order to illustrate the difference between the variance estimators (12.25) and (12.26), consider the case where the sample is drawn by Poisson sampling such that units are selected into the sample independently by Bernoulli trials 186 FITTING GLMS UNDER INFORMATIVE SAMPLING with probabilities of success p t  Pr(t P s). Simple calculations imply that for this case the randomization variance estimator (12.26) has the form ^ V R ( ^ b 3s )  [ ^  W Ã U ( ^ b 3s )] À1  s (1 À p t )[w t d Ut ( ^ b 3s )] 2 @ A [ ^  W Ã U ( ^ b 3s )] À1 (12X27) where ^  W Ã U ( ^ b 3s )   W 3s, e ( ^ b 3s ). Thus, the difference between the estimator de- fined by (12.25) and the randomization variance estimator (12.26) is in this case in the weighting of the products w t d Ut ( ^ b 3s ) by the weights (1 Àp t ) in the latter estimator. Since 0 ` (1 À p t ) ` 1, the randomization variance estimators are smaller than the variance estimators obtained under the sample distribution. This is expected since the randomization variances measure the variation around the (fixed) population values and if some of the selection probabilities are large, a correspondingly large portion of the population is included in the sample (in high probability), thus reducing the variance. Another plausible variance estimation procedure is the use of bootstrap samples. As mentioned before, under general conditions on the sample selec- tion scheme listed in PKR, the sample measurements are asymptotically inde- pendent with respect to the sample distribution, implying that the use of the (classical) bootstrap method for variance estimation is well founded. In con- trast, the use of the bootstrap method for variance estimation under the randomization distribution is limited, and often requires extra modifications; see Sitter (1992) for an overview of bootstrap methods for sample surveys. Let ^ b s stand for any of the preceding estimators and denote by ^ b b s the estimator computed from bootstrap sample b (b  1 F F F B), drawn by simple random sampling with replacement from the original sample (with the same sample size). The bootstrap variance estimator of ^ b s is defined as ^ V boot ( ^ b s )  B À1  B b1 ( ^ b b s À b boot )( ^ b b s À b boot ) 0 (12X28) where b boot  B À1  B b1 ^ b b s X It follows from the construction of the bootstrap samples that the estimator (12.28) estimates the unconditional variance of ^ b s . A possible advantage of the use of the bootstrap variance estimator in the present context is that it accounts in principle for all the sources of variation. This includes the identification of the form of the expectations E s (w t jy t , x t ) when unknown, and the estimation of the vector coefficient l indexing that expectation, which is carried out for each of the bootstrap samples but not accounted for by the other variance estimation methods unless the coefficients l are considered as part of the unknown model parameters (see Section 12.3.2). VARIANCE ESTIMATION 187 12.5. SIMULATION RESULTS simulation results 12.5.1. Generation of population and sample selection In order to assess and compare the performance of the parameter estimators, variance estimators, and the test statistic proposed in Sections 12.3 and 12.4, we designed a Monte Carlo study that consists of the following stages: A. Generate a univariate population of x-values of size N  3000, drawn independently from the discrete U[1, 5] probability function, Pr(X  j)  0X2, j  1 F F F 5. B. Generate corresponding y-values from the logistic probability function, Pr( y t  1jx t )  [ exp ( b 10  b 11 x t )]aC Pr( y t  2jx t )  [ exp ( b 20  b 21 x t )]aC Pr( y t  3jx t )  1 ÀPr( y t  1jx t ) À Pr( y t  2jx t ) (12X29) where C  1  exp ( b 10  b 11 x t )  exp ( b 20  b 21 x t ). Stages A and B were repeated independently R  1000 times. C. From every population generated in stages A and B, draw a single sample using the following sampling schemes (one sample under each scheme): Ca. Poisson sampling: units are selected independently with probabilities p t  nz t a  N u1 z u , where n  300 is the expected sample size and the values z t are computed in two separate ways: Ca(1): z t (1)  Int[(5a9)y 2 t u t  2x t ]; Ca(2): z t (2)  Int[5u t  2x t ]X (12X30) The notation Int[ Á ] defines the integer value and u t $ U(0, 1). Cb. Stratified sampling: the population units are stratified based on either the values z t (1) (scheme Cb(1)) or the values z t (2) (scheme Cb(2)), yielding a total of 13 strata in each case. Denote by S (h) ( j) the strata defined by the values z t ( j) such that for units t P S h ( j), z t ( j)  z (h) ( j), j  1, 2. Let N (h) ( j) represent the corres- ponding strata sizes. The selection of units within the strata was carried out by simple random sampling without replacement (SRSWR), with the sample sizes n (h) ( j) fixed in advance. The sample sizes were determined so that the selection probabilities are similar to the corresponding selection probabilities under the Poisson sam- pling scheme and  h n h ( j)  300, j  1, 2. The following points are worth noting: 1. The sampling schemes that use the values z t (1) are informative, as the selection probabilities depend on the y-values. The sampling schemes that 188 FITTING GLMS UNDER INFORMATIVE SAMPLING [...]... est 0 .63 Mean Std est Boot 0 .66 0.30 0.31 0.20 0.19 0.22 0.50 0.51 0 .68 0 .63 0 .66 0.50 0.51 0.20 0.20 0.22 W-weighting (Equation (12.15)): True values Mean estimate Empirical Std Mean Std est Mean Std est Boot 1.00 1.00 0.71 0 .65 0 .68 0.30 0.31 0.22 0.21 0.22 0.50 0.52 0.72 0 .66 0 .69 0.50 0.51 0.22 0.21 0.22 Ordinary MLE (Equation (12. 16) ): True values 1.00 Mean estimate 0.99 Empirical Std 0 .63 Mean... (complex) 0.37 (0. 16) 0.35 (0. 16) 0.32 (0.13) 0. 165 (0.028) 0.150 (0.024) 0.143 (0.022) 0.049 (0.018) 0.034 (0.011) 0.034 (0.009) 1951 60 OLS GLS (iid) GLS (complex) 0. 56 (0.11) 0.85 (0.09) 0.85 (0.09) 0.1 46 (0.021) 0.109 (0.047) 0.1 06 (0.044) 0.048 (0.015) 0.0 26 (0.047) 0.0 26 (0.045) After 1 960 OLS GLS (iid) GLS (complex) 0.49 (0.08) 0.41 (0.07) 0.40 (0.07) 0.155 (0.018) 0.154 (0.0 16) 0.150 (0.0 16) 0.071 (0.014)... T at which data are collected correspond to the time points for which the model of interest is specified In this case, longitudinal analysis may be considered as a form of multivariate analysis of the vectors of T responses ( yil , F F F , yiT ) and the question of how to handle complex sampling schemes in the selection of s is a natural generalisation of this question for cross-section surveys More... usually not tied to the timing of the waves of the longitudinal survey For example, it is unlikely that the timing of births, marriages or job starts will coincide with the waves of a survey Since such timing is central to the aims of event history analysis, it is natural to define event history models in a continuous time framework, unconstrained by the timing of the survey data collection, and this is... characteristics of individuals and temporal `transitory' variation within individuals Another important use of random effects models in the analysis of longitudinal data is in allowing for the effects of time-constant unobserved covariates in regression models (e.g Solon, 1989; Hsiao, 19 86; Baltagi, 2001) Failure to allow for these unobserved covariates in regression analysis of cross-sectional survey data may... model, longitudinal data enable the `cross-sectional' variance s2 s2 of yit to be decomposed into the variance s2 of the `permanent' 1 2 1 component yi and the variance s2 of the `transitory' component at each wave 2 2 06 RANDOM EFFECTS MODELS FOR LONGITUDINAL SURVEY DATA This may aid understanding of the mobility of individuals over time in terms of their place in the distribution of the response variable... Mean estimate Empirical Std Mean Std est Mean Std est Boot 1.00 1.05 0 .66 0 .62 0 .63 0.30 0.30 0.21 0.20 0.22 0.50 0.52 0 .67 0 .63 0 .65 0.50 0.51 0.21 0.20 0.23 Ordinary MLE (Equation (12. 16) ): True values 1.00 Mean estimate 1.03 Empirical Std 0 .60 Mean Std est 0.58 Mean Std est Boot 0. 56 4 b11 0.30 0.30 0.19 0.19 0.20 0.50 0.49 0 .61 0.59 0.57 0.50 0.52 0.20 0.19 0.20 Our last comment refers to Table... (12.15)): True values 1.00 Mean estimate 1.00 Empirical Std 0 .67 Mean Std est 0 .63 Mean Std est Boot 0 .65 0.30 0.31 0.21 0.19 0.21 0.50 0.53 0 .67 0 .62 0 .64 0.50 0.51 0.20 0.19 0.20 Ordinary MLE (Equation (12. 16) ): True values 1.00 Mean estimate 0.29 Empirical Std 0 .60 Mean Std est 0.59 Mean Std est Boot 0.57 0.30 0.43 0.19 0.18 0.18 0.50 0. 06 0 .60 0.58 0.58 0.50 0.59 0.18 0.18 0.18 191 SIMULATION RESULTS... Std est Mean Std est Boot 1.00 1.05 0.55 0.58 0 .66 0.30 0.30 0. 16 0.18 0.21 0.50 0.53 0.58 0.58 0 .64 0.50 0.50 0.17 0.18 0.20 W-weighting (Equation (12.15)): True values 1.00 Mean estimate 1.07 Empirical Std 0.58 Mean Std est 0 .61 Mean Std est Boot 0 .63 0.30 0.29 0.18 0.19 0.21 0.50 0.55 0 .61 0 .61 0 .63 0.50 0.50 0.18 0.19 0.21 Ordinary MLE (Equation (12. 16) ): True values 1.00 Mean estimate 0.37 Empirical... proportion of the population of pensionable age We first assess the fit of Models A and B (defined in Section 14.1) for each of the four cohorts The results are presented in Table 14.1 We use goodness-offit tests based on the covariance structure approach of Section 14.2, with three choices of the matrix V in (14.8): 214 RANDOM EFFECTS MODELS FOR LONGITUDINAL SURVEY DATA Table 14.1 Goodness -of- fit test . Std 0 .67 0.21 0 .67 0.20 Mean Std est. 0 .63 0.19 0 .62 0.19 Mean Std est. Boot 0 .65 0.21 0 .64 0.20 Ordinary MLE (Equation (12. 16) ): True values 1.00 0.30 0.50 0.50 Mean estimate 0.29 0.43 0. 06 0.59 Empirical. 0.50 Mean estimate 0.99 0.31 0.51 0.51 Empirical Std 0 .67 0.20 0 .68 0.20 Mean Std est. 0 .63 0.19 0 .63 0.20 Mean Std est. Boot 0 .66 0.22 0 .66 0.22 W-weighting (Equation (12.15)): True values 1.00. est. 0 .65 0.21 0 .66 0.21 Mean Std est. Boot 0 .68 0.22 0 .69 0.22 Ordinary MLE (Equation (12. 16) ): True values 1.00 0.30 0.50 0.50 Mean estimate 0.99 0.31 0.50 0.51 Empirical Std 0 .63 0.19 0 .63 0.19 Mean

Ngày đăng: 14/08/2014, 09:20

Từ khóa liên quan

Tài liệu cùng người dùng

  • Đang cập nhật ...

Tài liệu liên quan