The full text on this page is automatically extracted from the file linked above and may contain errors and inconsistencies.
Federal Reserve Bank of Chicago Estimation of a Transformation Model with Truncation, Interval Observation and Time-Varying Covariates Bo E. Honoré and Luojia Hu WP 2009-16 Estimation of a Transformation Model with Truncation, Interval Observation and Time–Varying Covariates∗ Bo E. Honor´† e Luojia Hu‡ First complete version: December 2007. This Version: September 2009. Abstract Abrevaya (1999b) considered estimation of a transformation model in the presence of left–truncation. This paper observes that a cross–sectional version of the statistical model considered in Frederiksen, Honor´, and Hu (2007) is a generalization of the e model considered by Abrevaya (1999b) and the generalized model can be estimated by a pairwise comparison version of one of the estimators in Frederiksen, Honor´, and e Hu (2007). Specifically, our generalization will allow for discretized observations of the dependent variable and for piecewise constant time–varying explanatory variables. ∗ This research was supported by the National Science Foundation and the Gregory C. Chow Econometric Research Program at Princeton University. We thank seminar participants at Rice University, Universit´ e Paris 1 Panth´on–Sorbonne and the Federal Reserve Bank of Chicago as well as members of Princeton’s e Microeconometric Reading Group for comments. The opinions expressed here are those of the authors and not necessarily those of the Federal Reserve Bank of Chicago or the Federal Reserve System. † Mailing Address: Department of Economics, Princeton University, Princeton, NJ 08544-1021. Email: honore@Princeton.edu. ‡ Mailing Address: Economic Research Department, Federal Reserve Bank of Chicago, 230 S. La Salle Street, Chicago, IL 60604. Email: lhu@frbchi.org. 1 1 Introduction The transformation model h (Ti∗ ) = g (Xi β) + εi (1) is often used to model durations. In models like this it is important to allow for right censoring and sometimes also for left truncation because the samples used in many applications include spells that are in progress at the start of the sample period. See Abrevaya (1999b). It is also sometimes desirable to allow for the dependent variable to be discretized so that one observes only whether it falls in a particular interval. So the observed duration, Ti , would be t if Ti∗ ∈ (t − 1, t]. See Prentice and Gloeckler (1978) or Meyer (1990). Moreover, in duration models it is often interesting to allow for time–varying covariates, which are not easily directly incorporated into the transformation model (Flinn and Heckman (1982)). The contribution of this paper is to specify a statistical model that allows for interval observations and time–varying covariates but which simplifies to a model with interval observations from (1) when the covariates are time invariant. We then propose an estimator for the parameters of the model. The estimator can be interpreted as a generalization of the truncated maximum rank correlation estimator proposed in Abrevaya (1999b). Consider first the transformation model, (1), with strictly increasing h (·) and g (·) and with εi independent of Xi and continuously distributed with full support. In this model P (Ti∗ > t| Xi ) = P (h (Ti∗ ) > h (t)| Xi ) = P (g (Xi β) + εi > h (t)| Xi ) = 1 − F (h (t) − g (Xi β)) where F is the CDF for εi . This gives P ( Ti∗ > t| Xi , Ti∗ > t − 1) = 1 − F (h (t) − g (Xi β)) 1 − F (h (t − 1) − g (Xi β)) where the assumption that εi has full support guarantees that the denominator is not 0. When 1−F (·) is log–concave (which is implied by the density of εi being log–concave; see Heckman and Honor´ (1990)), the right–hand side is an increasing function of g (Xi β) and e 2 hence of Xi β . See the Appendix. This means that one can write the event Ti∗ > t| Xi , Ti∗ > t − 1 in the form 1 {Xi β > η it } for some (possibly infinite) random variable η it that is independent of Xi and has CDF 1−F (h(t)−·) . 1−F (h(t−1)−·) Therefore, if we define Yit ≡ 1 {Ti∗ ∈ (t − 1, t]} = 1 {Ti = t} then we can write Yit = 1 {Xi β − η it ≤ 0} for t such that Yil = 0 (2) l≤t−1 In other words, a transformation model with discretized observations of the dependent variable and log–concave errors is a special case of the model Yit = 1 {Xit β − η it ≤ 0} for t such that Yil = 0 (3) l≤t−1 where the difference between (2) and (3) is that the latter allows for time–varying covariates. Note that this line of argument is valid even if Ti is left truncated. It is interesting to note that Abrevaya (1999b) also assumes log–concavity of 1 − F (·).1 Note that although log–concavity implies an increasing hazard for h (T ∗ ), it does not impose such a restriction on T ∗ .2 Equation (3) is a cross–sectional version of the model considered by Frederiksen, Honor´, e and Hu (2007). It is well understood that estimators of panel data models can be turned into estimators of cross–sectional models by considering all pairs of observations as units in a panel. See, for example, Honor´ and Powell (1994), who apply the idea in Honor´ (1992) e e 1 The assumption of log–concavity also appears elsewhere in the literature on truncated regression models. See, for example, Lee (1993) and (Honor´ 1992). e ∗ Let λ (·) denote the hazard for ε. The hazard for h (T ∗ ) is then − ∂ log(P ( h(T )>t|X)) = ∂t ∂ log(1−F (t−g (X β ))) − = λ (t − g (X β)). When 1 − F (·) is log–concave this is an increasing function of t. On ∂t 2 ∗ T the other hand, the hazard for T ∗ is − ∂ log(P (∂t >t|X)) ∗ = − ∂ log(P ( h(T )>h(t)|X)) = λ (h (t) − g (X β)) h (t). ∂t 2 The derivative of this with respect to t is λ (h (t) − g (X β)) h (t) + λ (h (t) − g (X β)) h (t), which can be of either sign. 3 for panel data censored regression models to all pairs of observations in a cross section. The insights in Frederiksen, Honor´, and Hu (2007) can therefore be used to construct an e estimator of β. We pursue this in Section 3 after discussing the model in more detail in Section 2. 2 The model Consider a spell with integer–valued duration, Ti , that starts at (integer–valued) time −Vi ≤ 0. Following the discussion above, we model the event that the spell lasts s periods conditional on lasting at least s − 1, by the qualitative response model Yis = 1 {Xis β − η is ≤ 0} , (4) where η is is independent of Xis and the distribution of η is is allowed to change over time. When there is left truncation, one must distinguish between duration time and calendar time. We will index the observables, Y and X, by calendar time, and the unobservable η by duration time. At first sight, this difference is confusing, but it is necessitated by the fact that the discussion in the previous section implied that one should allow the distribution of η is to vary by duration time. On the other hand, it seems natural to denote the first observation for an individual by t = 1. With this notation, we assume that we observe (Yit , Xit ) starting at t = 1, where Yit = 1 Xit β − η i,t+Vi ≤ 0 . (5) With this notation, ηs with the same time subscript will have the same distribution under the class of models discussed above. We will let Ti denote the first time that Yit equals 1. Since Yit is not defined after the end of the spell, and since we want to allow for random right censoring, we assume that we observe Yit from t = 1 until, and including, Ti or until a random censoring time Ci − 1 (whichever comes first). In other words, we observe (Yit , Xit ) for t = 1, 2, ..., T i where T i = min {Ti , Ci − 1} So when an observation is censored, Ci will be 4 the first time period in which individual i is not observed. We also assume that we observe the presample duration, Vi , for each observation. The statistical assumption on the errors in (5) is that conditional on Vi and on {Yis = 0 for s < t}, η i,t+Vi is independent of Ci , {Xis }s≤t . As explained in Section 1, if the errors are log–concave and the covariates are time–invariant, this is exactly what is implied by an underlying transformation model for Ti∗ , where we observe whether a spell that started at time −Vi and was in progress at time t − 1 is still in progress at time t. Note that when the covariates are time–varying they are not restricted to be strictly exogenous, and that the censoring times can be covariate–dependent, as long as they do not depend on the ηs. In the next section we will apply the insight of Frederiksen, Honor´, and Hu (2007) to e construct an estimator for β under these assumptions when the researcher has access to a random sample of individuals. 3 The estimator The key insight for the construction of the estimator can be easily illustrated if we ignore censoring first (so T i = Ti for all i). Let t1 and t2 be arbitrary. Consider the two events A = {Ti = t1 , Tj > t2 } and B = {Ti > t1 , Tj = t2 } where t1 + Vi = t2 + Vj . Under the stated assumptions, it then follows immediately from Lemma 1 of (Frederiksen, Honor´, e 1 >2 P ( A| A ∪ B, Xit1 , Xjt2 , Vi , Vj ) =1 2 <1 2 and Hu 2007) that if (Xit1 − Xjt2 ) β > 0, if (Xit1 − Xjt2 ) β = 0, if (Xit1 − Xjt2 ) β < 0. This suggests estimating β by maximizing i<j Ti t1 =1 Tj t2 =1 1 {t1 + Vi = t2 + Vj } · 1 {Ti = t1 , Tj > t2 } · 1 (Xit1 − Xjt2 ) b > 0 + 1 {Ti > t1 , Tj = t2 } · 1 (Xit1 − Xjt2 ) b < 0 5 (6) (6) is the same as one of the objective functions in (Frederiksen, Honor´, and Hu 2007), e except that that paper considers a panel data situation. It is convenient to rewrite (6) as i<j 1 {Tj + Vj > Ti + Vi > Vj } · 1 XiTi − Xj,Ti +Vi −Vj b > 0 +1 {Vi < Tj + Vj < Ti + Vi } · 1 (7) XiTj +Vj −Vi − Xj,Tj b < 0 This has the same structure as the maximum rank correlation estimator developed in Han (1987). When there is censoring, (6) can be modified to i<j Ti t1 =1 Tj t2 =1 1 {t1 + Vi = t2 + Vj , t1 < Ci , t2 < Cj } (8) · 1 {Ti = t1 , Tj > t2 } · 1 (Xit1 − Xjt2 ) b > 0 +1 {Ti > t1 , Tj = t2 } · 1 (Xit1 − Xjt2 ) b < 0 . And equation (7) can be rewritten as i<j 1 T j + Vj > Ti + Vi > Vj , Ti < Ci · 1 +1 Vi < Tj + Vj < T i + Vi , Tj < Cj · 1 XiTi − Xj,Ti +Vi −Vj b > 0 (9) XiTj +Vj −Vi − Xj,Tj b < 0 . The intuition for the estimator is essentially based on pairwise comparisons. Specifically, we compare an individual i who was observed to fail at time Ti (and thus had a complete duration Ti + Vi ) to all other observations j that survived up to the same duration. At the true parameter value β, if the index for individual i at the time he/she failed, XiTi β, is larger than the index for the comparable individual j at the time that corresponds to the same duration, Xj,Ti +Vi −Vj β, then individual j is likely to survive longer than individual i (that is, T j + Vj > Ti + Vi ). Note the set of comparison observations j includes censored spells provided the censoring time occurs after Ti + Vi − Vj . The additional inequality in the indicator function, Ti + Vi > Vj , ensures that the time at which j is being compared to i is within the sample period (i.e., not truncated). 6 Again this estimator has the same structure as Han’s (1987) maximum rank correlation estimator and the asymptotic distribution is therefore the one given in Sherman (1993) under the regularity conditions stated there. 4 Asymptotic properties Consistency and asymptotic normality can be established as in (Sherman 1993), Abrevaya (1999b) or Khan and Tamer (2007). First note that some normalization of the parameter is needed, since the parameter vector is only identified up to scale. For example, we can normalize the last component of β to be 1. The two key assumptions for consistency of the estimator are (1) at least one component of the explanatory variable X is continuously distributed with full support, and (2) the error η has full support. Without the first assumption, the parameter is not identified, since a small change in the parameter value could leave the ranking of the index unchanged. The second assumption on the error guarantees that the set of effective observations that make a nonzero contribution to the objective function is not empty. Both assumptions are standard in the semi-parametric estimation literature. To establish asymptotic normality results, we need some additional notations. Denote Di = 1 {Ti < Ci }, which is an observable variable indicating a complete (uncensored) spell. The objective function can be rewritten as 1 n (n − 1) n Di · 1 Ti + Vi > Vj , Tj + Vj > Vi , T j + Vj > Ti + Vi i=1 j=i ·1 XiTi b > Xj,Ti +Vi −Vj b . Define the function τ t, t, d, v, {xs }s≤t , b ≡ E Di · 1 Ti + Vi > v, t + v > Vi , t + v > Ti + Vi · 1 XiTi b > xTi +Vi −v b 7 (10) +E d · 1 t + v > Vj , Tj + Vj > v, T j + Vj > t + v · 1 xt b > Xj,t+v−Vj b = E Di · 1 Ti + Vi > v, t + v > Vi , t + v > Ti + Vi · 1 XiTi b > xTi +Vi −v b +E d · 1 t + v > Vi , Ti + Vi > v, T i + Vi > t + v · 1 xt b > Xi,t+v−Vi b . Following Theorem 4 of Sherman (1993), we have √ ∼ n β − β −→ N 0, 4Γ−1 ∆Γ−1 (11) where Γ = E 1 Ti , T i , Di , Vi , {Xis } , β ∆ = E with 2τ 1τ Ti , T i , Di , Vi , {Xis } , β and 1τ Ti , T i , di , Vi , {Xis } , β denoting the first– and second–derivative operator, respectively. 2 Following Sherman (1993), we can further express the variance-covariance matrix in terms of “model primitives.” Specifically, ∆ = VX Xs2 − µs1 Xs2 β S T, T , D, V, s1 , s2 , Xs2 β gXs1 β Xs2 β s1 ,s2 and Γ = EX Xs2 − µs1 Xs2 β Xs2 − µs1 Xs2 β s1 ,s2 S7 T, T , D, V, s1 , s2 , Xs2 β · gXs1 β Xs2 β where Xs2 is composed of the first K − 1 coordinates of Xs2 (the ones corresponding to the piece of β that is not normalized to 1), gXs1 β (λ) is the marginal density of Xs1 β, µs1 (λ) = E Xs1 | Xs1 β = λ and S t, t, d, v, s1 , s2 , λ = E Ai t, t, d, v, s1 , s2 Xis1 β = λ 8 and Ai t, t, d, v, s1 , s2 = Di · 1 Ti + Vi > v, t + v > Vi , t + v > Ti + Vi , Ti = s1 , Ti + Vi − v = s2 −d · 1 t + v > Vi , Ti + Vi > v, T i + Vi > t + v, t = s2 , t + v − Vi = s1 The asymptotic variance matrix can be estimated by plugging in the estimator β and calculating sample analogs of Γ and ∆ using numerical derivatives based on a smoothed version of τ . See Section 6 for more discussion. 5 Relationship to other estimators The estimator proposed in the previous section is related to a number of existing estimators, and it coincides with some of them in special situations. For example, when Ci = 2 for all i, and with no left truncation (so Vi = 0 for all i), (5) is a standard discrete choice model, and in that case the objective function in (9) becomes n 1 {Yi > Yj } · 1 (Xi − Xj ) b > 0 + 1 {Yi < Yj } · 1 (Xi − Xj ) b < 0 i<j which is the objective function for Han’s (1987) maximum rank correlation estimator. When there is left truncation, and the covariates are time invariant and there is no censoring, the estimator defined by maximizing (9) is the same as the truncated maximum rank correlation estimator in (Abrevaya 1999b). This is most easily seen by noting that without censoring and with time–invariant covariates, (10) becomes 1 n (n − 1) = 1 n (n − 1) n 1 {Ti + Vi > Vj , Tj + Vj > Vi , Tj + Vj > Ti + Vi } · 1 Xi b > Xj b i=1 j=i n 1 {Ti + Vi > Vj , Tj + Vj > Vi } 1 {Tj + Vj > Ti + Vi } · 1 Xi b > Xj b . i=1 j=i Except for the difference in notation and the normalization by n (n − 1), this is exactly equation (7) in Abrevaya (1999b). 9 Khan and Tamer (2007) consider a model with left censoring as well as right censoring, whereas we allow for left truncation as well as right censoring. When there is neither left censoring nor left truncation, and when the covariates are time invariant, the estimator defined by maximizing (9) coincides with the estimator proposed by Khan and Tamer (2007), except that ours applies to discretized durations and theirs to exactly measured durations, and we allow for time–varying covariates. Whether right censoring or right truncation is more interesting depends on the specific application. Left truncation will, for example, be relevant if as in Frederiksen, Honor´, and Hu (2007), the duration of interest is the length of e employment on a given job, and one has information on a sample of workers observed between two fixed points in time. In this case the durations are left truncated, because spells that ended before the start of the sampling will not appear in the data, and it is crucial for our approach that one observes the duration of employment in the current job at the start of the sampling. On the other hand, models with both left and right censoring are, for example, useful for estimation of models where the dependent variable is a fraction, which is restricted to be between zero and one, and where both zeros and ones are likely to be observed. See, for example Alan and Leth-Petersen (2006). Both Khan and Tamer (2007) and we allow the censoring points at the right to be observed only when the observation equals a censoring point. Khan and Tamer (2007) also allow the left–censoring point to be unobserved when an observation is not left–censored, whereas we assume that the truncation point is observed for everybody who is not truncated, but not for truncated durations. Both papers assume that one observes the actual duration, and not just the duration from the censoring/truncation point. In the duration contexts we have in mind, this is the most severe assumption.3 The framework here is also closely related to standard statistical duration models with discretized observations. The proportional hazard model can be written as Z (T ) = −x β + ε, 3 See also Heckman and Singer (1986) for a discussion of the effect of different sample schemes on the analysis of duration data. 10 where Z is the log integrated baseline hazard and ε has an extreme value distribution. Prentice and Gloeckler (1978), Meyer (1990) and (Hausman and Woutersen 2005) study a version of this model with interval observations. Meyer (1990) and Hausman and Woutersen (2005) also allow for time–varying explanatory variables and for ε to be a sum of an extreme value distributed random variable and a random variable that captures unobserved heterogeneity. While the estimation in Meyer (1990) is likelihood–based and hence fundamentally different from ours, the structure of the estimator proposed in Hausman and Woutersen (2005) shares many of the features of the estimator proposed here. The main difference is that theirs is based on a comparison of the integrated hazards rather than just the current index, X β. As a result, the approach does not seem to generalize to models with left truncation. On the other hand, log–concavity plays no role in Hausman and Woutersen (2005). 6 Monte Carlo experiment In this section, we conduct a small scale Monte Carlo study to illustrate the proposed estimation method and investigate its finite sample performance. We also demonstrate how to conduct inference and examine how good an approximation the asymptotic distribution provides for finite samples. The designs are based on the following: • All of the designs have two explanatory variables. • β = (β 1 , β 2 ) = (1, 2) . The parameter of interest is θ = β 2 /β 1 . The fact that this is one dimensional greatly simplifies the computations. • Time–varying intercept β 0 = −4 + (s/10)1.2 . This introduces duration dependence beyond the duration dependence introduced by the shape of F and by the choice of h (·). 11 • The time between the start of a spell and the first period of observation is uniformly distributed on the integers between 1 and 5. • The censoring time is generated as the minimum of ten periods from the start of the spell and Q periods from the start of the sample, where Q is uniformly distributed on the integers between 1 and 8. We consider a number of designs within this framework. Design 1: Dynamic Probit. The two explanatory variables are generated by i.i.d. draws from a bivariate normal distribution with zero means, variances equal to 2 and 1, respectively, and covariance equal to 1. In this design, η i,t is i.i.d N (0, 4). Design 2: Transformation Model Hazard. This design is set up as a generalization of a transformation model. Specifically, using the notation of Section 1, we assume that h (u) = log (u), g (u) = u, and ε ∼ N (0, 1). Using the derivation in Section 1, this yields P (Ti = t) = P Ti∗ < t| {Xis }s≤t , Ti∗ > t − 1 = 1 − 1 − Φ (log (t) − Xit β) . 1 − Φ (log (t − 1) − Xit β) As in Design 1, the two explanatory variables are generated by i.i.d. draws from a bivariate normal distribution with zero means, variances equal to 2 and 1, respectively, and covariance equal to 1.. Design 3: Feedback. Recall that our model does not require the explanatory variables to be strictly exogenous. In this design we therefore allow for feedback from the error η to future values of X. Specifically, we follow Design 1 except that the explanatory variables are defined by • X2t = η t−1 for t > 1 and standard normal for t = 1. • X1t = X2t + N (0, 1). 12 Design 4: Covariate–Dependent Censoring and Truncation. Our model allows censoring and truncation to be correlated with explanatory variables. In this design, we follow the basic structure of Design 1 but let censoring be defined by the outcome of a probit with explanatory variable X1t . Design 5: Dynamic Probit 2. This design is like Design 1 except that • X2s ∼ N (0.5 − 0.2s, 1). • X1s = 1 {X2s + N (0, 1) > 0} • η i,s ∼ N (0, 0.1 + (0.15s)) The summary statistics for the five designs are reported in Table 1. For each design, 100,000 observations are drawn from the underlying data–generating process. We then compute the fraction of the sample that is censored, the fraction that is truncated, and the mean and standard deviation of the underlying duration. Below, we report Monte Carlo results for the point estimates of β as well as for the performance of tests statistics based on the asymptotic distribution in Section 4. To do this, we estimate the components of the variance of the estimators by sample analogs of smoothed versions of the components.4 Recall that Γ=E 4 2τ Ti , T i , Di , Vi , {Xis } , β A recent paper by Subbotin (2007) has shown that the nonparametric bootstrap can be used to estimate the quantiles and variance of various maximum rank correlation estimators. The structure of our estimator is essentially the same as that of the maximum rank correlation estimators he considers. We therefore conjecture that the bootstrap could have been used to estimate the variance in our case as well, although this would increase the computational burden. 13 and ∆=E 1τ Ti , T i , Di , Vi , {Xis } , β 1τ Ti , T i , Di , Vi , {Xis } , β where τ t, t, d, v, {xs }s≤t , b = E Di · 1 Ti + Vi > v, t + v > Vi , t + v > Ti + Vi · 1 XiTi b > xTi +Vi −v b +E d · 1 t + v > Vi , Ti + Vi > v, T i + Vi > t + v · 1 xt b > Xi,t+v−Vi b . We then estimate τ by the smoothed version τ = t, t, d, v, {xs }s≤t , b 1 n n + Di · 1 Ti + Vi > v, t + v > Vi , t + v > Ti + Vi · Φ i=1 n 1 n d · 1 t + v > Vi , Ti + Vi > v, T i + Vi > t + v · Φ i=1 XiTi b − xTi +Vi −v b h xt b − Xi,t+v−Vi b h Then 1τ 1 nh = t, t, d, v, {xs }s≤t , b n Di · 1 Ti + Vi > v, t + v > Vi , t + v > Ti + Vi i=1 XiTi b − xTi +Vi −v b XiTi − xTi +Vi −v h +d · 1 t + v > Vi , Ti + Vi > v, T i + Vi > t + v xt b − Xi,t+v−Vi b φ xt − Xi,t+v−Vi , h φ and 2τ = −1 nh3 t, t, d, v, {xs }s≤t , b n Di · 1 Ti + Vi > v, t + v > Vi , t + v > Ti + Vi i=1 XiTi b − xTi +Vi −v b φ XiTi − xTi +Vi −v 14 XiTi b − xTi +Vi −v b h XiTi − xTi +Vi −v . +d · 1 t + v > Vi , Ti + Vi > v, T i + Vi > t + v xt b − Xi,t+v−Vi b XiTi b − xTi +Vi −v b φ h xt − Xi,t+v−Vi And therefore Γ= and 1 ∆= n 1 n xt − Xi,t+v−Vi . n 2τ Ti , T i , Di , Vi , {Xis } , β i=1 n 1τ Ti , T i , Di , Vi , {Xis } , β 1τ Ti , T i , Di , Vi , {Xis } , β . i=1 As mentioned earlier, β is only identified up to scale. One possibility is to normalize one of the coefficients to 1, and hence essentially focus on β 2 /β 1 or β 1 /β 2 . Unfortunately, this normalization will lead to different MAE and RMSE depending on which of the coefficients is normalized. So if one were to compare different estimators, one might reach different conclusions depending on a seemingly innocent normalization. This is unsatisfactory in models where there is only one parameter. For this reason, it is likely to be better to consider θ = log (β 2 /β 1 ) = log (β 2 ) − log (β 1 ) the parameter of interest. This means that the true parameter is log (2) ≈ 0.693 for all of the designs. We estimate θ by a grid search over the interval between − log (6) and log (6) with equal grids of size 1 . 200 Since the parameter of interest is one dimensional, the line search is feasible despite the fact that the calculation of the objective function requires O (n2 ) operations. When it is of higher dimension, it would be beneficial to use the method described in Abrevaya (1999a) to calculate the objective function in O (n · log (n)) operations. We calculate the variance of θ by applying the so– called δ–method to (11). For each design, the Monte Carlo experiment is conducted with 5,000 replications for each of the 5 sample sizes: 100, 200, 400, 800 and 1,600. The results are reported in Tables 2–6. Overall, the results across the designs are broadly consistent with predictions from the asymptotic theory. Some additional remarks are in order. First, the results illustrate the consistency of the estimator, since both the median ab15 solute error (MAE) and root mean squared error (RMSE) decrease as sample size increases. Moreover, the estimator is close to median unbiased even for small sample sizes. Second, the theory predicts that the estimator converges to the true parameter value √ at the rate n . This is borne out in the simulation as the MAE and RMSE decrease √ toward zero at a rate of approximately 2 when the sample size is doubled. For example, a regression of the log of the median absolute error on the log of the sample size (and design dummies) yields a coefficient of −0.543 with a standard error of 0.006. Third, to examine the normality prediction from the asymptotic theory, we estimate the density for θ − θ and plot the kernel estimate in Figures 1-5. The left–hand side of each figure gives the estimated density of the estimator of (β 2 /β 1 ) centered at the true value. They show severe asymmetry in the distribution of the estimator: it tends to be skewed to the left, especially in small samples. As mentioned, this is expected because of the somewhat unnatural normalization. The right–hand side of the figures shows the estimated density of the estimator of log (β 2 /β 1 ), again centered at the true value. There one can see that the distribution becomes more symmetric and closer to normal as sample size increases. Finally, the asymptotic theory suggests that we can conduct inference using t–tests. Under the null, the test statistic should follow a standard normal distribution. In the simulation, we compute a t–statistic for each of the 5,000 estimates θ and calculate the fraction of times the null is rejected at the 20% level. We focus on tests with (nominal) size of 20% rather than the conventional 5% because the results for the latter are likely to be more erratic for a finite number of simulations. The results are reported for various bandwidths used in the estimation of the asymptotic variance-covariance matrix of the estimator. In general, the rejection rate is closer to the nominal size of the test when the bandwidth is smaller and the sample size is larger. For example, for the bandwidth (0.05, 0.20) and sample size 1,600, the reject rate is 21.8%, 22.7%, 20.8% and 21.6% for Designs 1 through 4.5 These 5 Different bandwidths are used in estimating the matrix ∆ and Γ. The latter is based on a second derivative, and one would therefore expect it to require a larger bandwidth than the former. 16 are close to being statistically indistinguishable from the nominal size. The performance of the test under some other combinations of bandwidth and sample size is less encouraging. The test also performs less well under Design 5. We speculate that this is because of the discreteness of xi1 . The last row reports the reject rates computed using the average of the variance–covariance matrix estimated over all the bandwidth choices. Overall, the t-test tends to over–reject the null. It is interesting to compare our results to a standard logit or probit estimation of (5) where one uses x1 , x2 and time dummies as explanatory variables. Since we expect them to perform comparably, we focus on the logit maximum likelihood estimator.6 Designs 1, 3 and 4 are all correctly specified probit models, so one would expect the logit estimator to do well for this design. This is confirmed in panels one, three and four of Table 7. The bias is small and the MAE and RMSE fall at a rate close to root–n. It is less clear what to expect for Designs 2 and 5. Panel 2 of Table 7 shows that the logit estimator does well for Design 2. It appears to be close to unbiased and its MAE and RMSE fall at a rate close to root–n. One potential explanation for this is that misspecified maximum likelihood estimators often do well when the explanatory variables are jointly normally distributed. See, for example, Ruud (1983). Design 5 shows a situation where the logit estimator does relatively poorly. The bias is quite high, and as a result, the MAE and RMSE do not fall rapidly as the sample size increases. 7 Conclusion In this paper we propose a generalization of the transformation model that is appropriate for studying duration outcome with truncation, censoring, interval observations of the dependent variable, and time-varying covariates. We develop an estimator for this model, discuss its 6 Since our estimator of θ was calculated by a grid search over the interval between − log (6) and log (6), we censored the logit maximum likelihood estimator of β 2 /β 1 to be in the interval between 17 1 6 and 6. asymptotic properties and investigate its finite sample performance via a Monte Carlo study. Overall, the results suggest that the estimator performs well in finite samples, and the asymptotic theory provides a reasonably good approximation to the distribution. We also investigate test–statistics for the estimator. Those require estimation of the asymptotic variance of the estimator. This is somewhat sensitive to different choices of bandwidth. Investigating the optimal bandwidth choice in this case could be an interesting future research topic. References Abrevaya, J. (1999a): “Computation of The Maximum Rank Correlation Estimator,” Economics Letters, 62, 279–285(7). (1999b): “Rank Estimation of a Transformation Model With Observed Truncation,” Econometrics Journal, 2(2), 292–305. Alan, S., and S. Leth-Petersen (2006): “Tax Incentives and Household Portfolios: A Panel Data Analysis,” Center for Applied Microeconometrics, University of Copenhagen, Working paper number 2006-13. Flinn, C. J., and J. J. Heckman (1982): “Models for the Analysis of Labor Force Dynamics,” in Advances in Econometrics, ed. by R. Basmann, and G. Rhodes. JAI, Greenwich, Conn. ´ Frederiksen, A., B. E. Honore, and L. Hu (2007): “Discrete Time Duration Models with Group–Level Heterogeneity,” Journal of Econometrics, 141, 1014–1043. Han, A. (1987): “Nonparametric Analysis of a Generalized Regression Model,” Journal of Econometrics, 35, 303–316. Hausman, J. A., and T. Woutersen (2005): “Estimating a Semi-Parametric Duration Model without Specifying Heterogeneity,” Johns Hopkins University. 18 ´ Heckman, J. J., and B. E. Honore (1990): “The Empirical Content of the Roy Model,” Econometrica, 58, 1121–1149. Heckman, J. J., and B. Singer (1986): “Econometric Analysis of Longitudinal Data,” in Handbook of Econometrics, ed. by Z. Griliches, and M. D. Intriligator, vol. 3, pp. 1689– 1763. North-Holland, Handbooks in Economics series, book. ´ Honore, B. E. (1992): “Trimmed LAD and Least Squares Estimation of Truncated and Censored Regression Models with Fixed Effects,” Econometrica, 60, 533–565. ´ Honore, B. E., and J. L. Powell (1994): “Pairwise Difference Estimators of Censored and Truncated Regression Models,” Journal of Econometrics, 64, 241–78. Khan, S., and E. Tamer (2007): “Partial Rank Estimation of Duration Models with General Forms of Censoring,” Journal of Econometrics, 136(1), 251–280. Lee, M.-j. (1993): “Quadratic mode regression,” Journal of Econometrics, 57(1-3), 1 – 19. Meyer, B. D. (1990): “Unemployment Insurance and Unemployment Spells,” Econometrica, 58, 757–782. Prentice, R. L., and L. A. Gloeckler (1978): “Regression Analysis of Grouped Survival Data with Application to Breast Cancer Data,” Biometrics, 34, 57–67. Ruud, P. A. (1983): “Sufficient Conditions for the Consistency of Maximum Likelihood Estimation Despite Misspecification of Distribution in Multinomial Discrete Choice Models,” Econometrica, 51(1), 225–228. Sherman, R. T. (1993): “The Limiting Distribution of the Maximum Rank Correlation Estimator,” Econometrica, 61, 123–137. Subbotin, V. (2007): “Asymptotic and Bootstrap Properties of Rank Regressions,” SSRN eLibrary. 19 8 Appendix Assume that H (·) is a log–concave function and let f (w) = H (a2 − w) H (a1 − w) where a2 > a1 . Let w1 < w2 and ∆a = a2 − a1 , ∆w = w2 − w1 and λ= ∆a ∆a + ∆w then a2 − w2 = λ (a2 − w1 ) + (1 − λ) (a1 − w2 ) so by concavity of ln (H (·)), ln (H (a2 − w2 )) > λ ln (H (a2 − w1 )) + (1 − λ) ln (H (a1 − w2 )) . (12) Also a1 − w1 = (1 − λ) (a2 − w1 ) + λ (a1 − w2 ) so ln (H (a1 − w1 )) > (1 − λ) ln (H (a2 − w1 )) + λ ln (H (a1 − w2 )) Adding (12) and (13) yields ln (H (a2 − w2 )) + ln (H (a1 − w1 )) > ln (H (a2 − w1 )) + ln (H (a1 − w2 )) and ln (f (w2 )) − ln (f (w1 )) = (ln (H (a2 − w2 )) − ln (H (a1 − w2 ))) − (ln (H (a2 − w1 )) − ln (H (a1 − w1 ))) > 0 Hence f is an increasing function. 20 (13) Table 1: Summary Statistics for the Designs Design 1 Design 2 Design 3 Design 4 Fraction Truncated 0.260 0.349 0.316 0.260 Fraction Censored 0.297 0.100 0.190 0.425 Mean Duration 5.317 3.592 4.369 5.317 Standard Deviation of Duration 3.422 2.023 2.918 3.422 Median MAE Mean RMSE 0.05, 0.20 0.05, 0.40 0.10, 0.20 0.10, 0.40 0.20, 0.20 0.20, 0.40 0.40, 0.20 0.40, 0.40 average Table 2: Results for Design 1 Performance of Estimator n = 100 n = 200 n = 400 n = 800 n = 1600 0.721 0.695 0.702 0.693 0.693 0.331 0.215 0.155 0.105 0.070 0.727 0.706 0.702 0.698 0.694 0.488 0.331 0.227 0.155 0.106 Significance when testing at 20% level n = 100 n = 200 n = 400 n = 800 n = 1600 0.332 0.240 0.212 0.207 0.218 0.163 0.152 0.187 0.241 0.271 0.421 0.323 0.280 0.256 0.246 0.243 0.232 0.261 0.296 0.302 0.494 0.382 0.323 0.285 0.260 0.324 0.301 0.305 0.326 0.319 0.536 0.423 0.350 0.303 0.269 0.390 0.347 0.337 0.343 0.329 0.289 0.259 0.261 0.274 0.273 21 Design 5 0.183 0.543 7.270 3.937 Median MAE Mean RMSE 0.05, 0.20 0.05, 0.40 0.10, 0.20 0.10, 0.40 0.20, 0.20 0.20, 0.40 0.40, 0.20 0.40, 0.40 average Median MAE Mean RMSE 0.05, 0.20 0.05, 0.40 0.10, 0.20 0.10, 0.40 0.20, 0.20 0.20, 0.40 0.40, 0.20 0.40, 0.40 average Table 3: Results for Design 2 Performance of Estimator n = 100 n = 200 n = 400 n = 800 n = 1600 0.718 0.713 0.698 0.693 0.692 0.540 0.370 0.255 0.175 0.125 0.656 0.712 0.705 0.697 0.695 0.752 0.553 0.392 0.274 0.186 Significance when testing at 20% level n = 100 n = 200 n = 400 n = 800 n = 1600 0.423 0.350 0.268 0.236 0.227 0.164 0.175 0.180 0.224 0.265 0.515 0.422 0.335 0.280 0.253 0.269 0.252 0.245 0.273 0.296 0.578 0.477 0.372 0.302 0.269 0.359 0.317 0.293 0.303 0.313 0.620 0.514 0.394 0.317 0.279 0.426 0.363 0.322 0.319 0.323 0.333 0.299 0.274 0.273 0.275 Table 4: Results for Design 3 Performance of Estimator n = 100 n = 200 n = 400 n = 800 n = 1600 0.683 0.699 0.693 0.696 0.688 0.400 0.263 0.174 0.120 0.085 0.691 0.716 0.707 0.700 0.694 0.569 0.405 0.271 0.185 0.124 Significance when testing at 20% level n = 100 n = 200 n = 400 n = 800 n = 1600 0.389 0.289 0.214 0.201 0.208 0.216 0.150 0.143 0.202 0.255 0.468 0.372 0.288 0.264 0.247 0.280 0.226 0.225 0.273 0.302 0.551 0.455 0.345 0.308 0.273 0.368 0.307 0.287 0.316 0.327 0.608 0.503 0.384 0.330 0.287 0.438 0.370 0.325 0.344 0.342 0.326 0.268 0.247 0.266 0.273 22 Median MAE Mean RMSE 0.05, 0.20 0.05, 0.40 0.10, 0.20 0.10, 0.40 0.20, 0.20 0.20, 0.40 0.40, 0.20 0.40, 0.40 average Median MAE Mean RMSE 0.05, 0.20 0.05, 0.40 0.10, 0.20 0.10, 0.40 0.20, 0.20 0.20, 0.40 0.40, 0.20 0.40, 0.40 average Table 5: Results for Design 4 Performance of Estimator n = 100 n = 200 n = 400 n = 800 n = 1600 0.706 0.693 0.703 0.698 0.693 0.385 0.260 0.178 0.118 0.085 0.704 0.703 0.704 0.701 0.694 0.557 0.397 0.273 0.182 0.124 Significance when testing at 20% level n = 100 n = 200 n = 400 n = 800 n = 1600 0.355 0.280 0.235 0.195 0.216 0.176 0.161 0.186 0.219 0.270 0.437 0.369 0.307 0.250 0.249 0.250 0.235 0.255 0.269 0.300 0.515 0.428 0.356 0.280 0.268 0.335 0.305 0.307 0.305 0.320 0.568 0.472 0.383 0.296 0.280 0.402 0.353 0.340 0.324 0.333 0.300 0.272 0.270 0.259 0.276 Table 6: Results for Design 5 Performance of Estimator n = 100 n = 200 n = 400 n = 800 n = 1600 0.711 0.704 0.682 0.698 0.693 0.567 0.409 0.285 0.195 0.135 0.646 0.710 0.720 0.718 0.703 0.789 0.612 0.438 0.305 0.207 Significance when testing at 20% level n = 100 n = 200 n = 400 n = 800 n = 1600 0.518 0.414 0.365 0.355 0.372 0.506 0.464 0.567 0.637 0.661 0.548 0.485 0.429 0.399 0.401 0.528 0.557 0.638 0.666 0.677 0.607 0.541 0.469 0.425 0.411 0.584 0.618 0.672 0.681 0.683 0.631 0.566 0.480 0.431 0.412 0.627 0.648 0.682 0.684 0.683 0.511 0.503 0.505 0.508 0.513 23 Median MAE Mean RMSE Median MAE Mean RMSE Median MAE Mean RMSE Median MAE Mean RMSE Median MAE Mean RMSE Table 7: Performance of the Logit MLE Design 1 n = 100 n = 200 n = 400 n = 800 0.699 0.699 0.692 0.694 0.240 0.165 0.114 0.084 0.715 0.702 0.696 0.695 0.369 0.252 0.173 0.123 Design 2 n = 100 n = 200 n = 400 n = 800 0.696 0.703 0.692 0.689 0.413 0.282 0.199 0.139 0.691 0.720 0.701 0.694 0.639 0.443 0.303 0.214 Design 3 n = 100 n = 200 n = 400 n = 800 0.691 0.692 0.695 0.692 0.302 0.202 0.137 0.098 0.711 0.710 0.701 0.696 0.454 0.307 0.209 0.144 Design 4 n = 100 n = 200 n = 400 n = 800 0.687 0.695 0.696 0.696 0.288 0.201 0.138 0.097 0.706 0.702 0.701 0.699 0.438 0.302 0.209 0.146 Design 5 n = 100 n = 200 n = 400 n = 800 0.884 1.008 1.046 1.055 0.521 0.406 0.361 0.362 0.761 1.012 1.080 1.082 0.948 0.685 0.540 0.476 24 n = 1600 0.693 0.059 0.693 0.085 n = 1600 0.695 0.100 0.695 0.149 n = 1600 0.694 0.069 0.695 0.103 n = 1600 0.692 0.067 0.693 0.100 n = 1600 1.063 0.370 1.073 0.428 Figure 1: Density of Estimation Error for Estimator and Its Log for Design 1 Figure 2: Density of Estimation Error for Estimator and Its Log for Design 2 25 Figure 3: Density of Estimation Error for Estimator and Its Log for Design 3 Figure 4: Density of Estimation Error for Estimator and Its Log for Design 4 26 Figure 5: Density of Estimation Error for Estimator and Its Log for Design 5 27 Working Paper Series A series of research studies on regional economic issues relating to the Seventh Federal Reserve District, and on financial and economic topics. U.S. Corporate and Bank Insolvency Regimes: An Economic Comparison and Evaluation Robert R. Bliss and George G. Kaufman WP-06-01 Redistribution, Taxes, and the Median Voter Marco Bassetto and Jess Benhabib WP-06-02 Identification of Search Models with Initial Condition Problems Gadi Barlevy and H. N. Nagaraja WP-06-03 Tax Riots Marco Bassetto and Christopher Phelan WP-06-04 The Tradeoff between Mortgage Prepayments and Tax-Deferred Retirement Savings Gene Amromin, Jennifer Huang,and Clemens Sialm WP-06-05 Why are safeguards needed in a trade agreement? Meredith A. Crowley WP-06-06 Taxation, Entrepreneurship, and Wealth Marco Cagetti and Mariacristina De Nardi WP-06-07 A New Social Compact: How University Engagement Can Fuel Innovation Laura Melle, Larry Isaak, and Richard Mattoon WP-06-08 Mergers and Risk Craig H. Furfine and Richard J. Rosen WP-06-09 Two Flaws in Business Cycle Accounting Lawrence J. Christiano and Joshua M. Davis WP-06-10 Do Consumers Choose the Right Credit Contracts? Sumit Agarwal, Souphala Chomsisengphet, Chunlin Liu, and Nicholas S. Souleles WP-06-11 Chronicles of a Deflation Unforetold François R. Velde WP-06-12 Female Offenders Use of Social Welfare Programs Before and After Jail and Prison: Does Prison Cause Welfare Dependency? Kristin F. Butcher and Robert J. LaLonde Eat or Be Eaten: A Theory of Mergers and Firm Size Gary Gorton, Matthias Kahl, and Richard Rosen WP-06-13 WP-06-14 1 Working Paper Series (continued) Do Bonds Span Volatility Risk in the U.S. Treasury Market? A Specification Test for Affine Term Structure Models Torben G. Andersen and Luca Benzoni WP-06-15 Transforming Payment Choices by Doubling Fees on the Illinois Tollway Gene Amromin, Carrie Jankowski, and Richard D. Porter WP-06-16 How Did the 2003 Dividend Tax Cut Affect Stock Prices? Gene Amromin, Paul Harrison, and Steven Sharpe WP-06-17 Will Writing and Bequest Motives: Early 20th Century Irish Evidence Leslie McGranahan WP-06-18 How Professional Forecasters View Shocks to GDP Spencer D. Krane WP-06-19 Evolving Agglomeration in the U.S. auto supplier industry Thomas Klier and Daniel P. McMillen WP-06-20 Mortality, Mass-Layoffs, and Career Outcomes: An Analysis using Administrative Data Daniel Sullivan and Till von Wachter WP-06-21 The Agreement on Subsidies and Countervailing Measures: Tying One’s Hand through the WTO. Meredith A. Crowley WP-06-22 How Did Schooling Laws Improve Long-Term Health and Lower Mortality? Bhashkar Mazumder WP-06-23 Manufacturing Plants’ Use of Temporary Workers: An Analysis Using Census Micro Data Yukako Ono and Daniel Sullivan WP-06-24 What Can We Learn about Financial Access from U.S. Immigrants? Una Okonkwo Osili and Anna Paulson WP-06-25 Bank Imputed Interest Rates: Unbiased Estimates of Offered Rates? Evren Ors and Tara Rice WP-06-26 Welfare Implications of the Transition to High Household Debt Jeffrey R. Campbell and Zvi Hercowitz WP-06-27 Last-In First-Out Oligopoly Dynamics Jaap H. Abbring and Jeffrey R. Campbell WP-06-28 Oligopoly Dynamics with Barriers to Entry Jaap H. Abbring and Jeffrey R. Campbell WP-06-29 Risk Taking and the Quality of Informal Insurance: Gambling and Remittances in Thailand Douglas L. Miller and Anna L. Paulson WP-07-01 2 Working Paper Series (continued) Fast Micro and Slow Macro: Can Aggregation Explain the Persistence of Inflation? Filippo Altissimo, Benoît Mojon, and Paolo Zaffaroni WP-07-02 Assessing a Decade of Interstate Bank Branching Christian Johnson and Tara Rice WP-07-03 Debit Card and Cash Usage: A Cross-Country Analysis Gene Amromin and Sujit Chakravorti WP-07-04 The Age of Reason: Financial Decisions Over the Lifecycle Sumit Agarwal, John C. Driscoll, Xavier Gabaix, and David Laibson WP-07-05 Information Acquisition in Financial Markets: a Correction Gadi Barlevy and Pietro Veronesi WP-07-06 Monetary Policy, Output Composition and the Great Moderation Benoît Mojon WP-07-07 Estate Taxation, Entrepreneurship, and Wealth Marco Cagetti and Mariacristina De Nardi WP-07-08 Conflict of Interest and Certification in the U.S. IPO Market Luca Benzoni and Carola Schenone WP-07-09 The Reaction of Consumer Spending and Debt to Tax Rebates – Evidence from Consumer Credit Data Sumit Agarwal, Chunlin Liu, and Nicholas S. Souleles WP-07-10 Portfolio Choice over the Life-Cycle when the Stock and Labor Markets are Cointegrated Luca Benzoni, Pierre Collin-Dufresne, and Robert S. Goldstein WP-07-11 Nonparametric Analysis of Intergenerational Income Mobility with Application to the United States Debopam Bhattacharya and Bhashkar Mazumder WP-07-12 How the Credit Channel Works: Differentiating the Bank Lending Channel and the Balance Sheet Channel Lamont K. Black and Richard J. Rosen WP-07-13 Labor Market Transitions and Self-Employment Ellen R. Rissman WP-07-14 First-Time Home Buyers and Residential Investment Volatility Jonas D.M. Fisher and Martin Gervais WP-07-15 Establishments Dynamics and Matching Frictions in Classical Competitive Equilibrium Marcelo Veracierto WP-07-16 Technology’s Edge: The Educational Benefits of Computer-Aided Instruction Lisa Barrow, Lisa Markman, and Cecilia Elena Rouse WP-07-17 3 Working Paper Series (continued) The Widow’s Offering: Inheritance, Family Structure, and the Charitable Gifts of Women Leslie McGranahan Demand Volatility and the Lag between the Growth of Temporary and Permanent Employment Sainan Jin, Yukako Ono, and Qinghua Zhang WP-07-18 WP-07-19 A Conversation with 590 Nascent Entrepreneurs Jeffrey R. Campbell and Mariacristina De Nardi WP-07-20 Cyclical Dumping and US Antidumping Protection: 1980-2001 Meredith A. Crowley WP-07-21 Health Capital and the Prenatal Environment: The Effect of Maternal Fasting During Pregnancy Douglas Almond and Bhashkar Mazumder WP-07-22 The Spending and Debt Response to Minimum Wage Hikes Daniel Aaronson, Sumit Agarwal, and Eric French WP-07-23 The Impact of Mexican Immigrants on U.S. Wage Structure Maude Toussaint-Comeau WP-07-24 A Leverage-based Model of Speculative Bubbles Gadi Barlevy WP-08-01 Displacement, Asymmetric Information and Heterogeneous Human Capital Luojia Hu and Christopher Taber WP-08-02 BankCaR (Bank Capital-at-Risk): A credit risk model for US commercial bank charge-offs Jon Frye and Eduard Pelz WP-08-03 Bank Lending, Financing Constraints and SME Investment Santiago Carbó-Valverde, Francisco Rodríguez-Fernández, and Gregory F. Udell WP-08-04 Global Inflation Matteo Ciccarelli and Benoît Mojon WP-08-05 Scale and the Origins of Structural Change Francisco J. Buera and Joseph P. Kaboski WP-08-06 Inventories, Lumpy Trade, and Large Devaluations George Alessandria, Joseph P. Kaboski, and Virgiliu Midrigan WP-08-07 School Vouchers and Student Achievement: Recent Evidence, Remaining Questions Cecilia Elena Rouse and Lisa Barrow WP-08-08 4 Working Paper Series (continued) Does It Pay to Read Your Junk Mail? Evidence of the Effect of Advertising on Home Equity Credit Choices Sumit Agarwal and Brent W. Ambrose WP-08-09 The Choice between Arm’s-Length and Relationship Debt: Evidence from eLoans Sumit Agarwal and Robert Hauswald WP-08-10 Consumer Choice and Merchant Acceptance of Payment Media Wilko Bolt and Sujit Chakravorti WP-08-11 Investment Shocks and Business Cycles Alejandro Justiniano, Giorgio E. Primiceri, and Andrea Tambalotti WP-08-12 New Vehicle Characteristics and the Cost of the Corporate Average Fuel Economy Standard Thomas Klier and Joshua Linn WP-08-13 Realized Volatility Torben G. Andersen and Luca Benzoni WP-08-14 Revenue Bubbles and Structural Deficits: What’s a state to do? Richard Mattoon and Leslie McGranahan WP-08-15 The role of lenders in the home price boom Richard J. Rosen WP-08-16 Bank Crises and Investor Confidence Una Okonkwo Osili and Anna Paulson WP-08-17 Life Expectancy and Old Age Savings Mariacristina De Nardi, Eric French, and John Bailey Jones WP-08-18 Remittance Behavior among New U.S. Immigrants Katherine Meckel WP-08-19 Birth Cohort and the Black-White Achievement Gap: The Roles of Access and Health Soon After Birth Kenneth Y. Chay, Jonathan Guryan, and Bhashkar Mazumder WP-08-20 Public Investment and Budget Rules for State vs. Local Governments Marco Bassetto WP-08-21 Why Has Home Ownership Fallen Among the Young? Jonas D.M. Fisher and Martin Gervais WP-09-01 Why do the Elderly Save? The Role of Medical Expenses Mariacristina De Nardi, Eric French, and John Bailey Jones WP-09-02 Using Stock Returns to Identify Government Spending Shocks Jonas D.M. Fisher and Ryan Peters WP-09-03 5 Working Paper Series (continued) Stochastic Volatility Torben G. Andersen and Luca Benzoni WP-09-04 The Effect of Disability Insurance Receipt on Labor Supply Eric French and Jae Song WP-09-05 CEO Overconfidence and Dividend Policy Sanjay Deshmukh, Anand M. Goel, and Keith M. Howe WP-09-06 Do Financial Counseling Mandates Improve Mortgage Choice and Performance? Evidence from a Legislative Experiment Sumit Agarwal,Gene Amromin, Itzhak Ben-David, Souphala Chomsisengphet, and Douglas D. Evanoff WP-09-07 Perverse Incentives at the Banks? Evidence from a Natural Experiment Sumit Agarwal and Faye H. Wang WP-09-08 Pay for Percentile Gadi Barlevy and Derek Neal WP-09-09 The Life and Times of Nicolas Dutot François R. Velde WP-09-10 Regulating Two-Sided Markets: An Empirical Investigation Santiago Carbó Valverde, Sujit Chakravorti, and Francisco Rodriguez Fernandez WP-09-11 The Case of the Undying Debt François R. Velde WP-09-12 Paying for Performance: The Education Impacts of a Community College Scholarship Program for Low-income Adults Lisa Barrow, Lashawn Richburg-Hayes, Cecilia Elena Rouse, and Thomas Brock Establishments Dynamics, Vacancies and Unemployment: A Neoclassical Synthesis Marcelo Veracierto WP-09-13 WP-09-14 The Price of Gasoline and the Demand for Fuel Economy: Evidence from Monthly New Vehicles Sales Data Thomas Klier and Joshua Linn WP-09-15 Estimation of a Transformation Model with Truncation, Interval Observation and Time-Varying Covariates Bo E. Honoré and Luojia Hu WP-09-16 6