Announcement

Collapse
No announcement yet.
X
  • Filter
  • Time
  • Show
Clear All
new posts

  • Sebastian Kripfganz
    replied
    1. No, adding more lags of the dependent variable as regressors does not mean that you also need to start with higher lags (a1) for the instruments. The reason for starting with the second lag is that the first lag is correlated with the first-differenced error term. The second lag is uncorrelated with the first-differenced error term if the errors are serially uncorrelated. This does not depend on the number of lags of the dependent variable used as regressors. In fact, the more lags of the dependent variable you use as regressors, the more likely it is that the errors are indeed serially uncorrelated.
    2. model(fod) has the advantage that the transformed errors are still serially uncorrelated if the untransformed errors were serially uncorrelated, while model(diff) produces first-order serial correlation in the transformed error term. As long as you make sure that your instruments are uncorrelated with the transformed error term, it should not really matter which of the two model transformations you use. However, there is one additional benefit of model(fod): If your panel data set is unbalanced with gaps, the model(diff) would lose more observations than model(fod) does. Regarding model(level), this still contains the unobserved time-invariant "fixed effects" (which are removed by the other model transformations) such that you need to take extra care to ensure that your instruments are uncorrelated with them. This can often be hard to justify. Please see my 2019 London Stata Conference presentation and the references therein for details: Kripfganz, S. (2019). Generalized method of moments estimation of linear dynamic panel data models. Proceedings of the 2019 London Stata Conference.
    3. If all firms stay in the same industry throughout the entire sample, i.e. if the industry classification remains constant over time, then model(diff), model(fod), and model(mdev) all account for these effects. In fact, they account for all time-invariant effects by removing them from the transformed model. If you still obtain estimates for some industry effects, this would mean that there must be some variation over time in the industry classification or that you have combined the model(mdev) instruments with further instruments for model(level).
    4. It looks like the two specifications should be the same, aren't they?

    Leave a comment:


  • Nishant Kathuria
    replied
    Hi Prof. Kripfganz,

    I am writing to ask very basic questions that may seem too primitive to you and the audience:

    1. If I use dependent var with 2 lags, for instance, xtdpdgmm L(0/2).y x1 x2 x3, model(diff) gmm(y, lag(a1 . )) gmm(x1 x2 x3, lag(a2. )) , do I need to start a1 with 3 and a2 with 1? My understanding is that if I use a1=1 or 2, then the y variable (because of L(0/2)) in the equation will be endogenous with the controls. Please let me know if I am correct.

    2. How do I decide if I need to use model(diff) or model(fod) or mod(level)?

    3. In a study with large # of firm-year observations, does the model(mdev) control for the industry effects automatically or do I need to add industry_effects* in the equation? When I use both the industry effects and model(mdev), very few industry effects appear and most of the industries are omitted.

    4. I am using a four-way interaction, with one of the variables as time effect such that t=0 or t=1, to test difference in differences.
    Can I use xtdpdgmm L(0/2).y c.x1##c.x2##c.x3##c.t, model(diff) gmm(y, lag(3 . )) gmm(c.x1##c.x2##c.x3##c.t, lag(1 . )) teffects instead of
    xtdpdgmm L(0/2).y c.x1##c.x2##c.x3##c.t, model(diff) gmm(y, lag(3 . )) gmm(c.x1#c.x2#c.x3#c.t c.x1#c.x2#c.x3 c.x1#c.x2#c.t c.x2#c.x3#c.t c.x1#c.x3#c.t c.x1#c.x2 c.x2#c.x3 c.x1#c.x3 c.x1#c.t c.x2#c.t c.x3#c.t x1 x2 x3 , lag(1 . )) teffects ?

    I look forward to your reply. Thank you so much for continued support

    Nishant

    Leave a comment:


  • Rita Juliana
    replied
    Thank you Prof Kripfganz for your comment and suggestion. I will try your suggestions, really appreciate your comment.

    Leave a comment:


  • Sebastian Kripfganz
    replied
    I do not have an answer to the memory question. As you quote from Roodman's paper, T should be relatively small. With small T, adding time dummies is indeed recommended and not an issue because the number of extra coefficients remains small. With such a huge T, the dynamic panel bias arising from the inclusion of the lagged dependent variable is indeed not a concern. You could thus simply use xtreg if all other variables are exogenous! If you need to treat other variables as endogenous, you would still need instrumental variables. You could possibly use xtivreg. You would not need to use a system-GMM estimator. Instruments for the first-differenced equation would be sufficient. If you still want to include a large number of time dummies, the community-contributed ivreghdfe command might be helpful as it allows to absorb these dummies without actually adding a coefficient for each of them.
    Last edited by Sebastian Kripfganz; 04 Jun 2020, 03:49.

    Leave a comment:


  • Rita Juliana
    replied
    Thank you for replying Prof. Kripganz,

    Originally posted by Sebastian Kripfganz View Post
    There appears to be a problem with insufficient memory space on your computer. This is not surprising given the dimensions of your data set. The xtdpdgmm command internally creates some matrices that will become very large with that many time periods. It is not optimized for such data sets as it is designed for estimations with relatively few time periods. Adding time effects is not a good idea in your case because this will create a huge number of extra coefficients to be estimated.
    in this case, how much memory space will be sufficient?

    also, according to roodman (2009), in his conclusion he suggests to include time dummies, i'm afraid that my result will be bias if i don't include time dummies, what do think?
    Include time dummies. The autocorrelation test and the robust estimates of the coefficient standard errors assume no correlation across individuals in the idiosyncratic disturbances. Time dummies make this assumption more likely to hold.
    another thing, roodman (2009) also mentioned this:
    Apply the estimators to “small T, large N” panels. If T is large, dynamic panel bias becomes insignificant, and a more straightforward fixed-effects estimatorworks. Meanwhile, the number of instruments in difference and system GMM tends to explode with T. If N is small, the cluster–robust standard errors and the Arellano–Bond autocorrelation test may be unreliable.
    in my case, is it okay to use OLS xtreg to estimate since my data have 100 T ? could you give me some suggestion? thank you so much

    Leave a comment:

Working...
X