Announcement

Collapse
No announcement yet.
X
  • Filter
  • Time
  • Show
Clear All
new posts

  • Carlo Lazzaro
    replied
    Mansi:
    another possible shortcoming of using OLS with count dependent variables is that the predicted values can be negative (which does not make sense for a regressand that can take on integer, non negative values only).

    Leave a comment:


  • Mansi Jain
    replied
    Dear Long Hong,

    Thank you for your question! I chose to use a count data model because my dependent variable can only be a non-negative integer and the linear model doesn't take into account the limited support for this variable. I'm also estimating the linear fixed effects model, but I think a Poisson model makes more sense for recreation count data.

    Regards,
    Mansi

    Leave a comment:


  • Mansi Jain
    replied
    Dear Jeff Wooldridge,

    Thank you so much for your advice! That's very helpful. I hadn't quite understood the mechanics of how AIC/ BIC are measuring fit and to what extent that translates to unbiased coefficients, so I really appreciate the clarity your comment gave me. What I'm gathering from this discussion is that I can use xtpoisson if I decide to cluster just at the unit level and xtpqml if I want to cluster at the park level. One last clarifying question if I use xtpqml: the standard errors should be 'robust' irrespective of whether I cluster at the level of the fixed effects or a higher level, right? I'm guessing this is the case but just want to be sure; the xtpqml documentation isn't very helpful on this front.

    Thank you so much again!

    Regards,
    Mansi



    Leave a comment:


  • Long Hong
    replied
    Hi Mansi: for your research questions to be answered, I think you can first focus on the linear model with fixed effects. Is there a specific reason why you do not want to use a linear model? (e.g. is misspecification a first-order issue in your estimation?) Hope it helps.

    Leave a comment:


  • Jeff Wooldridge
    replied
    Mansi: Here are some more thoughts. It appears you don't have panel data but grouped data. Is that correct? It doesn't change much. (Edit: I reread your initial post and see that you do have true panel data, possibly in addition to clusters at the park level.)

    1. If you want to cluster at the level that you are allowing fixed effects, vce(robust) and using the cluster(unit_id) with xtpqml should deliver essentially the same answer. They are both robust to any misspecification of the Poisson distribution and within-cluster correlation.
    2. xtpqml has one advantage: it allows clustering at a higher level than the fixed effects. So, if you think you have data clustered at, say, the school level, but FEs are included at the student level, you can cluster at the school level. I'm not sure why Stata 15 did not allow this. (I haven't checked with Stata 16).
    3. I wouldn't recommend a multilevel model because then you'll be back to assuming the covariates are independent of the heterogeneity. It is not worth the tradeoff. It's possible to allow both, but it's pretty advanced and not done by Stata (as far as I know). Instead, use the Poisson FE and cluster at the appropriate level.
    4. It's not very informative to compare measures of fit based on log likelihoods. Of course the NegBin could fit better than the Poisson if you are to choose a full distributional specification. But it's very likely that both are wrong because of the very precise variance/mean relationship and the serial independence. That's why you use FE Poisson because, for estimating the conditional mean, you have full robustness. This is a common mistake. In the cross-sectional case, NegBin effectively adds a parameter. So it will fit better than the Poisson in terms of log likelihoods. But the Poisson is fully robust to any distributional misspecification.
    Last edited by Jeff Wooldridge; 22 Jul 2019, 21:05.

    Leave a comment:


  • Mansi Jain
    replied
    Dear Carlo Lazzaro,

    Thank you so much! The link was helpful in understanding what is going on. I'll also look into using a mixed model to take into account the spatial correlation.

    Related to point (2) about clustering standard errors, I actually read some sources online that suggested that the vce(robust) option for xtpoisson actually clusters at the unit level by default, so those errors are corrected for misspecification as well as correlation within observations belonging to the same panel. I also read some other sources that suggested that doing vce(cluster unit_id) actually not only clusters standard errors at the unit level but is also robust to misspecification. Lastly, I read something that suggested using xtpqml instead of xtpoisson to get clustered errors that are also robust to problems like overdispersion.

    Stata documentation for the vce types above isn't very helpful, so I would love any help I can get to figure out which of the above are true.

    Thank you so much!

    Leave a comment:


  • Carlo Lazzaro
    replied
    Mansi:
    among many other entries on incidental parameter bias, the following one (and related references) can be helpful: http://methods.johndavidpoe.com/2016...rameters-bias/

    -cluster()-ing standard errors at park level does not necessarily take correlation within observations belonging to the same panel into account.
    That said, if your panel units are actually clustered within parks, you may want to consider a mixed model design (see -help meglm-).
    Last edited by Carlo Lazzaro; 22 Jul 2019, 09:51.

    Leave a comment:


  • Mansi Jain
    replied
    Dear Carlo Lazzaro and Jeff Wooldridge,

    Thank you so much for your help and advice! I really appreciate it. What I'm gathering from this discussion is that I don't even need to test for over dispersion - FE NegBin is definitely not the way to go, and as long as I cluster the standard errors, FE Poisson should work just fine.

    I think I'm still a little unsure on these three fronts:

    (i) Carlo Lazzaro, you mentioned that xtpoisson is a conditional fixed effects model rather than a pure fixed effects model. What is the difference between the two? I'm not able to find helpful references to explain that at a glance.

    (ii) I think I'm still uncertain about how to correct standard errors in the FE Poisson model. I see that I can use vce (cluster park_unit) to cluster standard errors at the park level to account for some level of autocorrelation between units in the same park. However, I can't do vce(robust) at the same time. I'm wondering how I can both have standard errors that are robust to misspecification (such as presumed equality of conditional mean and variance) and clustered at the appropriate level. In the worse case scenario, I would at least want them to be robust to misspecification and clustered at the unit level (so that observations from the same unit across time are not assumed to be independent). Perhaps the vce (robust) option in a panel data model like xtpoisson is already clustering at the unit level, but I'm not sure.

    (iii) For whatever reason, the AIC and BIC for an xtnbreg model I just tried to see what would happen is lower than the xtpoisson model. Given the problems Jeff Wooldridge mentioned above, should I ignore this entirely, or does this deserve some explanation/ attention?

    Thank you so much! I'm very grateful for your support.

    Leave a comment:


  • Jeff Wooldridge
    replied
    There are so many problems with the FE NegBin approach that it should probably never be used. When the model was originally proposed by Hausman, Hall, and Griliches (1984, Econometrica), it was thought that it allows two forms of heterogeneity. In my 1999 Journal of Econometrics paper I showed that, in fact, the model collapses to depend on only one heterogeneity parameter.

    Below is the list of shortcomings of the NB approach. The Poisson suffers none of them. I showed in my 1999 Journal of Econometrics paper that the Poisson FE estimator is completely robust to every failure of the Poisson assumptions -- except, of course, for having the correct conditional mean. The FENB in the panel data case does not come close to nesting the Poisson assumptions unless the heterogeneity is zero.

    Here's a summary of why I tell people to avoid FE NegBin. I hope it helps. There's no need to test for overdispersion. And, with short T panel data, that's not so easy, anyway. And because the Poisson FE is completely robust, you wouldn't do anything if you find it.

    1. FE NegBin imposes a very specific overdispersion of the form (1 + c(i)) where the mean effect is c(i). Why this would ever be true is beyond me.
    a. There's only one source of heterogeneity.
    b. Does not nest the Poisson except in the uninteresting case.
    c. The Poisson estimator allows any kind of variance-mean relationship. Some units can be overdispersed, some underdispersed. The same unit can exhibit both depending on the covariate values.
    2. FE NegBin imposes conditional independence.
    a. Serial correlation is not allowed.
    b. The Poisson FE allows any kind of serial correlation. One just needs to cluster the standard errors.
    3. FE NegBin is not known to be robust to failure of any of its assumptions.
    4. Time constant variables do not drop out in the FENB estimation!
    a. This is what people mean when they say it's not a "true" FE procedure.
    5. The actual estimation of the FENB model often fails to converge, very likely because of the weird overdispersion it requires for every unit i in the cross section.

    Leave a comment:


  • Carlo Lazzaro
    replied
    Mansi:
    the deeply missed Joe Hilbe, who was for many years a relevant contributor of this forum following textbooks in addition of being a very kind and approachable person off the list, left two comprehensive textbook on the topics you're interested in, that can well complete the other sources you quote (for the future, as reminded by the FAQ, please provide full references. By chance, I know the textbooks/papers you refer to, but others on this forum may ignore them simply because they nevere challenged themselves with count data models. Thanks)
    1) https://www.stata.com/bookstore/modeling-count-data/
    2) https://www.stata.com/bookstore/nega...al-regression/.

    As far as I can remember, cluster robust standard errors correct for apparent overdipersion, whereas -nbreg- is the way to go when you have detected real overdispersion (as it is often the case with -poisson-).
    However, you can still use cluster robust standard errors with -nbreg- if you take autocorrelation into account.
    You're right that -fe- in -xtpoisson- actually is conditional -fe- (incidental parameter bias, you know...); if you're looking for -fe- in -xtreg- fashion, you should perform a pooled -poisson- with -i.panelid- among predictors.

    Leave a comment:

Working...
X