Hello Statalists,
I am using poisson regression for a three-way count dataset (5*5*5) and p-value and BIC index indicate the different results in model selection
Model 1 is nested in Model 2. Model 1 has the p-value (<.05), which means that it does not fit the data, but its BIC values is lower (more negative) than that of Model 2. The p-values of Model 2 is larger than .05, which means that this model fits the data well. Deviance is reduced in Model 2 and Likelihood test prefers Model 2 over Model 1.
A paper that is relevant to my research topic shows the similar results and prefers Model 2 over Model 1.The authors note that BIC is used as a model selection criterion when the sample size so large that no model would fit the data. The Ns (before I collapsed the data into the three-way count format) range from 1296 to 5832.However, I may need to note, if relevant, that the authors only presents these results in their working paper version and in the publication paper, they used a different sampling method and does not report the model selection results.
My codes are as following. The variables a, b, c are 5-class categorical variables. I am using Stata 12 SE.
I am confused at the moment. Which model does fit the model better? Whether the interaction term (a##c) that is added in Model 2 is statistically meaningful?
My guess is that Model 2 fits the data better than does Model 1, because they are nested and likelihood test prefers Model 2. The interaction term (a##c) is statistically significant and should be taken into account in models, but its magnitude of the effect might be a bit limited.
I am not sure if my guess is right or not and how convincing it is.
I am using poisson regression for a three-way count dataset (5*5*5) and p-value and BIC index indicate the different results in model selection
Model 1 is nested in Model 2. Model 1 has the p-value (<.05), which means that it does not fit the data, but its BIC values is lower (more negative) than that of Model 2. The p-values of Model 2 is larger than .05, which means that this model fits the data well. Deviance is reduced in Model 2 and Likelihood test prefers Model 2 over Model 1.
A paper that is relevant to my research topic shows the similar results and prefers Model 2 over Model 1.The authors note that BIC is used as a model selection criterion when the sample size so large that no model would fit the data. The Ns (before I collapsed the data into the three-way count format) range from 1296 to 5832.However, I may need to note, if relevant, that the authors only presents these results in their working paper version and in the publication paper, they used a different sampling method and does not report the model selection results.
My codes are as following. The variables a, b, c are 5-class categorical variables. I am using Stata 12 SE.
Code:
. // Model 1 . quietly poisson count i.a##i.b i.b##i.c . estat gof Deviance goodness-of-fit = 116.6563 Prob > chi2(80) = 0.0047 Pearson goodness-of-fit = 118.975 Prob > chi2(80) = 0.0031 . quietly fitstat, saving(M1) . . // Model 2 . quietly poisson count i.a##i.b i.b##i.c i.a##i.c . estat gof Deviance goodness-of-fit = 58.46142 Prob > chi2(64) = 0.6718 Pearson goodness-of-fit = 58.84949 Prob > chi2(64) = 0.6586 . fitstat, using(M1) Measures of Fit for poisson of count Current Saved Difference Model: poisson poisson N: 125 125 0 Log-Lik Intercept Only: -1673.722 -1673.722 0.000 Log-Lik Full Model: -303.837 -332.934 29.097 D: 607.673(34) 665.868(59) -58.195(-25) LR: 2739.770(60) 2681.575(44) 58.195(16) Prob > LR: 0.000 0.000 0.000 McFadden's R2: 0.818 0.801 0.017 McFadden's Adj R2: 0.764 0.762 0.002 Maximum Likelihood R2: 1.000 1.000 0.000 Cragg & Uhler's R2: 1.000 1.000 0.000 AIC: 6.317 6.383 -0.066 AIC*n: 789.673 797.868 -8.195 BIC: 443.511 380.998 62.513 BIC': -2450.071 -2469.130 19.058 Difference of 19.058 in BIC' provides very strong support for saved model.
I am confused at the moment. Which model does fit the model better? Whether the interaction term (a##c) that is added in Model 2 is statistically meaningful?
My guess is that Model 2 fits the data better than does Model 1, because they are nested and likelihood test prefers Model 2. The interaction term (a##c) is statistically significant and should be taken into account in models, but its magnitude of the effect might be a bit limited.
I am not sure if my guess is right or not and how convincing it is.
Comment