To make a point to some friends about declining returns in the number of possible responses on a scale, and the number of items to be scaled, I wanted to whip up a quick Stata example. But short of taking the normal curve and chopping it up with a bunch of "replace x1_3=0 if x3<.5 and x1>-.5" etc (and it of course gets worse depending on the number of responses), I somehow am having a mental blockage.
Ideally the process would be this:
1) generate normal variable
2) generate a dozen more normal variables, with 50% of their variances shared with the variable from step 1 -- that is, a dozen indicators of the original variable, with known loadings.
3) for each of the dozen indicators of the original variable, "coarsen" them into three, five, seven, nine, and eleven-category ordinal variables, approximately normally distributed.
4) compute alphas for the different levels of coarsening, and for using different number of items (e.g. a twelve-by-six grid). Yes, I know alpha is inappropriate for ordinal variables, but half the point is to see where and how far you can violate assumptions and approach the results you get under normality.
Hmmmm... I know a simple lit search would turn up oodles of information on this issue, with tables just like I describe, but it seemed at first blush such a simple task, and some people respond better to numbers they just saw generated (plus, we could play additional what-ifs with CFAs and such).
If this is confusing or doesn't seem worth doing, please don't bother putting much thought into it; like I said, there already exists a pretty good literature on the basic theme. But if somebody wants a bit of a brain teaser and/or knows exactly how to crank this out, much obliged. Thanks for any thoughts!
BTW -- steps 1 and 2 are obvious no-brainers. Step 3 is where I get befuddled.
Coarsening them into *uniform* distribution is easy. But I want to have it roughly approximate normality, that is, more answers in the middle and fewer in the extremes, though I realize that "true" normality would be lost.
Step 4 I can do, though getting the relevant output compiled in an efficient and creative way would be a bonus.
Ideally the process would be this:
1) generate normal variable
2) generate a dozen more normal variables, with 50% of their variances shared with the variable from step 1 -- that is, a dozen indicators of the original variable, with known loadings.
3) for each of the dozen indicators of the original variable, "coarsen" them into three, five, seven, nine, and eleven-category ordinal variables, approximately normally distributed.
4) compute alphas for the different levels of coarsening, and for using different number of items (e.g. a twelve-by-six grid). Yes, I know alpha is inappropriate for ordinal variables, but half the point is to see where and how far you can violate assumptions and approach the results you get under normality.
Hmmmm... I know a simple lit search would turn up oodles of information on this issue, with tables just like I describe, but it seemed at first blush such a simple task, and some people respond better to numbers they just saw generated (plus, we could play additional what-ifs with CFAs and such).
If this is confusing or doesn't seem worth doing, please don't bother putting much thought into it; like I said, there already exists a pretty good literature on the basic theme. But if somebody wants a bit of a brain teaser and/or knows exactly how to crank this out, much obliged. Thanks for any thoughts!
BTW -- steps 1 and 2 are obvious no-brainers. Step 3 is where I get befuddled.
Coarsening them into *uniform* distribution is easy. But I want to have it roughly approximate normality, that is, more answers in the middle and fewer in the extremes, though I realize that "true" normality would be lost.
Step 4 I can do, though getting the relevant output compiled in an efficient and creative way would be a bonus.
Comment