The sample size isn't large enough for the test to mean anything.
The sample size seems more than adequate, since several significant associations were found. Why do you say it isn't large enough?
Linda
The sample size isn't large enough for the test to mean anything.
i disagree with the sample size being big enough. because the results are based on correlation the sample size has to be VERY BIG as any correlation could be a freak chance, especially in the hundreds as in this case. i think you'd need at least 10k to be sure, i may even push for a 100k.
like i've said, there doesn't even seem to be a correlation or at least a good clear correlation. whatever happened to personality being nurture and not nature? this sort of classification of people makes me very uncomfortable.
the sample size will technically never be big enough because it's completely correlation, but 430 (or however many) is no where near enough considering all the possible outcomes. huge freak chances often occur from more than 3 times the number of possible outcomes over 10k samples. 430 is not enough to come to ANY conclusion.But that's the point of using statistical methods! We all realize that we don't want to draw conclusions from freak occurrences. So subjecting something to significance testing is effectively saying "there is a one in 20 chance that this is a freak occurrence". You've got it backwards. If the sample size is too small, then any differences, including freaky chance differences, are likely to be not statistically significant. Small samples prevent you from detecting real differences. They do not artificially inflate your chances of concluding that spurious results are significant.
I don't think the problem with this study is sample size. It is with performing dozens of comparisons, but holding them to a standard that is only suitable for one comparison. There is a helluva difference between trying once to roll double sixes and trying 90 times to roll double sixes.
yes, it could have uses. but personality may not even be genetic so using genetics to determine someones personality may not have any value if it turns out personality is nurture rather than nature. if they want to show it's use with mental conditions, then they should have done so.And, even if these results are real and reproducible, their predictive ability may be miniscule - e.g. if there is a 10 percent chance you are neurotic (an example of a baseline frequency) and someone looks in your eyes and sees furrows, that may change it to a 10.5 percent chance that you are neurotic.
the sample size will technically never be big enough because it's completely correlation, but 430 (or however many) is no where near enough considering all the possible outcomes. huge freak chances often occur from more than 3 times the number of possible outcomes over 10k samples. 430 is not enough to come to ANY conclusion.
The practice of faking one's personality is much more common than you might think. Some do it for personal gain and awards, others for political power and re-election.That's the strange beauty of a personality test. Unlike an IQ test, personality can be faked, and people do fake. In fact (though I don't claim expertise here, I have read some of the lit), the current consensus is that willingness to fake a personality test is itself a personality trait, and that that variance is captured by the big 5, and that faking in general seems to have little effect on validity.
the sample size will technically never be big enough because it's completely correlation, but 430 (or however many) is no where near enough considering all the possible outcomes. huge freak chances often occur from more than 3 times the number of possible outcomes over 10k samples. 430 is not enough to come to ANY conclusion.
yes, it could have uses. but personality may not even be genetic so using genetics to determine someones personality may not have any value if it turns out personality is nurture rather than nature. if they want to show it's use with mental conditions, then they should have done so.
Also, the Meyer article I cited was not a pro-personality test article. The claim in the article was that psych testing in general produces validities that rival or beat those found in other areas of "real science".
I think it's ok to use the term "test" loosely, even in the context of determining whether some medical treatment works (as another example, the EEOC considers an employment intervew to be a test). Screening mammograms is a medical test, and it has only .32 validity for detecting breast cancer (within 1 year).
The validity of using aspirin to reduce heart attacks compared with the validity of using personality to predict job performance, imo, is apples to apples.
I cherry picked the examples because they seemed most interesting to me. The article has about 7 pages of examples. It's a good read for anyone interested in testing in general.
Linda, thanks!
I think in terms of effect sizes-- the mean difference between groups on some outcome measure.
I'm pretty sure effect size can by synonymous with validity.
Looking at people who do and do not take aspirin (categorizing people into discrete groups here to make the example easier to understand) the effect size is .08*. In other words, aspirin has a .08 validity / effect on reducing the incidence of heart attacks.
That's a small effect, but if you did an extreme groups comparison-- followed 10000 people around who never took aspirin and 10000 others who always did, at the end there would be a statistically significant mean difference showing less heart attacks in the aspirin group.
The .30 validity for C predicting job performance is small in the sense that only 9% (.3 x .3) of the variance in job performance is explained by C.
That said, it's free to measure and takes 5 minutes. The return on investment for using C to hire people will be amazingly high, even though the validity is .30.
And, going back to the effect size example, get 100 people low in C and compare them to 100 people high in C. I'd bet money you'd find significant and non-trivial differences in job performance across the two groups.
*I'm pretty sure the correlation coefficient is the effect size, but not positive. I'm more used to seeing effect sizes that are caluclated by subtracting the mean difference between 2 groups, and then dividing by some measure of error (d ' ). Does anyone know how to calculate effect sizes from correlation coefficients, or is the correlation the measure of the effect?