Rotman School of Management
&
Department of Psychology
University of Toronto
Poster presented at the APA/Yale Conference on Intelligence, June 2000
The underlying assumption in much of the work on intelligence is that, with increasing g, there is increasing ability to do many different things. A number of researchers have recently rediscovered a challenge to this point of view: the so-called Divergence Hypothesis.
It was as far back as 1927 that Spearman first noted that g followed a law of diminishing returns (Spearman, 1927: p. 219). He labeled this the divergence hypothesis. This states that at high levels of g, abilities are not as closely associated as they are at lower levels of g. Evans (1999) notes that:
One can visualize this in three dimensions by imagining the various abilities as flowers arranged in a narrow vase - at the bottom, they are bound together tightly, at the top they spread out broadly. (p. 1059)
Despite the enormous amount of research on intelligence and the arguments about whether or not single (Jensen, 1998) or multiple factors (Gardner, 1983) are required to assess intelligence, the divergence hypothesis was lost sight of until fairly recently. This despite the fact that such a hypothesis implies the presence of a single factor at the lower end of the g spectrum and multiple factors at the higher end of the g spectrum.
Detterman & Daniel (1989) were the first, albeit inadvertently (Deary & Pagliari, 1991) to rediscover this law of diminishing returns. At high levels of g they found low intercorrelations between tests in the WAIS-R and the WISC-R. Their work was followed by a number of other investigators using a variety of different intelligence tests (ASVAB, DAT-T, etc.). We summarize these in Table 1 and a summary chart (Figure 1) based on a meta-analysis of the data(1). These data show clearly that as g increases, the intercorrelation between the various scales decreases. The pattern is fairly consistent for the many different ways in which g was measured (i.e., how the sample was classified to identify the high and low g groups). This was usually done by rotating through all the scales in the test as the classifier and looking at the correlations between the remaining tests in the high and low groups. Only Evans (1999) tried to find independent assessments of g that were not part of the test battery under investigation. Most authors have argued that classification on scales that are highly g-loaded result in a stronger divergence effect. This meta-analysis does not support this position.(2)
If g is no longer homogeneous but becomes less strong as intelligence increases, new implications for the prediction of job performance emerge: at high levels of g, relevant specific abilities may come into their own as predictors.(3)
Work over the past decade on validity generalization suggests that a single predictor - intelligence - is associated with performance in a variety of different jobs (Hunter, 1986). This work has been undertaken in large sample studies that included both civilian (Pearlman, Schmidt, & Hunter, 1980) and military occupations (McHenry, Hough, Toquam, Hanson, & Ashworth, 1990; Olea & Ree, 1994; Ree, Earles, & Teachout, 1994).
There are two ways of exploring differential prediction: an additive model and an interactive multiplicative model. In the additive model, performance would be viewed as a consequence of g and specific abilities (ss).
In the interactive model, we would expect that g would predict performance when g is low while the ss would predict performance when g is high. This implies that performance is a function of g*ss. So the declining positive manifold implies an interactive rather than an additive differential prediction model.
There has been some testing of the joint impact of general and specific ability using the additive model. Ree and his associates have been notable in their attempt to examine psychometric g(4) as a predictor; others (Pearlman et al., 1980; McHenry et al., 1990) have used specific test clusters from well-validated tests. Both groups of researchers find only a small increment in predictive validity for both training scores and performance ratings (r = .02)(5) when specific ability is added to an equation containing general ability. As these researchers use an additive model, it may not be surprising that the incremental validity is low. The appropriate model, implied by the declining positive manifold, is something quite different: an interactive model in which g is both a predictor of performance and a moderator of the relationship between a specific ability and performance. We expect that the best predictor will be the set: g, task relevant specific ability, and the product of g and the specific ability. When g is low, we would not expect the specific ability(ies) to contribute to predictability, when g is high, we would expect the specific abilities to have additional predictive power.
This is what is needed for further advance in this area: critical tests of the g only model and of the two incremental validity models, additive and interactive. This will require:
Abelson, A. R. (1911). The measurement of mental ability of "backward" children. British Journal of Psychology, 4, 269-314.
Busemeyer, J. R., & Jones, L. E. (1983). Analysis of multiplicative combination rules when the causal variables are measured with error. Psychological Bulletin, 93, 549-562.
Deary, I. J., & Pagliari, C. (1991). The strength of g at different levels of ability: Have Detterman and Daniel rediscovered Spearman's "Law of Diminishing Returns"? Intelligence, 15, 247-250.
Deary, I. J., Egan, V., Gibson, G. J., Austin, E. J., Brand, C. R., & Kellaghan, T. (1996). Intelligence and the differentiation hypothesis. Intelligence, 23, 105-132.
Detterman, D. K., & Daniel, M. H. (1989). Correlates of mental tests with each other and with cognitive variables are highest for low IQ groups. Intelligence, 13, 349-359.
Evans, M. G. (1991). The problem of analyzing multiplicative composites: Interactions revisited. American Psychologist, 46, 6-15.
Evans, M. G. (1999). On the asymmetry of g. Psychological Reports, 85, 1059-1069.
Gardner, H. (1983). Frames of Mind: The Theory of Multiple Intelligences. New York, NY: Basic Books.
Hunter, J. E. (1986). Cognitive ability, cognitive aptitudes, job knowledge, and job performance. Journal of Vocational Behavior, 29, 340-362.
James, L. R., Demaree, R. G., Mulaik, S. A., & Ladd, R. T. (1992). Validity generalization in the context of situational models. Journal of Applied Psychology, 73, 3-14.
Jensen, A. (1998). The g factor: The science of mental ability. Westport, CT: Praeger.
Jones, G. E., & Ree, M. J. (1998). Aptitude test score validity: No moderating effect due to job ability requirement differences. Educational and Psychological Measurement, 58, 284-294.
Legree, P. J., Pifer, M. E., & Grafton, F. C. (1996). Correlations among cognitive abilities are lower for higher ability groups. Intelligence, 23, 45-57.
Lubinski, D., & Humphreys, L. G. (1990). Assessing spurious "moderator effects": Illustrated substantively with the hypothesized ("synergistic") relation between spatial and mathematical ability. Psychological Bulletin, 107, 385-393.
Lynn, R. (1992). Does Spearman's g decline at high IQ levels? Some evidence from Scotland. Journal of Genetic Psychology, 153, 229-230.
McHenry, J. J., Hough, L. M., Toquam, J. L., Hanson, M. A., & Ashworth, S. (1990). Project A validity results: The relationship between predictor and criterion domains. Personnel Psychology, 43, 335-354.
Olea, M. M., & Ree, M. J. (1994). Predicting pilot and navigator criteria: Not much more than g. Journal of Applied Psychology, 79, 845-851.
Pearlman, K., Schmidt, F. L., & Hunter, J. E. (1980). Validity generalization results for tests used to predict job proficiency and training success in clerical occupations. Journal of Applied Psychology, 65, 373-406.
Ree, M. J., Earles, J. A., & Teachout, M. S. (1994). Predicting job performance: Not much more than g. Journal of Applied Psychology, 79, 518-524.
Saunders, D. R. (1956). Moderator variables in prediction. Educational and Psychological Measurement, 16, 209-222.
Spearman, C. E. (1927). The Abilities of Man. London, UK: Macmillan.
Zedeck, S. (1971). Problems with the use of moderator variables. Psychological Bulletin, 76, 295-310.
Table 1: Studies of the Asymmetry of g | ||||||||||||||
Investigator | Date | Test | Classification | Gender | Sample Size | Correlation | Sample | Correlation | Sample | Correlation | Sample | Correlation | Sample | Correlation |
Measure | hi | Hi | Upper | Upper | Middle | Middle | Lower | Lower | Lowest | Lowest | ||||
Evans | 1999 | ASVAB | Electronic Scale | Male | 214 | 0.54 | 214 | 0.75 | ||||||
Electronic | Female | 214 | 0.39 | 214 | 0.85 | |||||||||
mixed set of tests | both | 160 | 0.59 | 160 | 0.71 | |||||||||
Otis | both | 119 | 0.35 | 119 | 0.51 | |||||||||
Legree, Piper, Grafton | 1996 | ASVAB | Coding Speed | both | 1834.6 | 0.4 | 1834.6 | 0.45 | 1834.6 | 0.5 | 1834.6 | 0.51 | 1834.6 | 0.6 |
Auto & Shop Information | both | 1834.6 | 0.42 | 1834.6 | 0.41 | 1834.6 | 0.46 | 1834.6 | 0.51 | 1834.6 | 0.5 | |||
Electronics | both | 1834.6 | 0.3 | 1834.6 | 0.31 | 1834.6 | 0.36 | 1834.6 | 0.41 | 1834.6 | 0.42 | |||
Mechanical | both | 1834.6 | 0.3 | 1834.6 | 0.3 | 1834.6 | 0.36 | 1834.6 | 0.45 | 1834.6 | 0.49 | |||
Math Knowledge | both | 1834.6 | 0.24 | 1834.6 | 0.24 | 1834.6 | 0.29 | 1834.6 | 0.4 | 1834.6 | 0.41 | |||
Arithmetic Reasoning | both | 1834.6 | 0.23 | 1834.6 | 0.2 | 1834.6 | 0.25 | 1834.6 | 0.36 | 1834.6 | 0.41 | |||
General Science | both | 1834.6 | 0.24 | 1834.6 | 0.29 | 1834.6 | 0.35 | 1834.6 | 0.36 | 1834.6 | 0.35 | |||
Numerical Operations | both | 1834.6 | 0.41 | 1834.6 | 0.44 | 1834.6 | 0.45 | 1834.6 | 0.39 | 1834.6 | 0.38 | |||
Paragraph Comprehension | both | 1834.6 | 0.36 | 1834.6 | 0.36 | 1834.6 | 0.36 | 1834.6 | 0.38 | 1834.6 | 0.36 | |||
Word Knowledge | both | 1834.6 | 0.32 | 1834.6 | 0.29 | 1834.6 | 0.25 | 1834.6 | 0.23 | 1834.6 | 0.3 | |||
Deary, Egan, Gibson, Austin, Brand, Kellaghan | 1996 | DAT-T (UK) | Verbal Reasoning | both-young | 353 | 0.269 | 382 | 0.355 | ||||||
Verbal Reasoning | both-old | 366 | 0.243 | . | 370 | 0.349 | ||||||||
Numerical Ability | both-young | 739 | 0.309 | 70 | 0.39 | |||||||||
Numerical Ability | both-old | 950 | 0.311 | 1106 | 0.355 | |||||||||
Abstract Reasoning | both-young | 207 | 0.406 | 649 | 0.347 | |||||||||
Abstract Reasoning | both-old | 96 | 0.425 | 515 | 0.365 | |||||||||
Clerical Speed & Accuracy | both-young | 298 | 0.574 | 389 | 0.55 | |||||||||
Clerical Speed & Accuracy | both-old | 385 | 0.499 | 399 | 0.504 | |||||||||
Mechanical Reasoning | both-young | 319 | 0.431 | 343 | 0.437 | |||||||||
Mechanical Reasoning | both-old | 337 | 0.442 | 348 | 0.447 | |||||||||
Space Relations | both-young | 238 | 0.366 | 253 | 0.446 | |||||||||
Space Relations | both-old | 237 | 0.379 | 252 | 0.454 | |||||||||
Spelling | both-young | 311 | 0.409 | 368 | 0.416 | |||||||||
Spelling | both-old | 321 | 0.471 | 101 | 0.457 | |||||||||
Language Usage | both-young | 288 | 0.365 | 317 | 0.373 | |||||||||
Language Usage | both-old | 308 | 0.356 | 312 | 0.392 | |||||||||
Determan & Daniel | 1989 | Wais-R | Information | both | 111 | 0.41 | 514 | 0.38 | 697 | 0.52 | 466 | 0.6 | 120 | 0.82 |
vocabulary | both | 128 | 0.34 | 472 | 0.38 | 697 | 0.53 | 474 | 0.7 | 109 | 0.79 | |||
WISC-R | information | both | 168 | 0.28 | 535 | 0.3 | 842 | 0.39 | 525 | 0.48 | 130 | 0.55 | ||
vocabulary | both | 164 | 0.32 | 525 | 0.3 | 837 | 0.4 | 518 | 0.36 | 156 | 0.55 | |||
Lynn | 1992 | WISC-R Scotts | vocabulary | both | 270 | 0.2 | 270 | 0.14 | 270 | 0.17 | 270 | 0.38 | 270 | 0.44 |
Abelson (in Spearman) | 1927 | ? | ? | both | 78 | 0.47 | 22 | 0.782 | ||||||
1. Note that the within cell variances were higher than the variance due to sampling error. This is probably a consequence of a) the extremity of the selection, b) the measures used, and c) the classification scales used.
2. One of the highly g-loaded scales in Deary et al. (1996), Analytic Reasoning, showed an "increasing returns" effect. When this was omitted from the high g analysis, the results were no different.
3. My expectation is that this is true when jobs requiring a high level of ability are involved.
4. This is assessed as the first principal component of a factor analysis
5. The use of the components of a principal axis analysis ensures that the general factor is extracted from all scales and that all components extracted are orthogonal. This ensures that subsequent analysis adds unique variance in a regression equation. The use of real tests ensures that meaningful subsets of skill and ability are used, when g is extracted, subsequent components do not seem to reflect particular bundles of competencies; though the use of a Vernon-like hierarchical structure would enable both g and meaningful components to be used - no analyses used this technique.
6. By this I mean the level of ability required, not the kind of ability required which is what is addressed in Jones and Ree (1998)