Role of Mental Abilities and Conscientiousness in Explaining College Grades

The prediction of college academic performance, often measured using grade point averages (GPAs), continues to be a research topic in applied psychology (e.g., Biderman et al., 2018; Dahlke et al., 2019; Moore & Ones, 2019). Applied psychologists are involved in developing and validating academic testing programs (e.g., MCAT, SAT, ACT) and those in academia have grants to study predictors of academic performance (Sackett et al., 2008; Schmitt et al., 2011) and publish related research (e.g., Kuncel et al., 2001, 2007; Kuncel & Hezlett, 2007).

Research on predictors of job and training performance indicated that g is primarily responsible for the predictive power of standardized tests of mental abilities, with specific abilities often adding little or nothing to prediction of performance (e.g., Brown et al., 2006; Ree et al., 1994; Ree & Carretta, 2022). However, there have been recent calls to reexamine the importance of specific abilities in personnel selection, including a series of articles in a special issue of the Journal of Personnel Psychology (Krumm et al., 2014) and an article by Murphy (2017), which suggested the conclusion that specific abilities add little value was “premature.” Indeed, a study by Nye et al. (2022) provided evidence that the incremental validity of specific abilities over g is greater when for those specific abilities with correlations with g and Briet et al. (2024) reported higher incremental validity for specific abilities that are aligned with a criterion.

In contrast to the large body of research on the role of g vs. specific abilities in predicting job/training performance, there is considerably less research on this topic using academic performance. Much of the large-sample research focuses on the validity of specific operational tests (e.g., the SAT, the ACT), which include only a small number of subtests. Jensen (1998) noted the lack of research, stating “Surprisingly little of this applied literature on test validity, however, examines the degree to which g itself, as compared to other factors and specificity, contributes to tests’ validity. Fortunately, the few studies that focus specifically on this question have been conducted by the armed forces and by the U.S. Employment Service of the Department of Labor [i.e., Hunter & Schmidt]. These studies, based on huge samples, are technically excellent” (p. 271). More recently, Peng and Goodrich (2020) noted that g strongly predicts academic success with a validity coefficient of .42, reinforcing the dominance of general cognitive ability over specific abilities in academic performance.

One exception is a study by Cucina et al. (2016) who examined the criterion-related validity of g vs. specific abilities in predicting high school academic performance using two datasets. In the first dataset, they used the Armed Forces Vocational Aptitude Battery (ASVAB) to predict high school GPAs (obtained using transcripts) and found that much of the criterion-related validity was due to g. In the second dataset, they used mental abilities test scores from Project TALENT, a large nationwide study conducted by I/O psychologist John Flanagan and associates (1964). Self-reported GPA served as the criterion and big five personality scale scores served as a secondary predictor. Again, much of the criterion-related validity was due to g (specific abilities often lacked incremental validity). One exception that occurred in both studies was that mathematical knowledge added incremental validity over g.

There is strong evidence that conscientiousness adds predictive validity. A meta-analysis by Mammadov (2022) found that cognitive ability and conscientiousness were both strong predictors of multiple levels (including undergraduate) of academic performance and that conscientiousness added incremental validity. Another meta-analysis by Vedel (2014) reported that among the big five, conscientiousness was the strongest predictor of GPA. Friedrich and Schütz (2023) found conscientiousness and intelligence independently predict grades in 3,775 German 13th-graders, with no interaction, supporting independent effects and rejection of a multiplicative model. Lechner et al. (2021) found g and conscientiousness independently predict achievement in 18,637 German students, with weak math synergy. Wang and Zhang (2021) showed cognitive ability and self-control, akin to conscientiousness, independently predicted high school performance without interaction, suggesting applicability to college settings.

The Role of High School (HS) GPA in Predicting College GPA

The use of standardized testing for college admissions remains controversial, with many critics arguing for more weight given to HS GPA or the use of HS GPA in lieu of standardized test scores. Even proponents of the SAT and ACT (e.g., Kobrin et al., 2011) suggest test scores and HS GPA should be used in combination (e.g., using a regression equation). A large study by Mattern and Patterson (2013) indicated that HS GPA has an observed validity of .370 (compared to .367 for the SAT) and adds incremental validity (multiple R = .468) using Freshman GPA as the criterion. Cucina et al. (2016) found that HS GPA correlated .32 to .40 with g and .32 with conscientiousness.

Based on the overlap and intercorrelations between g, HS GPA, and college GPA, it is important to include HS GPA in any studies examining the contribution of g vs. specific abilities to predicting college GPA. Additionally, it might be possible that HS GPA has a mediatory role in the relationship between the individual differences of g and conscientiousness with college GPA. These individual differences may lead to individual differences in HS GPA, which in turn lead to individual differences in college GPA. Harackiewicz et al. (2002) reported that semester grades in psychology courses mediated the relationship between ability, goals, and high school performance in predicting grades later in student’s undergraduate careers. Research by Radunzel and Noble (2012) and Sanchez (2024) indicates that first-year college grades play a mediating role in the prediction of college degree completion and cumulative college grades by admissions predictors (e.g., ACT scores and high school grades).

Overview of Study and Hypotheses

In this study, we extended the work of Cucina et al. (2016) who examined the validity of mental ability tests and conscientiousness in predicting HS GPA using the Project TALENT dataset. Their work focused on the base year data collection of Project TALENT, which included over 300,000 participants (American Institutes for Research, 1960a/b). A five-year follow-up study was later conducted and included self-reported college GPA, which was not included in either the dataset or used by Cucina et al. (2016).

We examined the criterion-related validity of a large battery of mental abilities tests, HS GPA, and a measure of conscientiousness in the prediction of college GPA. Past research suggests that g, not s, predicts job, training, and performance, thus we predict the same finding for our study (Hypothesis 1), with one exception noted below. We hypothesized that conscientiousness will predict college GPA (Hypothesis 2), based on past meta-analyses (e.g., McAbee & Oswald, 2013). Our third hypothesis is that conscientiousness and g will interact when predicting performance, following the classic equation (Maier, 1955; Vroom, 1960, 1964):

Performance = motivation × ability

Our fourth hypothesis is that mathematical knowledge will add incremental validity over g, based on research by Cucina et al. (2016); this is the exception to Hypothesis 1. Hypothesis 5 states that HS GPA will mediate the prediction of college GPA from g and conscientiousness.

We used the dataset for the Project TALENT study; although it was conducted some time ago, it makes up for its age by virtue of the sheer size of the dataset and the number of tests administered. Project TALENT has data on over 300,000 high school students on a battery of 59 mental abilities tests as well as personality tests and other measures (e.g., interests, background questionnaires). Of these students, 35,446 participated in the five-year follow-up study and obtained a Bachelor’s degree, and self-reported college grades.

This dataset has been used in a large number of studies, primarily conducted in the 1960s-1970s (e.g., Cureton, 1968; see Campbell, 1979); however, it continues to be used in research today (e.g., Arneson et al., 2011; Major et al., 2012; Reeve, 2001, 2004; Reeve et al., 2006; Waters, 2007). Despite a current bibliography of over 356 publications, we were unable to locate a study that investigated the relationship between the 59 mental abilities tests, conscientiousness, HS grades, and college GPA. One exception is the study by Arneson et al. (2011) who investigated the linearity of the relationship between a composite cognitive measure from the dataset and college GPA for 15,040 participants. As the focus of their study was on the linearity of test-performance relationships, they did not test the central hypotheses in our study.

Method

We used the base- and five-year follow-up datasets from Project TALENT (American Institutes for Research, 1960; Flanagan, 1973; Flanagan et al., 1961, 1964, 1971a, 1971b; Flanagan & Jung, 1971; Wise et al., 1979). The Project TALENT researchers provided a credibility index for each participant and suggested that cases with an unacceptable index should be removed before conducting analyses, which we did. We removed cases with missing data on one or more of the abilities. It was necessary to remove these cases because scores on all tests were needed to compute variables for g. Replacing missing values using expectation-maximization did not have a substantive impact on the correlations between the variables of interest; therefore, we decided against replacing missing values. Finally, we removed participants who had not obtained a Bachelor’s degree and that lacked college GPA data from the five-year follow-up study. Table 1 presents the demographics for the cases in our final dataset.

HS GPA. In addition to completing the mental abilities tests, participants also completed a “Student Information Blank” which contained background questions. Eight questions concerned HS grades and the Project TALENT researchers created a weighted composite using seven of these questions. The items and weights are provided in Table 2 and were obtained from pages 36 and R-4 of Wise et al. (1960). The Project TALENT participants collectively attended 1,214 schools, and there were 1,095 schools that had 30 or more cases in the data. We also converted HS grades (for schools with 30 or more cases) to z-scores to address the possibility that different schools might have different grading standards.

Table 1.Demographic Statistics.
Variable Frequency Percent
Gender
Male 18,979 52.8
Female16,95347.2
Race/Ethnicitya
White (non-Hispanic) 34,073 94.8
Black (non-Hispanic) 666 1.9
Asian 331 0.9
Native American 10 <.1
Missing/Other 852 2.4
Grade
Ninth 8,508 23.7
Tenth 9,232 25.7
Eleventh 9,442 26.3
Twelfth 8,750 24.4
Mean SD
Ageb 15.64 1.22
Missing 453 1.3%
Total 35,932

Notes: aWe use modern terminology to describe the races/ethnicities. When Project TALENT was begun (in the 1960s) other terms were used for these groups. bAge was missing for 453 cases (1.3% of the sample).

College GPA. The five-year follow-up survey included the following question: “As an undergraduate what is or was your average grade in your major subject and all subjects?” Participants indicated the letter grade corresponding to the average grades in their major and overall. We used the overall GPA as the criterion. We recoded the letter grades into a traditional 4-point system (i.e., A=4; A-=3.7; B+=3.3; B=3; B-=2.7; C+=2.3; C=2; C-=1.7; D+=1.3; D=1; D-=.7). The use of self-reported grades is a limitation; however, Kuncel et al. (2005) found that self-reported HS grades have an observed correlation of .82 with grades from school records and college grades had a correlation of .90. They stated that self-reported grades “generally predict outcomes to a similar extent as actual grades” (p. 76).

Table 2.Items, Options, and Weights for HS Grades Criterion Variable.
Item # Text Item Weight
106 My gradesa in mathematics have been 2
107 My grades in science courses have been 2
108 My grades in foreign languages have been 2
109 My grades in history and social studies courses have been 2
110 My grades in English courses have been 2
111 My grades in vocational courses have been 0
112 My grades in business or commercial courses have been 1
113 My grades in all courses starting with the night grade have been 1
Option Text Option Weight
A All A’s or equivalentb 50
B Mostly A’s or equivalent 40
C Mostly A’s and B’s or equivalent 30
D Mostly B’s and C’s or equivalent 20
E Mostly C’s and D’s or equivalent 10
F Mostly D’s or below or equivalent 0

Notes: Items, options, and weights obtained from pages 36 and R-4 of Wise, McLaughlin, & Steel (1960). aParticipants were instructed to only report semester grades and to leave items blank if they did not complete any courses in a particular area. bThe term “equivalent” was defined as follows: A = “Excellent; 90-100”; B = “Good; 80-89”; C = “Average; 70-79”; D = “Fair; 60-69”; below D = “Failing; 59 or lower.”

Conscientiousness. Ten personality scales were administered with the Project TALENT mental abilities tests. Similar to Cucina et al. (2016), we used Reeve et al.'s (2006) work to compute big five scale scores.

Mental Abilities Tests. In Table 3, we present the names/abbreviations for all tests. Each test was rationally linked to Carroll’s (1993) three-stratum-theory. Major et al. (2012) independently linked a subset of the tests to Carroll’s abilities. Both linkages were used and are shown in Table 3. In general, the first half of the tests measure crystallized intelligence and Flanagan et al. (1962) included these “Information” tests as measures of general knowledge (especially knowledge gained through general life rather than formal schooling) and these fall squarely within Carroll’s (1993, p. 624) definition of crystallized intelligence (i.e., “mental processes that reflect… the effects of experience, learning, and acculturation”). The remaining tests measure specific abilities such as verbal, quantitative, spatial, perceptual speed, and memory. An analogue to the Raven’s Progressive Matrices (Raven et al., 2004), Test 290 Abstract Reasoning, was also administered.

Table 3.Names, Abbreviations, and Factors for Mental Abilities Tests.
Ability Abbreviation Carroll Broad Factor Carroll Narrow Factor
1 2 3 1 2 3
Information Tests
Vocabulary T102VOCA 2C V VL
Literature T103LITE 2C V K2
Music T104MUSI 2C V K2
Social Studies T105SOCI 2C V K2
Mathematics T106MATH 2C KM N A3
Physical Science T107PHYS 2C V A7
Biological Science T108BIOL 2C V A7
Scientific Attitude T109SCIE 2F 2R RG/I SP
Aeronautics and Space T110AERO 2C V A7
Electricity and Electronics T111ELEC 2C V MK A7
Mechanical T112MECH 2C V MK
Farming T113FARM 2C V MK
Home Economics T114HOME 2C V K2
Sports T115SPOR 2C V K2
Art T131ART 2C V K2
Law T132LAW 2C V K2
Health T133HEAL 2C V A7
Engineering T134ENGI 2C V A7
Architecture T135ARCH 2C V K2
Journalism T136JOUR 2C V K2
Foreign travel T137FORE 2C V K2
Military T138MILI 2C V K2
Accounting, Business, Sales T139ACCO 2C V K2
Practical Knowledge T140PRAC 2C V K2
Clerical T141CLER 2C V K2
Bible T142BIBL 2C V K2
Colors T143COLO 2C V K2
Etiquette T144ETIQ 2C V K2
Hunting T145HUNT 2C V K2
Fishing T146FISH 2C V K2
Other Outdoor Activities T147OTHE 2C V K2
Photography T148PHOT 2C V K2
Games (Sedentary) T149GAME 2C V K2
Theater and Ballet T150THEA 2C V K2
Food T151FOOD 2C V K2
Miscellaneous Test Score T152MISC 2C V
Non-Information Tests
Vocabulary T162VOCA 2C V VL
Memory for Sentences T211MEMO 2Y MA
Memory for Words T212MEMO 2Y MA
Disguised Words T220DISG 2R 2F 2C FW RG/I V
Spelling T231SPEL 2C SG V
Capitalization T232CAPI 2C MY SG V
Punctuation T233PUNC 2C MY V
English Usage T234ENGL 2C MY V
Effective Expression T235EFFE 2C MY V
Word Functions in Sentences T240WORD 2C MY V
Reading Comprehension T250READ 2C RC V
Creativity T260CREA 2R FI FE
Mechanical Reasoning T270MECH 2V VZ MK
Visualization in 2 Dimensions T281VISU 2V VZ SR
Visualization in 3 Dimensions T282VISU 2V VZ
Abstract Reasoning T290ABST 2F RG/I
Math I. Arithmetic Reasoning T311MTH1 2F RQ
Math II. Introductory HS Mathematics T312MTH2 2C KM N A3
Math III. Advanced HS Mathematics T333MTHe 2C KM N A3
Arithmetic Computation (R - 3W) T410ARIT 2S N
Table Reading (R - W) T420TABL 2S P
Clerical Checking (R - 3W) T430CLER 2S P
Object Inspection (R - W) T440OBJE 2S P

Notes: Carroll’s (1993) Broad abilities are as follows: 2C = Crystallized Intelligence; 2F = Fluid Intelligence; 2R = Broad Retrieval Ability; 2Y = General Memory & Learning; 2V = Broad Visual Perception; 2S = Broad Cognitive Speediness. Carroll’s Narrow Abilities are as follows: V = Verbal (Printed) Language Comprehension; VL = Lexical Knowledge (Vocabulary); K2 = Cultural Information (e.g., art, history; see Carroll, page 521); KM = Knowledge of Mathematics (see page 523-4); N = Numerical Facility; A3 = Tested Math Achievement (see page 532); A7 = Tested Science Achievement (see page 524); RG/I = General Reasoning Sequential Reasoning/Induction (these were combined per Colberg, 1985; Colberg et al., 1985); SP = Sensitivity to Problems; MK = Mechanical Knowledge (see page 525); MA = Associative Memory; FW = Word Fluency; SG = Spelling Ability; MY = Grammatical Sensitivity; RC = Reading Comprehension; FI = Ideation Fluency; FE = Expressional Fluency; VZ = Visualization; SR = Spatial Relations; RQ = Quantitative Reasoning; P = Perceptual Speed.

Results

Descriptive statistics and reliabilities for the research variables are provided in Table 4. To address Hypothesis 1, we examined the criterion-related validity of the tests. These analyses are summarized in Table 5.

We computed the bivariate observed correlation (robs) between each test and college GPA. Using the final restricted raw dataset, we conducted a Principal Component Analysis (PCA) on the tests and computed scores on the first unrotated principal component to serve as an estimate of g (see Jensen & Weng, 1994, for a justification). Two types of g scores were computed, one using all 59 tests, and the other omitting the Information tests. We also computed the operational validity (ρov), which corrects for criterion unreliability and multivariate range restriction (using the full sample of over 300,000 cases to estimate unrestricted variance-covariance matrix) and the true score validity (ρTS), which corrects for multivariate range restriction and predictor and criterion unreliability (using the full sample to estimate reliabilities). A PCA was also conducted on the full unrestricted Project TALENT dataset (which also included data from individuals that did not attend college). The scoring weights for both types of g scores were obtained and later used with the multivariate corrected correlation matrix to compute composite g and personality scores.

Table 4.Descriptive statistics, reliabilities, and g-loadings for cognitive tests.
Restricted Validation Sample Unrestricted Sample
Descriptive Statistics g-loading Descriptive Statistics g-loading
Ability N M SD Reliability Using All tests Using Non-Info tests N M SD Reliability Using All tests Using Non-Info tests
T102VOCA 35,932 15.62 3.10 .50 .81 .66 321,589 12.23 4.06 .71 .86 .74
T103LITE 35,932 16.11 4.09 .64 .74 .60 321,589 12.35 4.54 .71 .80 .69
T104MUSI 35,932 8.56 2.62 .58 .60 .50 321,589 6.26 2.94 .67 .71 .62
T105SOCI 35,932 18.87 3.79 .64 .70 .52 321,589 14.38 5.24 .81 .81 .69
T106MATH 35,932 13.71 5.03 .78 .77 .71 321,589 8.45 5.17 .79 .77 .72
T107PHYS 35,932 11.61 3.73 .69 .69 .54 321,589 8.32 4.07 .74 .73 .61
T108BIOL 35,932 7.41 2.09 .40 .60 .45 321,589 5.90 2.38 .54 .67 .55
T109SCIE 35,932 7.11 1.59 .10 .53 .47 321,589 5.95 1.93 .39 .65 .59
T110AERO 35,932 4.88 2.52 .57 .54 .33 321,589 3.71 2.30 .48 .56 .41
T111ELEC 35,932 9.28 4.47 .67 .53 .35 321,589 7.31 4.03 .59 .54 .40
T112MECH 35,932 11.21 3.72 .52 .52 .33 321,589 9.79 3.93 .57 .54 .39
T113FARM 35,932 8.55 2.06 .41 .45 .34 321,589 7.44 2.48 .59 .58 .48
T114HOME 35,932 11.52 3.64 .49 .29 .31 321,589 10.43 3.67 .50 .40 .41
T115SPOR 35,932 8.66 2.85 .51 .46 .30 321,589 6.77 3.01 .56 .59 .47
T131ART 35,932 7.90 2.21 .53 .61 .46 321,589 6.21 2.56 .65 .73 .62
T132LAW 35,932 5.69 1.72 .41 .62 .44 321,589 4.58 1.81 .47 .67 .54
T133HEAL 35,932 6.99 1.52 .24 .50 .39 321,589 5.85 2.00 .56 .67 .58
T134ENGI 35,932 3.46 1.19 .17 .46 .32 321,589 2.89 1.31 .32 .55 .43
T135ARCH 35,932 3.33 1.34 .35 .51 .38 321,589 2.60 1.33 .34 .55 .45
T136JOUR 35,932 2.31 0.79 .09 .53 .40 321,589 1.70 0.98 .41 .64 .54
T137FORE 35,932 3.39 1.21 .34 .61 .43 321,589 2.52 1.35 .47 .66 .52
T138MILI 35,932 2.95 1.42 .34 .48 .30 321,589 2.24 1.41 .33 .53 .40
T139ACCO 35,932 5.75 1.82 .42 .61 .47 321,589 4.55 1.96 .50 .67 .57
T140PRAC 35,932 3.48 0.68 .01 .41 .31 321,589 3.09 0.91 .39 .57 .49
T141CLER 35,932 2.14 0.83 .13 .38 .33 321,589 1.85 0.92 .29 .43 .40
T142BIBL 35,932 9.08 3.21 .72 .48 .37 321,589 7.02 3.26 .73 .60 .52
T143COLO 35,932 1.72 0.96 .28 .31 .32 321,589 1.40 0.94 .25 .38 .37
T144ETIQ 35,932 1.29 0.69 .01 .24 .25 321,589 1.01 0.72 .10 .35 .35
T145HUNT 35,932 1.66 1.25 .25 .19 .07 321,589 1.62 1.23 .23 .20 .10
T146FISH 35,932 1.52 1.24 .32 .21 .10 321,589 1.38 1.16 .23 .24 .14
T147OTHE 35,932 5.72 1.66 .32 .50 .36 321,589 4.61 1.88 .47 .64 .52
T148PHOT 35,932 1.57 0.64 .01 .36 .28 321,589 1.30 0.71 .16 .48 .42
T149GAME 35,932 2.77 1.17 .34 .40 .29 321,589 2.23 1.12 .28 .46 .38
T150THEA 35,932 5.40 1.66 .50 .59 .48 321,589 4.22 1.79 .57 .68 .60
T151FOOD 35,932 1.83 1.07 .42 .41 .32 321,589 1.32 1.04 .39 .48 .40
T152MISC 35,932 5.81 1.55 .29 .61 .46 321,589 4.65 1.76 .45 .69 .57
T162VOCA 35,932 7.35 1.53 .28 .68 .54 321,589 5.86 2.18 .64 .79 .68
T211MEMO 35,932 9.90 2.93 .60 .18 .28 321,589 9.11 3.04 .63 .29 .36
T212MEMO 35,932 14.59 5.41 .82 .37 .48 321,589 11.61 5.38 .82 .50 .57
T220DISG 35,932 18.82 6.31 .84 .54 .62 321,589 15.05 6.92 .87 .65 .69
T231SPEL 35,932 11.05 2.69 .48 .44 .55 321,589 9.36 3.00 .58 .54 .62
T232CAPI 35,932 31.00 2.32 .57 .30 .40 321,589 29.44 3.79 .84 .48 .57
T233PUNC 35,932 21.09 3.58 .56 .57 .70 321,589 17.64 4.55 .73 .70 .79
T234ENGL 35,932 18.65 2.69 .28 .51 .60 321,589 16.66 3.29 .52 .62 .69
T235EFFE 35,932 9.80 1.76 .25 .45 .51 321,589 8.58 2.33 .57 .57 .63
T240WORD 35,932 14.83 5.24 .82 .58 .71 321,589 10.16 5.46 .83 .67 .75
T250READ 35,932 38.73 7.18 .67 .76 .74 321,589 30.04 10.57 .85 .85 .84
T260CREA 35,932 11.16 3.71 .66 .65 .63 321,589 8.69 3.96 .70 .71 .69
T270MECH 35,932 12.50 4.06 .66 .56 .50 321,589 10.39 4.30 .70 .60 .56
T281VISU 35,932 14.59 5.19 .75 .37 .44 321,589 12.68 5.75 .80 .43 .48
T282VISU 35,932 10.04 3.12 .61 .51 .56 321,589 8.46 3.26 .64 .56 .60
T290ABST 35,932 10.71 2.27 .38 .50 .60 321,589 8.85 3.03 .65 .64 .71
T311MTH1 35,932 11.14 3.03 .62 .67 .70 321,589 8.29 3.54 .72 .74 .76
T312MTH2 35,932 15.15 4.55 .73 .70 .75 321,589 10.47 4.76 .75 .74 .77
T333MTHe 35,932 4.79 2.92 .66 .61 .63 321,589 3.10 2.30 .45 .52 .53
T410ARIT 35,932 38.13 14.19 .67 .33 .45 321,589 28.26 21.06 .85 .46 .54
T420TABL 35,932 11.08 7.01 .68 .13 .22 321,589 9.55 10.31 .85 .25 .33
T430CLER 35,932 30.17 16.43 .77 .21 .33 321,589 25.22 20.45 .85 .28 .38
T440OBJE 35,932 22.85 7.01 .80 .18 .28 321,589 21.52 8.11 .85 .26 .35
g-all tests 35,932 0.97 0.77 .97c 1a 1a 321,589 0.00 1.00 .98c 1a (.92)b
g-non-Info tests 35,932 0.93 0.77 .95c (.88)b (.88)b 321,589 0.00 1.00 .97c (.92)b 1a
HS GPA 34,496 32.98 9.60 .85d .29 .36 307,302 25.54 10.04 .86d .32 .35
Std HS GPA 32,244 0.76 0.97 .85d .34 .41 305,167 0.00 1.00 .86d .37 .40
Neuroticism 35,932 0.53 1.77 .87e .15 .14 321,427 0.00 1.69 .86e .25 .23
Extroversion 35,932 0.77 3.56 .94e .05 .04 321,427 0.00 3.33 .93e .18 .16
Openness 35,932 0.28 1.05 .72e .06 .08 321,427 0.00 1.00 .69e .15 .17
Agreeable. 35,932 0.20 1.01 .79e .04 .06 321,427 0.00 1.00 .79e .17 .19
Conscient. 35,932 0.64 1.86 .93e .09 .12 321,427 0.00 1.79 .92e .20 .22

Note. n = 35,932. aSince the g-loadings were computed by correlating each test score with the g-scores (obtained from principal components analysis), the g-loading for the g-score with itself is 1.0. bThe correlation between the two g-scores is placed parenthetically in this column. cThe reliability estimate for g was obtained using the reliabilities for each subtest, the weights used to compute the first unrotated principal component (i.e., the weights used to compute the estimate of g), and Nunnally and Bernstein’s (1994) formula 7-18 for the reliability of the sum of a weighted linear combination. dTo compute the reliability of HS grades, we used the weights for the items measuring HS grades and the weights for the response options for those items to compute seven variables that were entered into a Coefficient Alpha analysis using SPSS. The seven variables predicted the HS grades with an R of 1.00 and the resulting Coefficient Alpha was .860, which was used for both the unstandardized and standardized HS grades variables. When analyzed separately by grade, the Coefficient alphas were as follows: ninth grade: .845; tenth grade: .855; eleventh grade: .865 ; twelfth grade: .871. We should note these are Coefficient Alphas for the composite of the self-reported grades and do not necessarily reflect the reliability of the true HS GPA that would be obtained from transcripts. eTo obtain the reliability for the conscientiousness score, we computed the reliability of a unit-weighted composite of standardized scores (Nunnally & Bernstein, 1994, formula 7- 12, p. 268) for the Tidiness and Maturity scales using the reliability estimates for these scales that were obtained by Reeve et al (2006).

Table 5.Incremental criterion-related validity of mental abilities tests for College GPA controlling for g score from non-Information tests.
Bivariate Incremental validity over g Our linkage Major et al.’s Linkage
Ability robs p ρOV ρTS robs·g p βobs p ΔRobs βOV ΔROV βTS ΔRTS SEM rSEM-implied rSEM-implied - robs SEM rSEM-implied rSEM-implied - robs
path p path p
g-all tests .27 <.01 .38 .38
g-no info .32 <.01 .41 .42
T102VOCA .22 <.01 .31 .37 .01 .12 .01 .11 <.01 .03 <.01 .03 <.01 -.02 <.01 .22 -.01 .22 <.01 .22 -.01
T103LITE .20 <.01 .31 .37 .01 <.05 .01 <.05 <.01 .07 <.01 .11 .01 -.01 .02 .20 -.01 .20 <.01 .22 -.02
T104MUSI .20 <.01 .29 .35 .06 <.01 .06 <.01 <.01 .07 <.01 .11 .01 .04 <.01 .16 .04 .16 <.01 .17 .03
T105SOCI .19 <.01 .30 .33 .03 <.01 .03 <.01 <.01 .06 <.01 .06 <.01 -.01 .07 .20 -.01 .20 <.01 .19 .00
T106MATH .21 <.01 .33 .37 -.03 <.01 -.04 <.01 <.01 .08 <.01 .10 <.01 -.01 .02 .23 -.03 .23 <.01 .26 -.05
T107PHYS .17 <.01 .26 .30 .00 .41 .00 .56 <.01 .01 <.01 -.01 <.01 -.03 <.01 .19 -.02 .19 <.01 .21 -.04
T108BIOL .14 <.01 .22 .30 -.01 .25 -.01 .25 <.01 .01 <.01 <.01 <.01 -.03 <.01 .17 -.03 .17 <.01 .16 -.03
T109SCIE .17 <.01 .25 .40 .03 <.01 .03 <.01 <.01 .04 <.01 .16 .01 .00 .74 .17 .00 .17
T110AERO .05 <.01 .12 .17 -.07 <.01 -.07 <.01 .01 -.06 <.01 -.12 .01 -.09 <.01 .15 -.11 .15 .99 .15 -.10
T111ELEC .05 <.01 .11 .14 -.08 <.01 -.08 <.01 .01 -.06 <.01 -.10 .01 -.10 <.01 .15 -.10 .15 .03 .13 -.09
T112MECH .00 .95 .06 .08 -.12 <.01 -.12 <.01 .02 -.10 .01 -.16 .02 -.14 <.01 .14 -.14 .14 <.01 .13 -.13
T113FARM .10 <.01 .16 .21 -.01 .06 -.01 .06 <.01 -.03 <.01 -.07 <.01 -.02 <.01 .12 -.02 .12
T114HOME .15 <.01 .20 .28 .06 <.01 .06 <.01 .01 .05 <.01 .08 .01 .06 <.01 .07 .09 .07
T115SPOR .03 <.01 .14 .19 -.07 <.01 -.07 <.01 .01 -.04 <.01 -.09 .01 -.09 <.01 .13 -.10 .13
T131ART .16 <.01 .26 .32 .02 <.01 .02 <.01 <.01 .02 <.01 .02 <.01 .00 .85 .17 -.01 .17 .93 .18 -.01
T132LAW .14 <.01 .23 .34 .00 .57 .00 .69 <.01 .02 <.01 .04 <.01 -.03 <.01 .17 -.04 .17 <.01 .16 -.02
T133HEAL .17 <.01 .27 .36 .06 <.01 .06 <.01 <.01 .06 <.01 .10 .01 .04 <.01 .13 .04 .13 .27 .14 .03
T134ENGI .04 <.01 .12 .21 -.06 <.01 -.06 <.01 .01 -.05 <.01 -.18 .02 -.08 <.01 .13 -.08 .13
T135ARCH .13 <.01 .20 .34 .01 <.05 .01 <.05 <.01 .03 <.01 .08 <.01 -.01 .32 .14 -.01 .14
T136JOUR .13 <.01 .23 .36 .00 .92 .00 .70 <.01 .03 <.01 .06 <.01 -.02 <.01 .15 -.02 .15
T137FORE .14 <.01 .21 .31 .00 .71 .00 .50 <.01 .01 <.01 -.01 <.01 -.02 <.01 .17 -.03 .17
T138MILI .05 <.01 .13 .23 -.05 <.01 -.05 <.01 <.01 -.02 <.01 -.06 <.01 -.08 <.01 .14 -.09 .14
T139ACCO .17 <.01 .26 .37 .02 .60 .02 <.01 <.01 .06 <.01 .11 .01 .00 .51 .17 .00 .17
T140PRAC .09 <.01 .18 .29 -.01 <.05 -.01 <.05 <.01 -.01 <.01 -.05 <.01 -.02 <.01 .11 -.02 .11
T141CLER .11 <.01 .18 .33 .01 .18 .01 .14 <.01 .05 <.01 .12 .01 .01 .28 .10 .01 .10
T142BIBL .19 <.01 .29 .34 .09 <.01 .09 <.01 .01 .11 .01 .14 .01 .06 <.01 .13 .06 .13 <.01 .13 .06
T143COLO .15 <.01 .18 .36 .05 <.01 .05 <.01 <.01 .04 <.01 .13 .01 .06 <.01 .07 .08 .07
T144ETIQ .11 <.01 .16 .51 .03 <.01 .03 <.01 <.01 .02 <.01 N/A N/A .03 <.01 .05 .06 .05
T145HUNT -.07 <.01 -.05 -.10 -.10 <.01 -.09 <.01 .01 -.08 .01 -.18 .04 -.11 <.01 .06 -.13 .06
T146FISH -.03 <.01 .00 .00 -.07 <.01 -.07 <.01 .01 -.05 <.01 -.11 .02 -.08 <.01 .06 -.10 .06
T147OTHE .07 <.01 .18 .26 -.04 <.01 -.04 <.01 <.01 -.04 <.01 -.12 .01 -.06 <.01 .14 -.06 .14
T148PHOT .07 <.01 .16 .40 -.01 <.05 -.01 <.01 <.01 .00 <.01 -.91 .02 -.02 <.01 .09 -.02 .09
T149GAME .07 <.01 .13 .25 -.03 <.01 -.03 <.01 <.01 -.03 <.01 -.11 .01 -.04 <.01 .11 -.04 .11
T150THEA .18 <.01 .27 .36 .04 <.01 .04 <.01 <.01 .05 <.01 .09 .01 .02 <.01 .16 .02 .16 <.01 .17 .01
T151FOOD .12 <.01 .19 .30 .02 <.01 .02 <.01 <.01 .04 <.01 .07 <.01 .00 .53 .11 .00 .11
T152MISC .15 <.01 .24 .36 .00 .56 .00 .54 <.01 .02 <.01 .02 <.01 -.02 <.01 .17 -.02 .17 <.01 .16 -.01
T162VOCA .19 <.01 .28 .35 .02 <.01 .02 <.01 .01 .02 <.01 .02 <.01 .00 .42 .19 .00 .19
T211MEMO .15 <.01 .18 .23 .07 <.01 .07 <.01 <.01 .04 <.01 .05 <.01 .05 <.01 .06 .09 .06 <.01 .07 .08
T212MEMO .20 <.01 .28 .31 .06 <.01 .07 <.01 .01 .07 .01 .08 .01 .00 .91 .13 .08 .13 <.01 .12 .08
T220DISG .20 <.01 .28 .30 .00 .13 .01 .11 <.01 .01 <.01 -.01 <.01 .02 <.01 .18 .02 .18 <.01 .17 .03
T231SPEL .23 <.01 .29 .38 .07 <.01 .08 <.01 .01 .08 .01 .16 .01 .08 <.01 .15 .08 .15 <.01 .16 .07
T232CAPI .13 <.01 .20 .22 .01 .09 .01 <.05 <.01 -.04 <.01 -.06 <.01 .00 .64 .11 .02 .11 <.01 .11 .03
T233PUNC .28 <.01 .36 .42 .09 <.01 .12 <.01 .01 .10 .01 .25 .01 .03 <.01 .20 .08 .20 .01 .20 .09
T234ENGL .20 <.01 .28 .39 .02 <.01 .02 <.01 <.01 .01 <.01 -.03 <.01 -.01 .19 .17 .04 .17 <.01 .16 .04
T235EFFE .15 <.01 .25 .33 -.01 <.05 -.01 <.05 <.01 <.01 <.01 -.03 <.01 -.03 <.01 .14 .02 .14 <.01 .13 .02
T240WORD .30 <.01 .37 .41 .11 <.01 .15 <.01 .02 .16 .01 .22 .02 .05 <.01 .19 .11 .19 <.01 .19 .10
T250READ .26 <.01 .36 .39 .04 <.01 .06 <.01 <.01 .08 <.01 .10 <.01 .01 .02 .25 .01 .25 .01 .23 .03
T260CREA .16 <.01 .25 .30 -.05 <.01 -.06 <.01 <.01 -.04 <.01 -.12 .01 -.04 <.01 .21 -.04 .21 .57 .19 -.02
T270MECH .05 <.01 .12 .14 -.14 <.01 -.15 <.01 .03 -.16 .02 -.25 .04 -.05 <.01 .16 -.12 .16 <.01 .18 -.13
T281VISU .07 <.01 .11 .12 -.08 <.01 -.09 <.01 .01 -.11 .01 -.14 .02 .00 .75 .12 -.06 .12 <.01 .12 -.05
T282VISU .11 <.01 .17 .21 -.09 <.01 -.10 <.01 .01 -.11 .01 -.22 .03 .01 .02 .16 -.05 .16 .59 .15 -.04
T290ABST .16 <.01 .25 .31 -.05 <.01 -.06 <.01 <.01 -.09 .01 -.31 .02 -.02 <.01 .18 -.02 .18 <.01 .16 .00
T311MTH1 .23 <.01 .32 .38 .02 <.01 .03 <.01 <.01 .02 <.01 <.01 <.01 .01 .04 .23 .01 .23 <.01 .21 .02
T312MTH2 .24 <.01 .34 .39 .01 <.05 .02 .05 <.01 .07 <.01 .11 <.01 .04 <.01 .22 .03 .22 <.01 .26 -.01
T333MTHe .15 <.01 .21 .31 -.07 <.01 -.09 <.01 .01 .02 <.01 .03 <.01 -.03 <.01 .20 -.05 .20
T410ARIT .17 <.01 .23 .25 .04 <.01 .04 <.01 <.01 .01 <.01 .01 <.01 .05 <.01 .05 .13 .05 <.01 .12 .05
T420TABL .04 <.01 .09 .10 -.03 <.01 -.03 <.01 <.01 -.03 <.01 -.04 <.01 -.02 <.01 .07 -.03 .07 <.01 .04 .00
T430CLER .11 <.01 .15 .16 .01 .19 .01 .18 <.01 .01 <.01 <.01 <.01 .02 <.01 .08 .03 .08 <.01 .05 .07
T440OBJE .05 <.01 .09 .10 -.04 <.01 -.04 <.01 <.01 -.05 <.01 -.06 <.01 -.03 <.01 .07 -.02 .07 .36 .07 -.02
2C .00 .57
2C & KM -.02 .01
2C & MY .07 <.01
2F .00 .98
2R .00 .94
2S .03 <.01
2V -.09 <.01
2Y .10 <.01
A6 .10 <.01
K0 .01 .21
K1 -.14 <.01
KM -.01 .01
P .04 <.01
VZ -.06 <.01
g .31 <.01 .33 <.01

Notes: For the analyses conducted using Major et al.'s (2012) linkage, we use the same abbreviations as the authors: K0 = General Verbal Information; K1 = Science Knowledge; A6 = English achievement; KM = Math achievement; P = Perceptual Speed; VZ = Visualization.

Multivariate range restriction was used instead of direct range restriction. The latter is most applicable to situations on which applicants are selected on a predictor test and criterion-related validity is measured using a restricted sample that passed the predictor test. The Project TALENT mental abilities tests were administered for research purposes and were not used to make college admissions decisions. However, given the use of college admissions requirements, including standardized admissions testing (e.g., the SAT), at the time Project TALENT was conducted, it is likely that range restriction occurred. Since direct range restriction corrections can underestimate corrected validity coefficients when selection is not direct (see Hunter et al., 2006), a multivariate range restriction correction procedure was conducted. Multivariate range restriction correction allows for range restriction corrections to be made simultaneously on a battery of predictor tests, such as the Project TALENT mental abilities tests. It also allows for corrections when differential range restriction occurs on a set of correlated predictors, which is possible for the Project TALENT mental abilities tests. A Monte Carlo simulation by Fife et al. (2016) indicated that multivariate range restriction corrections (conducted using the Pearson-Lawley method) “almost always do as good or better than” Hunter et al.'s (2006) Case IV corrections (which are considered a state-of-the-art procedure by many meta-analysts). Participants took the Project TALENT mental abilities tests in either the 9th, 10th, 11th, or 12th grade. Multivariate range restriction corrections were conducted separately for each grade using each grade’s unrestricted and restricted variance-covariance matrices.

To make the corrections, the restricted and unrestricted correlation matrices were obtained. Next, the correlations involving the criterion (i.e., college GPA) were corrected for criterion unreliability and the standard deviation for the criterion was adjusted using formula 6-28b from Nunnally and Bernstein (1994) so that it reflected the standard deviation of criterion true scores rather than the standard deviation for observed scores. The correction was conducted on the restricted correlation matrix since criterion reliability is assessed using a restricted sample (see Hunter et al., 2006). After converting these to variance-covariance matrices, the psychmeta package (Dahlke & Wiernik, 2019) in R was used to conduct the multivariate range restriction corrections. Next, the iopsych package (Goebl et al., 2016) in R was used to compute composite scores (e.g., personality scale scores, principal component scores, which are described above) for the corrected correlation matrix. After making all four corrections, a sample size-weighted Fisher z average was obtained across the four grade levels. The Fisher z values were then converted back to correlation coefficients and an operational validity matrix was produced. Finally, using the reliabilities from the full unrestricted sample, a true score corrected correlation matrix was also produced. The matrices were then imported into SPSS and regression analyses were conducted on the matrices to examine the operational and true score incremental validities for each test over g.

In the final restricted raw dataset, the g scores using all 59 tests had an observed criterion-related validity of .27. Correcting for multivariate range restriction and criterion unreliability yielded an operational validity (ρov) of .38. Further correcting for predictor reliability yielded a true score validity (ρTS) of .38 as well. The g scores that omitted crystallized knowledge had an observed validity of .32, an operational validity of .41, and a true score validity of .42. We focused our analysis on the g scores omitting crystallized knowledge since these had slightly higher criterion-related validity. We computed the partial correlations (robs.g) between each subtest and college GPA, controlling for the g scores, and we ran a series of regression analyses and recorded the incremental validity (ΔRobs) and observed standardized regression weights (βobs). The regression analyses were also conducted on the multivariate corrected correlation matrix to yield operational incremental validity estimates (ΔROV) and associated standardized regression weights (βOV).

Following recommendations by Schmidt et al. (1981), Schmidt and Hunter (1996, scenario 6), and Schmidt (2011), we also corrected the multivariate corrected correlation matrix for predictor unreliability and ran a true-score regression analysis to obtain true score incremental validities (ΔRTS) and true score standardized regression weights (βTS). These results can be found in Table 5. Note that the corrected true score correlation between the Etiquette Information Test (T144ETIQ) and g exceeded 1.00 and its β-weight and ΔR are reported as not applicable in Table 5.

Finally, we conducted three sets of Structural Equation Models (SEMs) using the two factor structures described earlier. We first estimated the standardized regression path coefficients from g to the criterion. Next, leaving the path from g to the criterion intact (and fixing it to the first value), we added a path from a second factor (e.g., 2C) to the criterion and recorded its estimate. This serves as an estimate of the unique contribution of the s factor over and above g. We repeated this analysis for each of the remaining factors; however, we only included two paths (one of which was from g, the other from the target factor) to the criterion in each model. We then conducted this analysis at the test score level; we conducted 59 SEMs, all of which had a path from g to the criterion and each of which had a different subtest-to-criterion path. For example, one model had paths from g, 2F, and Test 290 to the criterion and another had paths from g¸2Y, and Test 211 to the criterion.

The SEMs were conducted twice, once using our linkage of the tests to Carroll’s (1993) factors and once using Major et al.'s (2012) linkage. To provide a better modeling of the factor structure, we included data from 486 cases that had scores on all tests, graduated from college, but lacked criterion data (using AMOS’s estimate means and intercepts option to handle the missing data). The fit statistics for CFAs of the Carroll linkage were mixed (χ2=261,426, df =1,642; CFI=.715; RMSEA=.066); however, those for the Major et al. linkage were better (χ2=66,263, df =599; CFI=.893; RMSEA=.055). We computed the model implied correlation between each subtest and college GPA using the model that only allowed g to load onto college GPA. The model implied correlations are shown in Table 5 along with the difference between them and the observed validities.

The test scores and the g scores were significantly related to college GPA. However, after controlling for g, the unique variance of the test scores had either negative validities or small validities that were often less than .10 (i.e., “Irish validities,” according to Kunin, 1965, and Seberhagen, 2007). There were four exceptions. The Word Functions in Sentences Test (T240WORD) had a partial correlation, β-weight, and standardized SEM path in excess of .10. This finding held up after correcting for multivariate range restriction and predictor reliability. The Bible Information Test (T142BIBL) and the Punctuation Non-Information Test (T233PUNC) had corrected operational validity standardized regression weights that were .10 or higher. Several more tests had corrected true score validity standardized regression weights that were .10 or higher, but are not discussed here since their operational validities were not .10 or higher. Additionally, the 2Y and A6 factors in the SEM had standardized paths of .10 or higher.

The observed validities were very close to those implied by the SEM model that only allowed g to predict, suggesting that much of the correlation between the tests and college GPA is attributable to the relationship between g and college GPA. Only two tests (T240WORD and T410ARIT) had observed correlations that were .10 or higher than the SEM model-implied correlations, suggesting that g did not capture all of the validity of these two tests. Thus, we found strong support for Hypothesis 1. In contrast to Cucina et al. (2016), we did not find that the two math knowledge tests (T106MATH and T312MTH2) added positive incremental observed validity over g. Neither test had appreciable operational or true score incremental validity over g, but both did have true score standardized regression weights of .10 or higher. That said, it was concluded that Hypothesis 4 was not supported.

Hypothesis 2 was supported as conscientiousness had an operational validity of .22. Although the other big five factors predicted the criterion, they did not add incremental validity over Conscientiousness.

To address Hypothesis 3, we used Hayes (2017, 2019) PROCESS 3.3 SPSS macro to determine if there was an interaction between conscientiousness and the g scores in predicting college GPA. Although statistically significant, this interaction term did not add appreciable incremental validity over the main effects (ΔR<.01). A ‘pick-a-point’ analysis of simple slopes indicated the conditional effects of the g scores at the 16th, 50th, and 84th percentiles to be .19 (p<.001), .20 (p<.001), and .21(p<.001), respectively. There were no statistically significant transition points using the Johnson-Neyman method within the range of Conscientiousness scores.

Hypothesis 5 was tested using model 4 of Hayes (2017, 2019) PROCESS 3.3 SPSS macro with g serving as the predictor and conscientiousness serving as a control with both being mediated by HS GPA in the prediction of college GPA. We ran this analysis twice (switching the roles of g and conscientiousness). The results can be found in Table 6, with a graphical depiction of the mediation models in Figures 1 and 2. The direct and indirect effects of both predictors were significant (as evidenced by confidence intervals not including zero after 10,000 bootstrap samples) providing support for mediation and Hypothesis 5, albeit with a small standardized effect size.

Figure 1
Figure 1.Mediation model with g-non-information tests as predictor and conscientiousness as covariate using standardized high school grades and completely standardized effects.
Figure 2
Figure 2.Mediation model with conscientiousness as predictor and g-non-information tests as covariate using standardized high school grades and completely standardized effects.

Finally, we also performed multiple regression using g (βobs=.17; βOV=.26; βTS=.24), conscientiousness (βobs=.06; βOV=.07; βTS=.05), and HS GPA (βobs=.34; βOV=.33; βTS=.37) to predict college GPA and obtained a multiple R of .46 (ρOV = .52; ρTS =.54).

Table 6.Mediation analysis results.
X Covariate M Effect
Type Unstandardized Completely Standardized
Coefficient Confidence Interval p
Lower Bound Upper Bound
g-non-Info tests Conscient. HS GPA Path a 4.15 4.31 4.27 <.001 .33
Path b 0.02 0.02 0.02 <.001 .32
Path c 0.20 0.19 0.21 <.001 .30
Path cʹ 0.13 0.12 0.14 <.001 .19
Path ab 0.07 0.07 0.07 .11
Covariate on M 1.14 1.09 1.19 <.001 .22
Covariate on Y 0.02 0.11 0.02 <.001 .06
Covariate Total 0.04 0.04 0.04 <.001 .13
g-non-Info tests Conscient. Std. HS GPA Path a 0.48 0.47 0.49 <.001 .38
Path b 0.18 0.17 0.19 <.001 .34
Path c 0.20 0.19 0.21 <.001 .30
Path cʹ 0.11 0.11 .012 <.001 .17
Path ab 0.09 0.08 0.09 .13
Covariate on M 0.12 0.11 0.12 <.001 .22
Covariate on Y 0.02 0.01 0.02 <.001 .06
Covariate Total 0.04 0.03 0.04 <.001 .13
Conscient. g-non-Info tests HS GPA Path a 1.14 1.10 1.19 <.001 .22
Path b 0.02 0.02 0.02 <.001 .32
Path c 0.04 0.03 0.04 <.001 .13
Path cʹ 0.02 0.01 0.02 <.001 .06
Path ab 0.02 0.02 0.02 .07
Covariate on M 4.15 4.03 4.27 <.001 .33
Covariate on Y 0.13 0.12 0.14 <.001 .19
Covariate Total 0.20 0.19 0.21 <.001 .30
Conscient. g-non-Info tests Std. HS GPA Path a 0.12 0.11 0.12 <.001 .22
Path b 0.18 0.17 0.19 <.001 .34
Path c 0.04 0.03 0.04 <.001 .13
Path cʹ 0.02 0.01 0.02 <.001 .06
Path ab 0.02 0.02 0.02 .08
Covariate on M 0.48 0.47 0.49 <.001 .38
Covariate on Y 0.11 0.11 0.12 <.001 .17
Covariate Total 0.20 0.19 0.21 <.001 .30

Discussion

The results of this study have implications in several domains. First, the finding that g predicts college GPA is particularly relevant for the ongoing discussion and criticism about college admissions tests. College admissions tests are g-loaded and provide a standardized method to compare and contrast college candidates from different backgrounds and high schools. They are also a way to admit students to colleges who are most likely to have success and by extension complete college and graduate. The finding that g predicts college GPA reinforces the legitimacy of using college admissions tests as part of the process to admit students into college. However, the lack of support for the traditional multiplicative model between motivation (defined here as conscientiousness) and ability does not apply to collegiate academic performance (given the miniscule interaction effect). Sackett et al. (1998) reported similar results for job performance; however, interaction effects are difficult to detect (McClelland & Judd, 1993) and additional research could be conducted.

Second, our finding that individual tests had very small validities with the criterion after controlling for g provides further disconfirmation of specific aptitude theory and suggests that the past I/O research we reviewed for the role of g versus specific abilities in the prediction of job and training performance also extends to college GPA. Those looking to predict college performance would do well to use a g-loaded predictor, such as general aptitude and ability tests, instead of predictors in sub-areas or domain-specific abilities. Future research should examine the role of g and specific abilities in predicting academic performance in primary and middle school.

There are several limitations to our study. The dataset is several decades old and so the results of this study should be replicated with newer data. Such a replication might include additional specific ability tests to better capture the 69 narrow abilities in Carroll’s (1993) Three-Stratum Theory and improve incremental validity over g and more current/up-to-date content in the Information tests. Second, the self-report for the collection of GPA data could have resulted in inaccuracies (e.g., inflation or embellishment). Third, our analyses did not include individuals who started but did not complete a Bachelor’s degree as we decided to focus on final overall cumulative grades rather than grades for Freshmen, for example. Finally, some of the tests included in Project Talent had low reliabilities. Thus, there is room for more investigation into specific aptitude theory.


Author Note

Portions of this paper were previously presented at the 2020 annual meeting of the Society for Industrial and Organizational Psychology. The views expressed in this paper are those of the authors and do not necessarily reflect the views of U.S. Customs and Border Protection or the U.S. Federal Government.