Roeper Review, 31:53–67, 2009 Copyright © The Roeper Institute ISSN: 0278-3193 print / 1940-865X online DOI: 10.1080/02783190802527372 UROR DELVING INTO DIMENSIONS OF MATHEMATICAL GIFTEDNESS Test of the Three-Mathematical Minds (M3) for the Identification of Mathematically Gifted Students Three-Mathematical Minds Ugur Sak In this study, psychometric properties of the test of the three-mathematical minds (M3) were investigated. The M3 test was developed based on a multidimensional conception of giftedness to identify mathematically talented students. Participants included 291 middle-school students. Data analysis indicated that the M3 had a .73 coefficient as a consistency of scores. Exploratory factor analysis yielded three separate factors explaining 55% of the total variance; however, one-factor solution seems to best fit the data. The convergent validity analysis showed that M3 scores had medium to high-medium correlations with teachers’ ratings of students’ mathematical ability and students’ ratings of their own ability and their liking of mathematics. The findings provide partial evidence for the validity of M3 test scores for the identification of mathematically gifted students. Whether giftedness is a single entity or it emerges in different forms has been a subject of numerous theoretical debates. Ideas related to this issue are abundant (Heller, Monks, Sternberg & Subotnik, 2000; Sternberg & Davidson, 1986, 2005). The assessment of multiple forms of giftedness in mathematics, using the three-mathematical minds model, based on an integration of the theory of successful intelligence (Sternberg, 1997), studies on expertise and theories of mathematicians about mathematical ability (e.g., Gould, 2001; Polya, 1954a, 1954b) is the subject of this study. First, I review theories and empirical research related to the nature of mathematical ability and mathematical giftedness. Then I present the three-mathematical minds model and the test of the three-mathematical minds I developed based on the three-mathematical minds model for the identification of mathematically gifted individuals. Recieved 1 July 2006; accepted 7 September 2007. Address correspondence to Ugur Sak, Anadolu University, College of Education, Department of Special Education, 26470 Eskisehir, Turkey. E-mail: [email protected] MATHEMATICAL ABILITY ACCORDING TO THE PSYCHOMETRIC VIEW In psychology, mathematical ability has often been studied using factor analysis. Mathematical ability takes many forms in factor analytic studies, depending on the nature of mathematical tasks. Mathematical tasks also vary, depending on the branches of mathematics such as arithmetic, algebra, geometry, numbers, or statistics, and on the cognitive processes such as induction, deduction, or computation. Following the advent of factor analysis, at least four or five factors underlying mathematical ability were found frequently. A numerical factor was found in early factor analytic studies of mathematical ability (Spearman, 1927; Thurstone, 1938; Werdelin, 1958) consisting mostly of addition, multiplication, and other arithmetical problems. Another common factor found was related to visual or spatial tasks that required the manipulation of objects in two- and three-dimensional space. Other common factors found in factor analytic studies related to mathematical ability were the reasoning factors, consisting mostly of induction (number series) and deduction tasks (Thurstone; Werdelin). Carroll (1993) reanalyzed 480 studies using exploratory factor analysis. Many of the studies also had datasets relevant to mathematical ability. His reanalysis indicated a 54 U. SAK hierarchical structure in cognitive abilities. In his reanalysis, Carroll found quantitative ability under the factor fluid intelligence. No unique mathematical ability existed according to the reanalysis. However, some first-level factors were found. These first-level factors were general sequential reasoning, quantitative reasoning, and induction. Carroll (1996) suggested that fluid intelligence was related to mathematical ability because most reasoning activities under fluid intelligence were associated with logical and quantitative concepts. However, most tasks used in prior factor analytic studies seem to measure aspects of analytical mathematical ability and mathematical knowledge, leaving out essential tasks underlying mathematical creativity. Mathematical creativity was not a subject in prior factor analytic studies with the exception of Sternberg’s study of successful intelligence, which indicated that separate analytical and creative abilities existed in mathematics (Sternberg, 2002). According to Sternberg (1997), the theory of successful intelligence has three aspects that underlie intellectual performance in academic domains, including mathematics. These are creative ability, analytical ability, and practical ability. Analytical ability involves comparing, contrasting, evaluating, and judging relatively familiar problems. Creative ability is invoked when the information processing factors of intelligence are applied to relatively novel problems or situations to create, design, imagine, suppose, explore, invent, or discover. Practical ability involves solving real-life problems. Practical ability is related to tacit knowledge. Sternberg and colleagues carried out a series of studies to investigate the validity of the theory of successful intelligence (see a complete review in Sternberg, 2002; Sternberg, Castejon, Prieto, Hautakami, & Grigorenko, 2001; Sternberg, Grigorenko, Ferrari, & Clinkenbeard, 1999), using the Sternberg Triarchic Abilities Test (STAT). Their factorial research revealed separate and uncorrelated analytical, creative, and practical factors in three domains: verbal, figural, and quantitative. Quantitative analytical ability was measured through a test of number series. Creative ability was measured by a test of novel number operations. Furthermore, quantitative practical ability was measured by scenarios, requiring students to solve real-life problems. CONCEPTIONS OF MATHEMATICAL ABILITY Beliefs about the nature of mathematics seem to contribute to conceptions of mathematical ability. For example, abstractions and generalizations are viewed to be the essence of mathematics, and mathematical thinking therefore is deemed mostly abstract and generalized thinking (Kanevsky, 1990; Krutetskii, 1976; Sriraman, 2003). In fact, beliefs about the nature of mathematics were found to influence how mathematicians carry out research (Sriraman, 2004a). Beliefs about the nature of mathematics and that of mathematical thinking therefore are seen in definitions of mathematical ability. For example, the mathematician Poincare (Gould, 2001) believed mathematical ability to be discernment. Conceptions of mathematical ability vary not only by individuals but also by disciplines. For example, conceptions of mathematicians differ from those of nonmathematicians who study mathematical ability. In the psychometric tradition, working memory and analytical processes have much value, as can be seen both in individual and in group IQ tests. In contrast, according to some mathematicians, intuition is as important as logic and power of memory in inventions in mathematics. Poincare (Gould, 2001), for example, pronounced that a strong memory or attention does not make people mathematicians, but intuition is the main instrument that enables them to conceive the structure or relations among mathematical entities. He stated that people with great memory and attention and the capacity for analysis also can be gifted in mathematics. They can learn every detail of mathematics, but they lack the ability to discover. There also is another type of mathematical ability, the mental calculator, who can make very complicated calculations very quickly. According to Hadamard (1945), only a few eminent mathematicians possess such talent. In addition to broad conceptions of mathematical ability discussed above, researchers have investigated specific characteristics of mathematical ability, particularly mathematical giftedness. For example, Krutetskii (1976) found the following characteristics to be essential for high mathematical ability: (a) the ability to comprehend the formal structure of a mathematical problem; (b) the ability to generalize numerical and spatial relations; (c) the ability to operate with numbers and other symbols; (d) the ability to switch from one mental operation to another; (e) the ability to grasp spatial concepts; and (f) the mathematical memory for mathematical generalizations and structures. Other researchers underscored specific features of mathematical ability. Among these features are (a) the ability to visualize mathematical problems or relations (Presmeg, 1986), (b) the ability to think recursively (Kieren & Pirie, 1991), (c) the ability to carry out analogy and heuristics (Polya, 1954a), (d) the ability to discern mathematical relations (Gould, 2001), and (e) the ability to make decisions (Ervynck, 1991). The aforementioned characteristics of high mathematical ability also were supported empirically by recent studies. For example, Sriraman (2003), based on analyses of gifted students’ mathematical problem-solving, found gifted students spending a great deal of time trying to understand problem situations and devising plans to solve these problems. In this study, gifted students were reported to utilize particular cases to make generalizations, to search for similarities among problems, and to employ analogies to explain similarities among problems. In another study, Sriraman THREE-MATHEMATICAL MINDS (2004b) observed gifted students using visualization and reversibility in constructing mathematical truths. According to Sriraman, gifted students’ problem-solving processes were guided by their strong intuition as they made conjectures and devised constructions to validate their initial conjectures. 55 mind. However, mathematical expertise, although more related to mathematical knowledge than cognitive processes, might be much more related to mathematical analysis and creativity than I consider. That is to say, analysts and creators differ in their use of cognitive processes, whereas experts differ in their knowledge structure. Levels of Mathematical Ability As with all intellectual abilities, mathematical ability shows developmental stages and levels within each stage and sublevels within each level. For example, Usiskin (2000) proposed seven hierarchical levels of mathematical ability based on educational attainment, mastery of mathematical concepts, and contribution to the domain of mathematics. At the lowest level, Level 0, are adults who even are unaware of arithmetic. The first level, basic talent, can be characterized as the development of ability and knowledge repertoire to reason about number concepts and arithmetic. The levels from the first to the sixth distinguish among talent levels that can be found among mathematics students from high school to the end of graduate school. However, Levels 6 and 7 characterize adult mathematicians who are on the peak of the mathematical domain. Level-6 mathematicians are exceptional and can be found in the top few percentage of mathematicians. At the seventh level are those mathematicians who make great contributions to the domain of mathematics. They are the geniuses of mathematics, such as Euler and Gauss. Note that Usiskin’s proposition of mathematical ability deals with how mathematical talent develops into hierarchical levels based on formal and informal mathematical experience. Levels of mathematical ability also can be seen in Poincare’s classification of two types of mathematicians, analysts and creators (Gould, 2001), although this hierarchical classification is implicit in his conception of mathematical ability. According to Poincare, analysts can do great analytical work but lack the ability to create, whereas creators have the potential to analyze and to invent. Hence, creative mathematicians can be considered a subset of analytical mathematicians. By the same token, Sriraman (2005) distinguished between mathematical giftedness and mathematical creativity, proposing that creative mathematicians make up a small portion among mathematicians, and mathematical giftedness does not imply mathematical creativity; that is, all creative mathematicians are mathematically gifted, but all mathematically gifted mathematicians are not creative mathematicians. Types of Mathematical Ability Among mathematicians, two kinds of ability characterize two types of mathematical minds; one that is analytical and one that is creative (Gould, 2001; Hadamard, 1945; Polya, 1954a). The third mind that I believe is important for understanding and assessing mathematical ability is the expert Analysts and Creators Analytical mathematical ability and creative mathematical ability are defined differently by those who deal with mathematics and the psychology of mathematics. Poincare (Gould, 2001), for example, defined creativity in mathematics as discernment or the ability to choose among mathematical combinations that are useful and analytical mathematical ability as the ability to dissect mathematical combinations. Creators make discoveries of theorems. Analysts, on the other hand, usually do microscopic work by analyzing mathematical rules, axioms, combinations, or theorems the creator already has discovered. The primary thinking tool of analysts is logic. Deductive reasoning is the particular case of this logic. The primary thinking tool of creators is mathematical induction by rule discovery, analogy, or mathematical constructions. Poincare (Gould, 2001) believed that the nature of mathematicians’ minds makes them either analysts or creators, and this can be seen in the way they approach a novel problem. That is, not only do the two minds work differently, but also the ways these two minds deal with a problem make them different. Analysts approach a problem by their logic, whereas creators approach the same problem by their intuition. That is to say, the nature of the problem does not change the nature of thinking of the two minds. Further, the analyst is weak in visualizing space, and the creator is weak in long calculations. Similar to Poincare’s thoughts about mathematical ability, Polya (1954a, 1954b) proposed two kinds of reasoning underlying mathematical ability. One is demonstrative reasoning. The principal function of demonstrative reasoning is to distinguish a proof from a conjecture or a valid argument from an invalid argument; thus, demonstrative reasoning ensures certainty in mathematics. The other type of reasoning is plausible reasoning. The primary function of plausible reasoning is to differentiate a more reasonable conjecture from a less reasonable conjecture by providing logical evidence. Empirical research also reveals evidence about how creative mathematicians produce creative work. For example, Hadamard (1945) uncovered the nature of creative mathematicians’ thinking. He reported the prominence of visual images, as well as sudden illuminations guided by intuition and subconscious processes during mathematical discoveries. Similarly, Sriraman (2004a) interviewed creative mathematicians about their problem-solving experience and work habits and identified common characteristics of mathematical creativity as intuition, proof, social interaction, 56 U. SAK imagery, and heuristics, in addition to the four-stage Gestalt model of preparation-incubation-illumination-verification. Knowledge Experts The concept of expertise refers to a well-organized body of accessible domain-specific knowledge and skills. Researchers have explored the nature of the skills and knowledge that underlie expert performance (e.g., Ackerman, 2003; Alexander, 2003; Chi, Glaser, & Farr, 1988; Ericsson, 2003; Hatano & Osuro, 2003; Lajoie, 2003). One of the key findings is that experts’ knowledge structures differ from novices (Chi et al., 1988). The kind of knowledge experts possess is characterized as involving an organized, conceptual structure or schematic representation of knowledge in memory. Their knowledge is organized around important ideas of their disciplines, and it includes information about conditions of applicability of key concepts and procedures. This type of representation of knowledge, as opposed to isolated facts, enables experts to think around principles when encountering a problem, whereas novices tend to solve problems by attempting to recall specific formulas that could be applicable to the problem (Chi et al.). For example, compared to novices, mathematics experts are more able to recognize patterns of information in situations that entail specific classes of mathematical solutions. Experts also show some variation. Some researchers underscored the importance of studying different forms of expertise, such as routine expertise and adaptive expertise (e.g., Alexander, 2003; Hatano & Inagaki, 1986; Hatano & Osuro, 2003). Routine expertise involves the quick and accurate solving of relatively familiar problems. In contrast, adaptive expertise involves processes that lead to innovation and those that lead to efficiency. Routine expertise is a form of knowledge expertise, whereas adaptive expertise is a form of analytical or creative mind. Those mathematicians who only teach mathematics courses at universities are good examples of knowledge experts or of routine expertise, whereas mathematicians who publish or review research or theoretical articles for scientific journals are good examples of analytical or creative mathematicians or of adaptive expertise. THE THREE-MATHEMATICAL MINDS MODEL (M3) From this author’s point of view, mathematical giftedness is the mathematical competence demonstrated in the form of production, reproduction, or problem-solving in any branch of mathematics at a given time and is recognized as remarkable by members of mathematical communities (e.g., teachers or mathematicians). The three-mathematical minds model is a tool to reconcile various views about giftedness. However, the three minds differ in three aspects (see Figures 1 & 2). The first difference is the cognitive components, such as ANALYST Creative analyst CREATOR Master Expert analyst Creative expert KNOWLEDGE EXPERT (Routine Expertise) FIGURE 1 The Three-Mathematical Minds Model and seven forms of mathematical giftedness or mathematical expertise. This model is based on patterns of giftedness proposed by Sternberg (2000), research on expertise and mathematical ability, and teachings of eminent mathematicians about mathematical talent. memory, intuition, or logic, to carry out cognitive tasks. The second difference is in cognitive tasks, such as routine work, novel work, or analytical work. Finally, they differ in their end-products as a function of applying certain cognitive components in different tasks, such as knowledge production, reproduction, or solving familiar problems. In the three-mathematical minds model, knowledge expertise refers to routine expertise, while adaptive expertise refers to creators and analysts. Knowledge experts might differ from analysts and creators in their knowledge representation, amount of knowledge, and experience but not necessarily in their cognitive ability and styles. Although their knowledge is specialized, representing domain specificity or task specificity, their expertise is in the form of routine expertise. Therefore, their cognitive end-products can be characterized as the reapplication of experience to solving familiar problems, which do not necessarily produce creative work. Unlike knowledge experts, creators and analytical thinkers manifest their work more as a function of their thinking, such as the way they approach a novel problem; the differences in their attentional ability, logic, and intuition; and the way they deal with information, such as to search for novelty or ambiguity. Creative mathematicians and analytical mathematicians, too, are experts, but in different forms. They not only are knowledge experts but also experts in how to think mathematically. As pointed out earlier, their expertise is in the form of adaptive expertise, leading to innovation or efficiency. Figure 1 shows that it is plausible to think of mathematical giftedness in seven forms like patterns of giftedness (Sternberg, 2000). That is, mathematical giftedness may be THREE-MATHEMATICAL MINDS 57 Mathematical Minds Knowledge Expert Memory Recall Creative Routine work Routine Problem Solving FIGURE 2 Intuition Induction Analytical Novelty Production Logic Deduction Proof Reproduction Major instruments of mathematical minds applied in cognitive tasks and their end-products. conceptualized based on the interactions of the three minds. For example, the interaction of knowledge and analytical ability produces an expert analyst, who is competent both in domain knowledge and in analysis. The interaction of knowledge and creativity makes a creative expert, who is a good intuitive free thinker and has remarkable domain knowledge. By the same token, the interaction of analysis and creativity gives birth to a creative analyst, who has both good, logical judgment and an a priori synthetic judgment. Finally, the interaction of all brings into being a master, who demonstrates remarkable analytical ability, domain knowledge, and creative productivity and who, no doubt, should be very rare. Compared to Usiskin’s (2000) hierarchical classification of mathematicians, expert and analytical minds of the three-mathematical minds model can be considered among level-5 mathematicians; creative minds represent level-6 mathematicians; and the interaction of the three, or the master, can be thought to be the level-7 mathematician. The three-mathematical minds model should be thought as an instrument for developing theory-driven tests of mathematical ability to assess students’ three primary cognitive abilities for production, reproduction, and routine problemsolving in the domain of mathematics. The rest of this article deals with research about the validity and psychometric properties of the test of the three-mathematical minds (M3) developed by the author based on the three-mathematical minds model to identify students with high ability in analytical, creative, and knowledge aspects of mathematical ability. The specific questions of exploration in this study were as follows: 1. What is the underlying structural validity of the test of the three-mathematical minds? 2. What psychometric properties does the test of the three-mathematical minds have? METHOD Participants The total number of participants was 291 (female = 133; male = 158), including sixth (n = 63), seventh (n = 117) and eighth (n = 111) grade students from four different schools. Schools A (n = 41) and B (n = 27) were located in one city, School D (n = 152) was in another city. School C (n = 71) was located in a rural area. All schools were located in the southwestern part of the United States. The socioeconomic status of students in each school varied from lower to upper. Students of Schools A and D came mostly from middle-class families. Students of Schools B and C mostly came from low-middle-class families. The participants’ ages ranged from 10.5 to 15.5. The mean age was 13.04. The racial distribution of participants was as follows: White (72.2%), Mexican American (14.4%), Asian (3.4%), Black (2.4%), American Indian (1.4%), and others (6.2%). Procedures Development of the Test of the Three-Mathematical Minds (M3) A team of experts, with the leadership of the author, developed mathematics problems according to the threemathematical minds model (M3) and the three-level cognitive complexity model (C3), as seen in Table 1. The team consisted of the following members: two mathematicians, one with a PhD in the science of mathematics and the other with a PhD in mathematics education; two mathematics teachers, who worked at a middle school and a high school; and the author, who specialized in the assessment of cognitive abilities, creativity, and giftedness. 58 U. SAK TABLE 1 Item Development According to the Three-Mathematical Minds Model and the Three-Level Cognitive Complexity Model Cognitive Complexity Factor Knowledge Analytical ability Creativity Subtests Algebra Geometry Statistics Linear Reasoning Conditional Reasoning Categorical Reasoning Induction Insight Selective Problem Solving Level I Level II Level III Factual Factual Factual 5 elements 1 condition 2 sets: 1 superset and 1 subset Free classification Team agreement Selective encoding Relational knowledge Relational knowledge Relational knowledge 5 elements and additions 2 conditions 3 sets: dissections, 1 superset, and 1 subset 1 relation Team agreement Selective encoding and selective combination Conceptual-schematic knowledge Conceptual-schematic knowledge Conceptual-schematic knowledge 6 elements, coefficients, and divisors 2 conditions and double negation 4 sets: 1 intersection, 2 dissections, 2 supersets, and 2 subsets 1 rule and 1 relation Team agreement Selective encoding, selective combination, and selective comparison First, the author developed 27 sample problems to measure the analytical mathematical ability, creative mathematical ability, and mathematical knowledge at three levels of cognitive complexity according to theories framing the M. Then, the sample problems were sent to the mathematicians for their review prior to the author’s initial meetings with them. The author met with each mathematician twice to review, revise, and develop new problems. The first meeting resulted in modifying 4 problems, developing 20 new problems, keeping 3 original problems unchanged, and omitting 20 problems. The second meeting ended with modifying 22 problems, developing 5 new problems, keeping 3 original problems unchanged, and omitting 2 problems, thus yielding a total of 30 problems. In the second phase, the final 30 problems were sent to the mathematics teachers to review the content and difficulty level of each problem according to the level of eighth-grade students’ mathematical ability. The teachers used the following scale to rate problems: 1 (very easy), 2 (easy), 3 (average), 4 (difficult), and 5 (very difficult). They rated the difficulty level of each problem by comparing it to other problems in the same subtest. For example, the problems in the insight subtest were compared only to each other, not to other problems in the other subtests. The reason for such a rating procedure was the author’s conviction that only problems of the same type should be compared in their difficulty. Because of strong agreement between the author and the mathematicians on the difficulty level of the problems, high correlations were expected between item cognitive complexity (ICC) levels of the 30 problems and difficulty levels of these problems as rated by the two teachers according to eighth-grade students’ levels of mathematical ability. However, correlational analysis indicated low and nonsignificant correlations between the ICC levels and the two teachers’ ratings of the difficulty levels of the problems (r = .29 and .33), contradicting the initial agreements. An examination of the ICC levels and teacher ratings of the difficulty levels of the problems showed that ICC levels of the problems in the induction and insight subtests diverged considerably from teacher ratings. The main reason for this divergence, perhaps, was the fact that these problems were not developed according to the C3, but the author and the mathematicians agreed on the difficulty level of each problem. Therefore, I set out a third phase, during which I revised 7 problems, replaced 3 problems, and omitted 3 additional problems. At the end of this process, the test included nine subtests for a total of 27 problems. Each subtest had three problems. The teachers’ ratings on the problems in the insight and induction subtests also were integrated in the final revision. Then, the mathematicians reviewed the final problems before the test was given to the participants. However, the problems were not sent again to the teachers for their ratings of item difficulties. As reported, in the results section, correlation between item difficulties and the ICC was moderately high, supporting the validity of the C3. Table 1 shows the contents and the psychological sources of the ICC levels revised after the final phase. Item Content and the Use of the M3 in Item Development A total of nine subtests were developed to measure three components of the three-mathematical minds as seen in Table 1: knowledge component, creativity component, and analytical component. It was hypothesized that each component of the three-mathematical minds was measured by three subtests. Problems in each subtest hypothetically were developed to measure an aspect of one of the minds. Studies on expertise provided the theoretical background in the development of knowledge component and problems in this component. Three branches of mathematics (algebra, geometry, and statistics) were used to develop three separate classes of problems. The theoretical purpose of these subtests was to measure factual, relational, and schematic knowledge to distinguish between those who demonstrated the knowledge of a novice and those who demonstrated knowledge possessed by experts. Algebra problems required knowledge of substitution, transposing, and factoring, as well as that of solving an algebra word problem by translating it into an equation with two unknowns. Geometry problems required knowledge of area, perimeter, and of angles as well as knowledge of relationships between area and perimeter. Statistics problems required knowledge of rate, percent, interest and data tables. Analytical ability was measured by three deductive subtests: linear syllogism, conditional syllogism, and categorical syllogism. Information processing research informed the development of these problems. In linear syllogism, two or more quantitative relations were given between each of two pairs of items, depending on the cognitive complexity of the item. One item of each pair overlapped with an item of another pair, such as A < 2B and B/2 > C. The task of the problemsolver was to figure out the letter with the smallest or the greatest value. In conditional syllogism, one condition was presented with five conclusions, of which only one option satisfied the condition. The following is an example of a conditional syllogism problem: If x < 0, then: (a) x2 > 2x; (b) x2 < 3x; (c) x2 < 0; (d) x2 < x + x; (e) x2 > x2. In categorical syllogism, participants had to figure out relationships between members of classes. Group memberships were presented in a table. Creativity ability was measured by three subtests: induction, selection, and insight. Induction problems can be characterized either by mathematical rule discovery or by rule production. Selection problems can be characterized by finding out relevant or irrelevant information, by selective encoding, by selectively combining encoded information, and by analogizing combinations to other constructions that are presented in the problem stem or among answer options, as related to the solution of the problem. The theoretical purpose of these problems was to measure selection in problem solving as articulated by Poincare (Gould, 2001) and Polya (1954a) and as theorized by Davidson and Sternberg (1984). The insight subtest contained problems of mathematical recreations or nonroutine problems. The theoretical background of these problems came from Gestalt psychology. Item Cognitive Complexity Item cognitive complexity refers to psychological sources of item difficulty, such as levels and kinds of knowledge or cognitive processes a problem requires for the solution. The difficulty level of each problem used in the M3 test, except the insight and induction problems, was developed according to the three-level cognitive complexity model (C3) developed by the author (see Table 1). The three-level cognitive complexity model was developed according to a performance continuum on which people can be categorized as novices, developing experts, and experts (see Figure 3) based on their intellectual performance on some tasks that are related to the domain of mathematics. Because the continuum reflects a developmental performance, Number of People THREE-MATHEMATICAL MINDS 59 Knowledge Domain Novices Level 1 Developing Experts Experts Level 2 Level 3 Levels of Expertise FIGURE 3 Continuum of expertise in a domain. People might differ in their level of expertise in different domains. As the level of expertise increases, the number of people decreases in a domain. An expert in one domain can be a novice in another domain. people may be found at every point on the continuum. Based on the continuum, therefore, expertise can be measured at different levels. Although I used three levels, novices, developing experts, and experts, other levels might be found by dividing the continuum at different points provided that these points have psychological meaning. Item Format, Scoring, and Test Administration Two types of item formats were used. Most items were presented in a multiple-choice format consisting of a problem stem and five answer options. Only one option was correct in these types of problems. The second format was of kind—more than one method and solution or more than one method but only one solution was accepted as correct. One point was given for each correct answer in both multiple choice and open problems. No point reduction was taken for wrong answers. Mathematics teachers administered the test during students’ regular mathematics classes in the beginning of the spring semester of the 2004–2005 school year. The testing was done in one sitting, taking about 45 minutes. The teachers read standard instructions before the testing. RESULTS Reliability of the Tests of the Three-Mathematical Minds Means and standard deviations for student variables and the M3 subtests by grade are presented in Table 2. The reliability of the test of the three-mathematical minds was investigated through Kuder-Richardson reliability analysis (KR20). The analysis showed a .73 coefficiency level for the entire test, slightly exceeding the minimum desired level (.70) for consistency of scores for psychological tests. Relationships Among the M3 Subtests The relationships among the subtests of the M3 were investigated using Pearson product-moment correlation coefficient. Preliminary analyses were performed to check the 60 U. SAK theoretically should correlate. Partial correlation coefficients were computed, while grade was controlled in the equation, to investigate the relationships between teachers’ rating of students’ mathematical ability, students’ rating of their own mathematical ability, and their liking of mathematics and the M3 subtests (Table 4). Correlations ranged from low to high-medium, with the majority of correlations being statistically significant. Particularly important were the correlations between the M3 total score and the other variables. These correlations, medium to high-medium, were as follows: .45 between the M3 total and teachers’ rating (p < .01), .36 between the M3 composite and students rating of their own ability (p < .01), and .35 between the M3 composite and students’ rating of their liking of mathematics. TABLE 2 Means and Standard Deviations for Student Variables and the M3 Subtests by Grade Grade 6 Variable Teacher rating Student liking Student rating Algebra Geometry Statistics Linear syllogism Conditional syllogism Categorical syllogism Selection Induction Insight M3 composite M Total 7 SD M 8 SD 4.00 .95 3.45 1.08 3.06 1.57 2.87 1.30 3.89 .96 3.39 1.00 1.60 .66 1.50 .71 .67 .90 .72 .86 1.11 .86 1.08 .80 .73 .87 .76 .72 .81 .82 .89 .83 .82 .92 .62 .79 .52 .72 .49 .62 .90 .79 .91 .66 .32 .56 .27 .47 7.49 3.99 7.23 3.26 M SD M SD 3.50 2.78 3.36 1.61 1.01 1.21 1.01 1.09 .85 .59 1.11 .53 9.03 1.41 1.43 1.08 .72 .90 .79 .90 .99 .94 .67 .71 .69 3.94 3.59 2.88 3.49 1.55 .82 1.13 .85 .95 .75 .54 .99 .38 7.98 1.23 1.41 1.04 .70 .90 .81 .83 .89 .88 .66 .71 .59 3.77 Grade Differentiation A one-way between-groups analysis of variance (ANOVA) was conducted to examine performance differences of sixth-, seventh-, and eighth-grade students on the M3 total score (see Table 2 for mean performance for each grade level). Levene’s test for homogeneity of variances in scores of the groups showed no violation (p = .077) (Tabachnick & Fidell, 1996). The analysis indicated statistically significant differences among the grades (F [2, 288] = 7.5, p < .001). The effect size, calculated using eta squared, was .05, a medium effect according to Cohen (1988). Post hoc tests using Tukey’s honestly significant difference (HSD) revealed the following mean differences among the grades. Eighth graders performed significantly higher than seventh and sixth graders (p < .01 and p < .05, respectively) and no significant performance difference existed between seventh and sixth graders on the M3 composite. These findings provided developmental evidence for the construct validity of the M3 in terms of developmental differentiation. assumptions of normality, linearity, and homoscedasticity. Algebra, statistics, and induction subtests distributed normally. Geometry, linear syllogism, conditional syllogism, categorical syllogism, insight, and selection subtests were positively skewed. That is, scores were scattered around low performance. Table 3 shows a number of statistically significant and nonsignificant correlations among the subtests, with the highest associations among the knowledge subtests: algebra, geometry, and statistics. Categorical syllogism and selection subtests had the lowest associations with the other subtests. In fact, the categorical subtest did not have any significant correlations with any other subtests, and the selection subtest had a statistically significant correlation only with the induction subtest (r = .16; p < .05). Item-Total Score, Item-Subtest, and Subtest-Total Score Correlations Convergent Validity Whether the M3 test items and subtests were homogenous or heterogeneous was investigated to determine the Convergent validity of the M3 was investigated to examine whether it correlated with other variables with which it TABLE 3 Bivariate Correlations Among the M3 Subtests Variables 1. Algebra 2. Geometry 3. Statistics 4. Linear syllogism 5. Conditional syllogism 6. Categorical syllogism 7. Selection 8. Induction 9. Insight 2 3 4 5 6 7 8 9 M3 Total .35** .30** .36** .28** .42** .25** .34** .47** .26** .26** .05 .03 .01 .08 .03 .03 .09 .03 .02 .16** .10 .24** .34** .26** .25** .26** .13* .06 .21** .36** .24** .17** .38** .00 .07 .12* .57** .73** .57** .59** .67** .32** .30** .55** .50** *p < .05; **p < .01 (two-tailed). THREE-MATHEMATICAL MINDS TABLE 4 Partial Correlations Between Student Variables and the M3 Subtests Correlations Subtest Grade Age Teacher rating Student liking Student rating Algebra Geometry Statistics Linear syllogism Conditional syllogism Categorical syllogism Selection Induction Insight M3 composite .02 .16** .06 .14* .13* .03 .05 .13* .16** .18** −.03 .11 −.02 .07 .05 .02 .01 .04 .10 .07 .26** .33** .38** .23** .31** .09 .13 .29** .25** .45** .20** .25** .24** .11 .24** .16** .07 .19** .18** .35** .21** .33** .34** .08 .27** .01 .07 .22** .19** .36** Note. The effect of grade was removed in correlational analyses between the M3 subtests and students variables. *p < .05; **p < .01 (two-tailed). degree to which the M3 items measured a unified construct or multiple constructs. The degree of homogeneity or heterogeneity of a test has some relevance to its construct validity. Because the M3 is a measure of multidimensional 61 aspects of mathematical ability, an investigation of item homogeneity and heterogeneity in the M3 provides information about its construct validity. Point biserial correlations between the M3 items and the subtests were computed. As seen in Table 5, correlations between the items and the M3 total score ranged from low to high, with all correlations being statistically significant except for item 26. It had a very low and statistically insignificant correlation with the M3 total score (r = .06). Items 2, 6, and 15 had low correlations with the M3 total, although the correlations were significant. The second mode of analysis was point biserial correlation computed between the items and the subtests to examine the homogeneity and heterogeneity of the items. In other words, I assumed that an item was supposed to correlate highly with the subtest in which it was located to show homogeneity and to correlate at a low level with a subtest in which it was not located to show heterogeneity. Correlations ranged from low negative to very high positive correlations. For example, the correlation coefficient was −.12 between item 26 and the subtest selection in which it was not located (p < .05). The correlation was .88 between item 7 and the subtest insight in which it was located (p < .01). TABLE 5 Item-Subtest and Item-Total Test Point Biserial Correlations Subtest Item-subtests Algebra Geometry Statistics L. syllogism Con. syllogism Cat. syllogism Selection Induction Insight Total 1. L. syllo 2. Algebra 3. Con. syllo 4. Geometry 5. Statistics 6. Selection 7. Insight 8. Induction 9. Con. syllo 10. L. syllo 11. Induction 12. Geometry 13. Statistics 14. Cat. syllo 15. Cat. syllo 16. Cat. syllo 17. Algebra 18. Selection 19. Selection 20. L. syllo 21. Insight 22. Con. syllo 23. Algebra 24. Statistics 25. Induction 26. Insight 27. Geometry .24** .47** .24** .20** .27** −.03 .15** .24** .14** .17** .10 .28** .18** .13* −.03 .00 .80** .03 .05 .10 .24** .28** .54** .13* .07 .02 .26** .38** .08 .29** .76** .34** .05 .29** .28** .28** .19** .18** .75** .25** .14** −.06 −.01 .36** .09 .01 .21** .30** .35** .13* .15** .15** .08 .56** .29** .13* .19** .29** .66** .02 .22** .29** .16** .12* .11* .20** .67** .04 −.05 .03 .26** .05 −.03 .06 .19** .17** .14** .38** .04 −.03 .27** .67** .05 .21** .30** .29** −.02 .13* .28** .16** .63** .10* .35** .15** .09 .01 .09 .33** .04 .03 .56** .20** .14* .05 .05 .05 −.03 .21** .17** .06 .72** .35** .31** .09 .31** .29** .61** .21** .23** .39** .20** −.02 −.06 .17** .34** .11* .08 .11* .29** .62** .16** .07 −.04 .09 .24** .08 .00 −.06 −.08 .01 .04 −.05 .09 −.01 .00 .13* .08 −.04 .74** .79** .51** .05 .11* .03 .08 .07 .12* .04 .16** .03 .01 .11* .01 −.09 .01 .03 .07 .69** .08 −.01 .10 .01 .08 .06 −.03 .06 .03 .13* .10* .54** .51** .02 .08 .20** −.03 .12* .05 −.12* .13* .24** .11* .16** .23** .23** .02 .19* .69** .08 .09 .51** .22** .21** .11* .06 .12* .27** .05 .04 .13* .11* .25** .02 .13* .58** −.06 .30** .08 .04 .29** .29** .30** −.01 .88** .12* .21** .22** .15** .27** .19** .05 −.02 −.06 .23** .11* .03 .03 .61** .24** .09 −.02 −.04 .36** .18** .46** .17** .44** .51** .51** .16** .41** .47** .37** .34** .33** .56** .37** .29** .15** .21** .56** .22** .14** .28** .42** .50** .23** .25** .18** .06 .49** Parentheses indicate items’ subtests. *p < .05; **p < .01 (two-tailed). 62 U. SAK What should be read from Table 5 are correlations between the items and the subtests in which the items are located. The results indicated that items had high or very high correlations with the subtest in which they were located, whereas items had very low to medium correlations with the subtest in which these items were not located. Only item 24 had a medium correlation with the statistics subtest in which it was located, and item 26 had a medium-level correlation with the insight subtest in which it was located; however, the correlations still were statistically significant (p < .01 for both items). As seen in Table 5, the pattern of correlations provided partial support for both homogeneity and heterogeneity of the M3 items because many items also had significant correlations with the other subtests in which they were not located; however, these correlations were low. Item Difficulty and Item Discrimination As reported earlier, item difficulties were developed according to the three-level cognitive complexity model (C3). Item difficulties were determined through computing the proportion of the participants who answered them correctly. As seen in Table 6, item difficulties ranged from .02 (extremely difficult) to .93 (very easy) with a mean of .29 (very difficult). Correlation between item difficulties and item cognitive complexity levels was found to be .64 (p < .01), a high correlation supporting the validity of the C3. The mean item difficulty of the entire M3 test seems to be very difficult for the entire population used in this study. Such a difficulty level results from the fact that item difficulties were constructed according to eighth-grade students’ level of mathematical ability, but sixth- and seventh-grade students also were participants in this study. TABLE 6 Item Difficulties (ID), Item Cognitive Complexity Levels (ICC), and Item Discriminations (D) Item ID ICC D Item ID ICC D 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 .36 .93 .42 .41 .58 .27 .30 .67 .20 .23 .09 .30 .34 .31 .30 1 1 1 1 1 1 1 1 2 2 2 2 2 1 2 .62 .12 .60 .77 .80 .19 .58 .64 .44 .41 .27 .76 .48 .39 .14 16 17 18 19 20 21 22 23 24 25 26 27 .15 .51 .17 .10 .26 .06 .33 .13 .15 .22 .02 .11 3 2 2 3 3 2 3 3 3 3 3 3 .21 .84 .26 .12 .36 .30 .72 .19 .27 .20 .03 .40 Mean SD .29 .20 2 .83 .41 .24 Note. Item difficulty range (ID): .80–1.00 very easy, .60–.79 easy, .40–.59 moderately difficult, 20–39 very difficult, and .00–.19 extremely difficult. Item discrimination: .50 and above high, .30–49 moderate, .15–29 low, .00–.15 very low and negative. Classical item discrimination analysis was used to examine item discrimination. The index for item discrimination was computed based on a comparison of the performance of the upper-25th percentile group and that of the lower-25th percentile group on the M3 total score. Table 6 shows the discrimination index for each item. Discrimination indices ranged from .12 (low discrimination) to .84 (very high discrimination). The M3 test had a mean of .41 discrimination index, which is a moderate level of discrimination. Factorial Structure of the Tests of the Three-Mathematical Minds Preliminary Analyses and Prior Decisions Prior to performing factor analysis, the suitability of data for factor analysis was assessed. Inspection of the correlation matrix revealed the presence of coefficients of .3 and above. The Kaiser-Meyer-Oklin value for sampling adequacy was .81, exceeding the recommended value of .6 (Kaiser, 1974). The Bartlett’s test of sphericity reached statistical significance (p < .001), which supported the factorability of the correlation matrix. Moreover, prior to performing factor analysis, three decisions were made. First, the type of factor analysis to be used was determined. Exploratory factor analysis “finds the one underlying factor model that best fits the data,” whereas confirmatory factor analysis “allows the researcher to impose a particular factor model on the data and then see how well that model explains responses” (Bryant & Yarnold, 1995, p. 109). Moreover, confirmatory factor analysis is “generally based on a strong theoretical and/or empirical foundation” (Stevens, 1996, p. 389). Because, the M3 is a new instrument and no studies have been done before on the instrument’s structural validity and the subtests in the M3 were only presupposed to measure some aspects of the three dimensions of mathematical ability, principal axis factoring was used to examine the factor structure of the M3 in this study. Second, two criteria were established to determine the number of factors in the factor analysis: (a) A factor should have an eigenvalue of 1.00 or above; and (b) at least 50% of the variance in the original data set should be explained by the number of factors to be determined. Third, the type of rotation to be used was determined. As recommended by Tabachnick and Fidell (2001), multiple rotations were examined using varimax, quartimax, equimax, and direct oblimin. Direct oblimin rotation was chosen for use as the underlying constructs were also correlated. Factor Analytic Findings The principal axis factoring of the nine subtests revealed the presence of three factors with eigenvalues exceeding 1. Because only three factors had an eigenvalue above 1.00 and explained more than 50% of the total variance, these THREE-MATHEMATICAL MINDS TABLE 7 Three-Factor Solution from Principal Axis Factoring With Direct Oblimin Rotation Factor Subtest 1 Geometry Linear syllogism Statistics Algebra Induction Conditional syllogism Selection Insight Categorical syllogism Percent of variance explained Eigenvalues .68 .58 .53 .51 .51 .40 2 3 .39 29.30 −.35 .31 13.58 12.16 2.82 1.11 1.03 Communality .40 .22 .20 .21 .18 .32 .04 .20 .03 three factors were investigated further, using direct oblimin rotation with Kaiser normalization. Factors that had less than .30 absolute values were not reported in the rotated solution. All subtests achieved loadings above .30 on at least one factor. As seen in Table 7, the three-factor solution explained 55.03% of the variance, with factor 1 contributing 29.30%, factor 2 contributing 13.58%, and factor 3 contributing 12.16%. Factor 1 consisted of geometry, algebra, statistics, linear syllogism, conditional syllogism, and induction subtests. Factor 2 consisted of the categorical subtest and the insight subtest. Factor 3 consisted of the selection subtest only. The first factor was labeled as the “knowledge-reasoning” factor, the second as the “analytical” factor, and the third as the “creativity” factor. The factor analytic findings provided partial support for the theoretical validity of the test of the three-mathematical minds. Although three separate factors were found, four subtests did not cluster in the factors for which they were developed. Indeed, the findings show that one-factor solution might be a better solution for the dataset used in this study. The induction subtest was developed to measure an aspect of the creative mind, but this subtest was found in the knowledge-reasoning factor. Conditional and linear syllogism subtests were developed to measure the analytical mind, but they too were found in the knowledge-reasoning factor. Finally, although the insight subtest was developed to measure an aspect of the creative mind, it did not cluster in the creativity factor. These findings mean that new subtests need to be developed to measure creative and analytical abilities. DISCUSSION In this research study, psychometric properties of the test of three-mathematical minds (M3) were investigated. The M3 63 test was developed based on a multidimensional conception of mathematical ability. The findings provided partial support for the reliability, convergent validity, and developmental validity of the M3 test. However, factorial findings did not show strong support for the factorial validity of the three-mind model. Reliability of the M3 The reliability coefficient may be interpreted in terms of the percentage of score variance attributable to different sources (Anastasi & Urbina, 1997). The analysis yielded a reliability coefficient of .73 for the entire test. Although the coefficient is above the minimum level, it is not very strong, in that over 25% of the variance in scores in the M3 is attributable to error variance. Recall that the M3 is not a measure of a single trait, but a measure of multidimensional aspects of mathematical ability. That is, the items in the M3 are heterogeneous; therefore, the interitem consistency of the M3 should not be expected to be very high because the interitem consistency is influenced largely by the heterogeneity of the behavior domain sampled (Anastasi & Urbina). The M3 has nine subtests measuring separate aspects of mathematical ability. Overall, the reliability finding shows a good level of reliability for this preliminary research. However, test– retest reliability might be a better reliability indicator for the Test of the Three-Mathematical Minds. Factorial Validity of the M3 Although the factor analysis revealed three separate factors out of nine subtests, most subtests clustered in one factor, which does not provide strong evidence for the factorial validity of the three-minds model. The factor analysis showed that some subtests of creative and analytical minds clustered in the knowledge factor, contradicting the author’s theoretical position. The author’s initial assumption was that geometry, algebra, and statistics subtests cluster in the knowledge factor; the three deductive subtests cluster in the analytical factor; and insight, selection and induction subtests cluster in the creativity factor. Table 7 shows loadings of each subtest on the factor in which it clustered. According to the analysis, geometry, algebra, statistics, linear syllogism, conditional syllogism, and induction subtests clustered in the first factor and explained almost 30% of the variance. This factor can be labeled the knowledge-reasoning factor. Note that the thirdlevel problems in the knowledge factor require some reasoning because these problems entail conceptual understanding of subject matter. For example, the level-three problem in the geometry subtest requires understanding perimeter and area relationships for the solution. What was unexpected was the clustering of induction, linear syllogism, and conditional syllogism subtests in the knowledge factor. They have low to medium, but statistically 64 U. SAK significant, correlations with the other subtests of the knowledge factor. An inspection of the problems in the induction subtest indicates that the first-level problem (categorizing numbers), and the second-level problem (finding out the sum of the internal angles of a shape of 14 sides) can be solved by using mathematical knowledge, as well as by inductive processes. However, the third-level problem (finding out the number of turns of a wheel) does not require mathematical knowledge. Needless to say, the firstand second-level problems contribute to the overlap between the knowledge factor and the induction subtest. Another reason for the overlap can be seen from the finding that the induction subtest has a high correlation with the geometry subtest because the second-level and the thirdlevel problems in the induction subtest also require spatial ability, whereas the first-level problem requires numerical ability. The subtests conditional syllogism and linear syllogism, which were supposed to cluster in the analytical factor, clustered in the knowledge factor. Linear syllogism and conditional syllogism problems require focused attention, as well as the comparing and contrasting aspects of analytical ability. Therefore, I predicted that the linear syllogism and conditional syllogism subtests would correlate highly with the categorical syllogism and cluster in the analytical factor, but they did not. The conditional syllogism problems and linear syllogism problems resemble algebra problems only in language, as they have algebraic symbols. Nevertheless, they require little algebraic knowledge. Conditional syllogism problems also require the use of coefficients and factors. Therefore, the correlation between algebra and the conditional syllogism subtest is .34, a statistically significant finding. As a matter of fact, the conditional syllogism correlates with the geometry subtest at a higher level, .47. This author believes that the overlap between the conditional syllogism and the knowledge factor occurs because of the similar nature of the problems in the conditional syllogism and the algebra subtest. In fact, some students, who were not good at algebra, might not have even attempted to solve conditional problems because of surface similarities between algebra and the conditional syllogism problems. The categorical syllogism subtest has very low correlations with the other subtests. The categorical syllogism subtest was found to be a separate factor in factor analysis explaining 12% of the variance. The categorical syllogism problems, like linear and conditional syllogism problems, require focused attention, contrasting, and comparing. Therefore, this factor was labeled as the analytical factor by this author. The categorical subtest has a significant correlation, .13, with the induction subtest only. However, the overlap is quite small, 1.6%. Keep in mind that a significant correlation might exist between two variables with a large sample size. Therefore, this correlation does not tell much about the relationship between the inductive processes and the categorical syllogism processes. Indeed, the finding might be just an artifact of the sample size. Furthermore, although the categorical syllogism subtest itself is a separate factor in the factor analysis, the difficulty level of the categorical problems might have contributed to the distinction of this factor. Categorical problems have a mean difficulty level of .25, a very difficult level (see Table 6). The second factor, which included selective problem solving subtest, was labeled as the creativity factor by this author because problems in this subtest requires an unusual mode of thinking—selective thinking. Recall that, as pointed out before (Davidson & Sternberg, 1984, 1986), many creative ideas emerge from sorting out related and unrelated information and combining them selectively or through the use of analogies. Problems in the selective problem-solving subtest require selective encoding, selective comparing, and selective contrasting. Overall, the selective problem-solving subtest seems to measure a different aspect of human ability. This difference can be seen more clearly from point biserial item subtest correlations in Table 5. Problems in the selective problem-solving subtest correlate significantly only with the total score of this subtest. Overall, the factor analysis did not show a clearcut picture about the existence of three distinct mathematical minds. However, this conclusion can be drawn only from the dataset used in this study, in which only nine subtests were used and the participants were middle-school students only. Different findings can be obtained in another study if more subtests and a different group of participants were used. Convergent Validity of the M3 Campbell (1960) pointed out that a psychological test should correlate with other variables to which it should be related theoretically to show construct validity. Convergent validity can be investigated through the correlation of the same ability measured by different tests or through the correlation of similar abilities measured by the same or different tests. In this study, the convergent validity of the M3 was investigated by correlating students’ M3 scores with teachers’ rating of students’ mathematical ability, students’ rating of their own mathematical ability, and students’ rating of their liking of mathematics. Note that the first two ratings are some kinds of measures of students’ mathematical ability. The liking of mathematics is not a measure of mathematical ability, but it should be associated theoretically with mathematical performance. As read from Table 4, correlations between the M3 scores and the ratings range from medium to high-medium, with all correlations being statistically significant at the .01 level. What is interesting in the table is the pattern of correlations between the M3 subtests and the ratings. Both teachers’ ratings and students’ ratings correlate with the pure THREE-MATHEMATICAL MINDS knowledge-reasoning subtests, such as algebra and statistics, much higher than their ratings with the other subtests, such as selective problem-solving and insight. These correlations might mean that teachers and students associate mathematical ability more with the amount of mathematical knowledge and achievement in math classes than with creative and analytical ability. Perhaps some creativity and analytical problems also were unusual to the students. Overall, these findings provide evidence for the convergent validity of the M3. Moreover, further research is needed to investigate associations between scores on the M3 and scores on other tests of mathematical ability for a more clearcut picture of the convergent validity evidence. An investigation of the relationship between scores on the M3 and grades in mathematics classes or performance on an achievement test also is needed to provide criterion-related validity evidence for the M3. Developmental Differences Among Students of Various Grade Levels A major criterion used in the validation of a number of intelligence tests is age differentiation (Anastasi & Urbina, 1997). However, the use of age in the validation of aptitude tests, such as the M3, is not appropriate because they measure ability that is influenced largely by school learning. Therefore, the major criterion for aptitude tests should be grade differentiation, which is what the author used to check the M3 against grade to determine whether the scores in the M3 show a progressive increase with grade during middle school (sixth through eighth grade). The findings provide partial validity evidence about whether the M3 discriminates among different grade levels. The partial evidence means that eighth-grade students scored significantly higher than seventh- and sixth-grade students; however, the sixth graders scored slightly higher than the seventh graders. Needless to say, the latter finding contradicts the former, even though the difference between the sixth graders and the seventh graders is not significant. The discussion of this contradiction follows. An inspection of the sixth- and the seventh-grade sample sizes shows a significant difference. The size of the sixth grade is almost half that of the seventh grade. This difference might have contributed to the performance difference between the sixth and seventh graders, favoring the sixth graders. Another reason for the contradiction might come from the statewide achievement difference between sixthand seventh-grade students. Further, developmentally, some peaks and slumps might exist in students’ performance during middle school (Charles & Runco, 2000; Sak & Maker, 2006). Briefly, the findings show partial, developmental evidence for the validity of the M3. Therefore, the author’s future research agenda includes administering the M3 to a different sample to see if the same or different results are obtained. 65 Item-Total Score, Item-Subtest, and Subtest-Total Score Correlations Internal consistency correlations, whether based on items or subtests, show the homogeneity and heterogeneity of the test items used to measure the ability domain sampled by the test. In other words, the internal consistency correlations might provide evidence related to whether items in a test battery measure the same construct, similar constructs or completely different constructs. Anastasi and Urbina (1997) stated that correlations among items, subtests, and the total score are expected to be significant if the test battery is constructed to measure a single, unified construct. The author’s assumption is that low correlations exist among items and between items and subtests that are developed to measure separate constructs even though they are parts of the same test battery. For example, the M3 was designed to measure three aspects of mathematical ability; therefore, low correlations should be expected among items measuring different aspects of mathematical ability, such as the creative mind and the analytical mind. On the other hand, high correlations should be expected among items measuring the same aspect of mathematical ability, such as the analytical mind. As seen in the point biserial correlational matrix in Table 5, 26 of the 27 items in the M3 test battery correlate significantly with the total score. Indeed, most correlations are within the range of medium to high, providing evidence of strong internal consistency of item-total score relationships. In other words, the 26 items differentiate among the respondents in the same direction, as does the entire test battery. However, item 26 has a low and nonsignificant correlation with the total score showing low discrimination. Therefore, this problem needs to be revised or removed from the test battery. Furthermore, bivariate correlations between the subtests and the M3 total score are significant, ranging from medium to high. The correlations between items and the total score and the correlations between the subtests and the total score provide evidence of internal consistency related to the construct validity of the M3. The correlation pattern in Table 5 also provides additional psychological evidence for the internal consistency of the M3. What is most important in this pattern are the correlations between the items and the subtests in which the items are located and the correlations between the items and the subtests in which the items are not located. As seen in the table, the items, such as 1, 10, and 20 located in the linear syllogism subtest, have high and significant correlations with the subtests in which they are located. In other words, each linear syllogism problem differentiates among individuals in the same direction, as does the entire subtest linear syllogism. This pattern repeats itself for the other subtests as well. However, the only items that do not have high correlations with their associated subtests are items 24 and 26, although they have moderate and significant correlations. 66 U. SAK To this author, more interesting than the high correlations are low-positive and low-negative correlations between the items and the subtests in which these items are not located. In other words, these items and the subtests are not developed to measure the same constructs. The correlation between item 25, an induction problem, and the subtest conditional syllogism, for example, is −.04, meaning that they do not measure the same ability and they differentiate among different abilities. Low negative or low positive and nonsignificant correlations exist between most items of the M3 and the subtests that measure different constructs. These findings show the internal consistency of the M3 at the item level and subtest level. beginning of the spring semester; therefore, the participants had not completed their associated grade level. As a result, they had not mastered mathematical knowledge and had not developed skills taught at the end of their respective grade level. The time of the school year in which the test was given might have contributed to item difficulty because problems were developed according to eighth-grade students’ level of mathematical ability as rated by mathematics teachers. Another limitation is related to the methodology. The M3 items were hypothetically assigned into their subtests; therefore, a subtest-factor analysis was used instead of an item-factor analysis. An item-factor analysis is recommended for future research. Item Difficulty and Item Discrimination A moderately high correlation was found between item cognitive complexities (ICC) and item difficulties (r = .64). A careful examination of Table 6 shows that 8 items have high-level discrimination indices, 5 items have moderatelevel discrimination indices, and the rest have low discrimination indices. Note that the level-three items, the most difficult items, have low discrimination powers. These problems have low discriminations not because they do not discriminate between high ability and low ability but because they are just too difficult; even some of the individuals with the highest abilities could not solve them. Keep in mind that the M3 problems were developed according to eighth-grade students’ general-mathematical ability level, but sixth- and seventh-grade students also were participants in this study. The problems might have been too difficult for these groups of students. If the comparison percentile is changed from upper 25% to upper 3%, discrimination indices of the items would change meaningfully. In conclusion, this study showed the strengths and weaknesses of the test of the three-mathematical minds. The M3 test is a new instrument and in need of improvement, particularly in item difficulty, item discrimination, and item content. As reported and discussed in the previous sections, it particularly has some weakness at the factorial level. It is clear from the findings that new subtests should be developed to measure creative and analytical abilities. LIMITATIONS A number of limitations exist in this study, most of which pertain to the sample. First, the sample is small and it is not an exact representation of the U.S. population because it was drawn only from the southwest region of the country. Second, the number of sixth-grade students is much less than that of the seventh- and eighth-grade students. This difference might have contributed to the performance variation among the groups. Third, the test was given in the REFERENCES Ackerman, P. (2003). Cognitive ability and non-ability trait determinants of expertise. Educational Researcher, 32, 15–20. Alexander, P. (2003). The development of expertise: The journey from acclimation to proficiency. Educational Researcher, 32, 10–14. Anastasi, A., & Urbina, S. (1997). Psychological testing. Upper Saddle River, NJ: Prentice Hall. Bryant, F. B., & Yarnold, P. R. (1995). Principal-components analysis and exploratory and confirmatory factor analysis. In L. G. Grimm & P. R. Yarnold (Eds.), Reading and understanding multivariate statistics (pp. 99–136). Washington, DC: American Psychological Association. Campbell, D. T. (1960). Recommendations for APA test standards regarding construct, trait, and discriminant validity. American Psychologist, 15, 546–553. Carroll, J. B. (1993). Human cognitive abilities: A survey of factor-analytic studies. New York: Cambridge University Press. Carroll, J. B. (1996). Mathematical abilities: Some results from factor analysis. In R. J. Sternberg & T. Ben Zeev (Eds.), The nature of mathematical thinking (pp. 1–26). Mahwah, NJ: Lawrence Erlbaum. Charles, R. E., & Runco, M. A. (2000). Developmental trends in evaluative and divergent thinking of children. Creativity Research Journal, 13, 417–437. Chi, M. T. H., Glaser, R., & Farr, M. J. (1988). The nature of expertise. Hillsdale, NJ: Lawrence Erlbaum. Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.). Hillsdale, NJ: Lawrence Erlbaum. Davidson, J. E., & Sternberg, R. J. (1984). The role of insight in intellectual giftedness. Gifted Child Quarterly, 28, 58–64. Davidson, J. E., & Sternberg, R. J. (1986). What is insight? Educational Horizons, 64, 177–179. Ericsson, K. A. (2003). The search for general abilities and basic capacities: Theoretical implications from the modifiability and complexity of mechanisms mediating expert performance. In R. J. Sternberg & E. L. Grigorenko (Eds.), The psychology of abilities, competencies, and expertise (pp. 93–125). New York: Cambridge University Press. Ervynck, G. (1991). Mathematical creativity. In D. Tall (Ed.), Advanced mathematical thinking (pp. 42–53). Dordrecht, The Netherlands: Kluwer Academic. Gould, S. J. (Ed.). (2001). The value of science: Essential writings of Henri Poincare. New York: The Modern Library. Hadamard, J. (1945). The psychology of invention in the mathematical field. New York: Dover. Hatano, G., & Inagaki, K. (1986). Two courses of expertise. In H. Stevenson, H. Azuma, & K. Hakuta (Eds.), Child development and education in Japan (pp. 262–272). New York: Freeman. THREE-MATHEMATICAL MINDS Hatano, G., & Osuro, Y. (2003). Commentary: Reconceptualizing school learning using insight from expertise research. Educational Researcher, 32, 26–29. Heller, K. A., Monks, F. J., Sternberg, R. J., & Subotnik, R. F. (2000). International handbook of giftedness and talent. New York: Elsevier. Kaiser, H. (1974). An index of factorial simplicity. Psychometrika, 39, 31–36. Kanevsky, L. S. (1990). Pursuing qualitative differences in the flexible use of a problem solving strategy by young children. Journal for the Education of the Gifted, 13, 115–140. Kieren, T., & Pirie, S. (1991). Recursion and the mathematical experience. In L. P. Steffe (Ed.), Epistemological foundations of mathematical experience (pp. 78–102). New York: Springer-Verlag. Krutetskii, V. A. (1976). The psychology of mathematical abilities in school children. (J. Kilpatrick & I. Wirszup, Eds.). (J. Teller, Trans.). Chicago: Chicago University Press. Lajoie, S. (2003). Transitions and trajectories for studies of expertise. Educational Researcher, 32, 21–25. Polya, D. (1954a). Induction and analogy in mathematics. Princeton, NJ: Princeton University Press. Polya, D. (1954b). Patterns of plausible inference. Princeton, NJ: Princeton University Press. Presmeg, N. C. (1986). Visualization and mathematical giftedness. Educational Studies in Mathematics, 17, 297–311. Sak, U., & Maker, C. J. (2006). Developmental variation in children’s creative mathematical thinking as a function of schooling, age and knowledge. Creativity Research Journal, 18, 279–291. Spearman, C. (1927). The abilities of man. London: Macmillan. Sriraman, B. (2003). Mathematical giftedness, problem solving, and the ability to formulate generalizations. Journal of Secondary Gifted Education, 14, 151–165. Sriraman, B. (2004a). The characteristics of mathematical creativity. The Mathematics Educator, 14(1), 19–34. 67 Sriraman, B. (2004b). Gifted ninth graders’ notions of proof: Investigating parallels in approaches of mathematically gifted students and professional mathematicians. Journal for the Education of the Gifted, 27, 267–292. Sriraman, B. (2005). Are giftedness and creativity synonyms in mathematics. Journal of Secondary Gifted Education, 17(1), 20–36. Sternberg, R. J. (1997). Successful intelligence. New York: Plume. Sternberg, R. J. (2000). Patterns of giftedness: A triarchic analysis. Roeper Review, 22, 231–235. Sternberg, R. J. (2002). Beyond g. In R. J. Sternberg & E. L. Grigorenko (Eds.), The general factor of intelligence (pp. 447–479). Mahwah, NJ: Lawrence Erlbaum. Sternberg, R. J., Castejon, J., Prieto, M. D., Hautakami, J., & Grigorenko, E. L. (2001). Confirmatory factor analysis of the Sternberg Triarchic Abilities Test (multiple-choice items) in three international samples. An empirical test of the triarchic theory of intelligence. European Journal of Psychological Assessment, 17, 1–16. Sternberg, R. J., & Davidson, J. E. (1986). Conceptions of giftedness. New York: Cambridge University Press. Sternberg, R. J., & Davidson, J. E. (2005). Conceptions of giftedness (2nd ed.). New York: Cambridge University Press. Sternberg, R. J., Grigorenko, E. L., Ferrari, M., & Clinkenbeard, P. (1999). A triarchic analysis of an aptitude-treatment interaction. European Journal of Psychological Assessment, 15, 1–11. Stevens, J. (1996). Applied multivariate statistics for the social sciences (3rd ed.). Mahwah, NJ: Lawrence Erlbaum. Tabachnick, B. G., & Fidell, L. S. (2001). Using multivariate statistics (4th ed.). Needham Heights, MA: Allyn & Bacon. Thurstone, L. L. (1938). Primary mental abilities. (Psychometric monographs No. 1). Chicago: University of Chicago Press. Usiskin, Z. (2000). The development into the mathematically talented. Journal of Secondary Gifted Education, 11(3), 152–163. Werdelin, I. (1958). The mathematical ability: Experimental and factorial studies. Lund, Sweden: C. W. K. Gleerup. AUTHOR BIO Ugur Sak, PhD, is associate professor and director of the graduate program on gifted education at the Department of Special Education, College of Education, Anadolu University, Eskisehir, Turkey. He is the founder of the Education Programs for Talented Students at Anadolu University. E-mail: [email protected]
© Copyright 2026 Paperzz