Methodological Issues in Developing a Learning Progressionbased Assessment System Jennifer Doherty, Karen Draney and Andy Anderson Michigan State University BEAR Center, UC Berkeley NARST 2012 Research Goal Develop a valid and reliable assessment system that measures student progress in terms of learning progression levels. Key Methodological Problem Each component of the system must meet two sets of criteria: – Statistical criteria based on measurement theory and practice – Conceptual criteria based on learning progression theory and practice, including conceptual coherence with other components of the assessment system Components of assessment system (NRC 2001) Components of assessment system (Wilson 2005) Component I: Construct Maps: Learning Progression Framework (Wilson 2005) Component I: Construct Maps: Learning Progression Framework Learning Progression Hypothesis: Consistency across processes with respect to principles and models. – Students who learn scientific discourse see how systems and processes are connected, applying principles and models across processes. – Lower-level students will not see scientific connections among processes, but their accounts will have similarities because they draw on a common pool of linguistic and conceptual resources. Component I: Construct Maps: Learning Progression Framework Implication of hypothesis: An assessment system should be able to measure levels of proficiency in the ability to: – trace matter and energy through systems – reason across scales for different processes that are consistent across processes. Component 2: Item design (Wilson 2005) Component 2: Item design • We want to measure students’ understanding of principles and models with minimal effects from scaffolding and local knowledge. Example Items • ENERPLNT (energy practice, plant growth process, MTF+CR) – Which of the following is(are) energy source(s) for plants? – Circle yes or no for each of the following: Water, Light, Air, Nutrients in soil, Plants make their own energy. – Please explain ALL your answers, including why the things you circled “No” for are NOT sources of energy for plants. • THINGTREE (mass/gas practice, plant growth process, MTF+CR) – A small oak tree was planted in a meadow. After 20 years, it has grown into a big tree, weighing 500 kg more than when it was planted. Do you think the tree will need any of the following things to grow and gain weight? – Circle yes or now for each of the following: Sunlight, Soil, Water, Air – If you circled yes, explain how the tree uses it. • TREEDECAYC (energy practice, decomposer process, CR) – A tree falls in the forest. After many years, the tree will appear as a long, soft lump on the forest floor. Is energy involved when the tree decays? – Circle one: Yes / No If your answer is yes, please explain how energy is involved. Component 3: Outcome space (Coding Rubrics) (Wilson 2005) Component 3: Outcome space (Coding Rubrics) Challenges: – Coding rubrics need to be aligned among items • among items of a single type of carbon transforming process (e.g., plant growth item). • among items of a single type of practice (e.g., tracing energy). – Coding rubrics need to be aligned with the learning progression framework • Responses that are coded at level 2 are not just partially incorrect. • They must be responses that contain indicators of level 2. (Alonzo et al. 2012) Component 3: Outcome space: Developing Coding Rubrics Methods: • Development of rubrics that combine the general Learning Progression framework levels with itemspecific level indicators • Iterative process of developmental coding and rubric revising • Iterative process of full coding, reliability checks, rubric revising, and recoding Interviews and written assessments that assess two dimensions of complex accounts (process and practice) Component 4: Measurement Model (Wilson 20005) How well do codes of the sample items fit the statistical measurement model? Discrimination ENERPLNT THINGTREE TREEDECAYC 0.54 0.50 0.47 Weighted Mean Square 1.02 0.88 1.11 Questions about individual items • Check coding rubrics for conceptual validity • Check the data on which step thresholds are based • Recognize the limitations of individual items Questions about whole assessment Criteria for results of validation analyses – Multi-dimensionality and discrimination: • a single dimension should account for most of the nonrandom variance • items should be aligned with that dimension. – Wright maps and item difficulty: • Step thresholds in approximately the same region of vertical space support claim of an underlying level of proficiency. – Alignment with student interviews • There should be correlation between a students written and interview codes High correlations among processes and practices support claim of an underlying level of proficiency. Animal Function Animal Function Animal Growth Combustion Cross Process Animal ComCross Decay Growth bustion Process Plant Growth 0.900 0.789 0.816 0.754 .828 0.862 0.818 0.764 0.837 0.847 0.743 0.847 0.856 0.924 Decay Plant Growth 0.880 Process Dimension Correlation Matrix Correlation between codes of Interviews and IRT student ability estimates (r=.526) Conclusions and Next Steps • We have multiple lines of evidence that these written assessments do a good job of categorizing most of the students that we have, Level 2 and Level 3 students. • Not a lot of good evidence we’re able to discriminate between Level 3 and Level 4 students. • Investigate conceptual validity of items that are too hard or too easy
© Copyright 2026 Paperzz