Schedule Oral Paper Sessions - International Test Commission

Schedule Oral Paper Sessions
1. Test development.
2. Tests for challenging concepts.
Chair: José Muñiz
3 July, 11.00-12.30
Using the method of plural identification
for test developing and test using
Olga Mitina
Chair: April Zenisky
3 July, 11.00-12.30
The measurement of heterogeneity in
adolescents’ expected political
participation
Joseph KuiFoon Chow
Kerry J. Kennedy
The Symptom Checklist-90-Revised: Do
we need more than the total score?
Iris A.M. Smits
Marieke E. Timmerman
Dick P. H. Barelds
Esther Allart
Rob R. Meijer
Adding a frame of reference to
personality measures: Comparing the
predictive validity of two types of
contextualization
DjurreHoltrop
Marise Ph. Born
Reinout de Vries
Anita de Vries
Using quantitative and qualitative
approaches for career preference
profiling and exploration
Marié de Beer
Enhancing test development through
digital technology: Blended and
distributed modes of validation
Susan Davis-Becker
Chad Buckendahl
Development of a recruitment selfevaluation tool for the Royal Air Force
Felicity Hill-Miers
You can be sitting next to a creative
genius
Sara IbéricoNogueira
Leonor Almeida(presenter)
Evaluation of emotional competencies
(EC): A questionnaire that fits EC theory
Sophie Brasseur
MoïraMikolajczak
Reducing bias in terms of culture: An
eight-country study on Explanations of
Unemployment
Kostas Mylonas
AdrianFurnham
3. Psychometric issues in test
construction.
Chair: JosèMuñiz
3 July, 13.45-15.15
Evaluating the short-answer item type in
a professional test
Brad Ching-Chao Wu
Peter Watkins
On the use of abbreviated tests in
psychology
Peter M. Kruyen
Wilco H.M. Emons
KlaasSijtsma
The split-half reliability coefficient:
Prevalence and controversy
Nancy Anderson
Verification of the unidimensionality of
tests with dichotomously scored items
Jacob ArieLaros
Maria FabianaDamásioPassos
Felipe Valentini
Applying a Bayesian approach of
modeling responses and response times
to intelligence assessment
Rudolf Debelak
4. Different aspects of IRT.
Chair: Paula Elosua
3 July, 13.45-15.15
Measuring response styles and response
processes in rating-scales through uniand multidimensional IRT-models
LaleKhorramdel
Matthias vonDavier
Practical findings and issues with IRT
item calibrations and model fit analyses
HanwookYoo
Ronald K. Hambleton
Free tools for learning and applying Item
Response Theory
Kyung T. Han
Lawrence M. Rudner
The performance of item fit indices of the
Rasch model in the presence of omitted
items
Christine Hohensinn
Klaus D. Kubinger
Application of hierarchical mixture IRT
models to international large-scale
assessment data
Maria Elena Oliveri
Matthias vonDavier
5. The use of tests in work-related
settings.
Chair: Iain Coyne
3 July, 15.45-17.15
Predictive validity in aviation
psychology: Really not much more than
g?
Oliver Zierke
6. Assessment of student
performance.
Chair: AlettaOdendaal
3 July, 15.45-17.15
Use of many-facets Rasch
measurement in performance
assessment in higher education
YesimCapa-Aydin
Competency potential constructs &
person-role fit technologies
RainerKurz
Rab MacIver
Predicting sales performance with the
GFP, the Big Five and narrow traits
Thomas B. Sitser
Dimitri van der Linden
Marise Ph. Born
Factorial structure of the School
Teacher Formative Assessment Practice
SawLanOng
Developing and measuring character
and leadership competencies
Elly Broos
J. De Bruin
L. De Boer-Gutter
Measuring person-environment fit with
different response conditions
IlkeInceoglu
Peter Warr
The Chinese teachers’ conceptions of
assessment – re-conceptualization and
cross cultural comparison
LingbiaoGao
Kerry J. Kennedy
Yang Wu
Revisiting the separation of
achievement goals for the case of
secondary school students
Flaviu A. Hodis
Georgeta M. Hodis
The effect of testing on student
achievement, 1910–2010
Richard P. Phelps
7. Differential Item Functioning and
Differential Prediction.
Chair: Jan Bogg
4 July, 8.45-10.30
DIF analysis of the Turkish Pupil
Monitoring System across school types
and its’ impact on test results
NedimToker
GirayBerberoğlu
SerkanArikan
Assessing dependability and safety
potential: Examining DIF in three diverse
language versions
Tanya Yao
MathijsAffourtit
IlkeInceoglu
Investigating measurement invariance of
national tests across different subgroups
in England using the Rasch model
Stephen Anwyll
Qingping He
Fatima Lampreia-Carvalho
Comparison of English and Spanish
translations of a national certification
examination
Ada Woo
MarijanaDragan
Sarah Hagge
Philip Dickison
Differential validity and differential
prediction in interest inventories
KatjaPäßler
Andrea Beinicke
Benedikt Hell
Two issues in differential item
functioning and two recently suggested
solutions
David Magis
Bruno Facon
Paul De Boeck
8.Test bias across cultures.
9. CAT &Testlet research.
Chair: Cheryl Foxcroft
4 July, 8.45-10.30
Equivalence of English-only and
bilingual mathematics tests of limited
English proficient Malaysian-students
S. Kanageswari d/o
SuppiahShanmugam
SawLanOng
South Africa and UAE student
performance progression in IBT
Mathematics and Science
Van Nguyen
Chair: April Zenisky
4 July, 11.00-12.30
Multidimensional computer adaptive
testing in non-cognitive assessments
Anna Brown
The Indonesian version of MMPI - 2 RF: A psychometric evaluation
Magdalena S. Halim
Robustness
of
full-information
maximum likelihood method for
adaptively administered test data
Kyung T. Han
FanminGuo
Adaptation of Wechsler Intelligence
Scale for Children, 4th edition in
Pakistan
SaimaAmbreen
AnilaKamal
Evaluation of booklet effect in PISA
mathematics across countries
Luc T. Le
The cross-cultural comparison of the
Cognitive Development Assessment test
items across Turkish and German
students
SerkanArikan
ÖmerAhmetKonak
GirayBerberoğlu
KutlayYağmur
Investigating adaptation equivalence of
the Indonesian version of the Dutch
Family Relation Test
EfiFitriana
UripPurwono
Marc Delsing
Kai Welzen
Identifying critical testlet features using
Tree-based Regression: An illustration
with the AR-section of the LSAT
Muirne C. S. Paap
Qiwei He
Bernard P. Veldkamp
Classification and diagnosis using
Bayesian mixture models in the next
generation of computerized adaptive
testing
LihshingLeigh Wang
10. New developments in work related
testing.
Chair: Iain Coyne
4 July, 11.00-12.30
Additional
relevance
for
cognitive
performance testing: Measuring employer
brand attractiveness
HaraldAckerschott
Johanna Beckert
Esther Kordecki
Jelena Strache
Annika van Veen
Effect of computer-based testing on
candidate perception of selection justice
AylinDincer
LeventSevinc
MerveSusut
IpekAltintas
A neural network predictive model to
estimate future transfer of knowledge to
the work environment
Eduardo C. Cascallar
Carla Quesada
New issues in the assessment of work
design
Enrique Merino-Tejedor
Manuel Fernández-Ríos
Implementing a new selection model
across 27 European countries
Markus Nussbaum
11. Test use in educational settings.
12. Item design and test performance.
Chair: AlettaOdendaal
4 July, 13.45-15.15
Admission test behaviors for African
citizens compared to United States and
Asian citizens
Eileen Talento-Miller
Ronald Sibert
Hillary Chan
Chair: Sverre Nielsen
4 July, 13.45-15.15
Accessibility or comparability: Priorities in
mathematics test item designs
Leanne Ketterlin-Geller
Diane Gifford
PoojaShivraj
Driving classroom based health via
assessment and reporting engines
Janet Clinton
John Hattie
Kathryn Cairns
Robert McNeill
Rob Kirton
The measurement of feedback from the
student perspective
Debra Masters
John Hattie
Assessing children’s analogy learning: Is
working memory a piece of the puzzle?
Claire Stevenson
Marian Hickendorff
Wilma Resing
Willem Heiser
Paul de Boeck
Effects of working memory and attention on
items of different levels of complexity and
difficulty
Mariel F. Musso
Eduardo C. Cascallar
The construction task: A method to avoid
distractor weakness in figural matrices
Nicolas Becker
N. Raffel
F. M. Spinath
The role of context in science items and its
relation to students’ performance
Maria Araceli Ruiz-Primo
Min Li
Score
report
interpretation
educators perspectives
SaadChahine
from
A cross-national study of student effort in
TIMSS Advanced
Hanna Eklöf
Development and validation of teachers’
beliefs about student learning and
assessment for learning practice scales
Kim Koh
Edmund Song
Yi Xian Ho (presenting author)
13. Faking and test security.
Chair: Jacques Grégoire
4 July, 15.45-17.30
Control of response distortion in
adaptive personality measurement
Katharina Lochner
MaikeWehrmaker
AchimPreuss
Controlling faking effects on
personality measures: An example
with aggression measures
Cristina Anguiano-Carrasco
Pere J. Ferrando
Andreu Vigil-Colet
Controlling
response
bias
in
aggression measures: The IndirectDirect Aggression Questionnaire
Mireia Ruiz-Pàmies
Andreu Vigil-Colet
Urbano Lorenzo-Seva
Cristina Anguiano-Carrasco
Evaluating the release of an item
bank to respond to security concerns
Chad W. Buckendahl
Russell W. Smith
High-tech cheating:
How new
technologies thwart test security
efforts
Steve Addicott
Secure mobile testing solution
DevRamnane
Romil Gupta
AnkitSolanki
14. Test development around the
world.
Chair: Jan Bogg
4 July, 15.45-17.15
Current status of test adaptation and use
in Nigeria
Andrew A. Mogaji
15. Standard setting and calibration.
Exploring the adaptation of international
computer based testing guidelines for a
digitally developing country
Cheryl Foxcroft
Andrea Watson
Bronwyn Seymour
Caroline Davies
Kim Hurter
Psychological
assessment
and
socioeconomic level in Argentina
Elena Brenlla
Score alignment and vertical articulation in
mathematics in the Turkish Pupil Monitoring
System
ÇiğdemİşGüzel
GirayBerberoğlu
SerkanArikan
NedimToker
The lack of relevance of scientific
validation in the application of
psychological tests in Brazil
Paulo Roberto Teixeira Jr.
Assessing early mathematics skills in
developing countries: Instrument design,
development, and validation
Linda Platas
Wendi Ralaingita
Leanne Ketterlin-Geller
Setting performance standards for partial matrix
sampling assessment design
Abdullah Ferdous
Chad Buckendahl
Validating performance level descriptors for
Advanced Placement® Environmental Science
Exam
Rosemary Reshetar
Michael Chajewski
Pamela Kaliski
Karen Lionberger
Barbara Plake
Evaluating test design through complementary
standard setting studies
Gad S. Lim
Chad W. Buckendahl
Brett P. Foley
Impact of high stakes national tests and
public examinations: A UK perspective
Dennis Opposs
Qingping He
SomiaNasim
Matthew Glanville
Stephen Anwyll (presenting author)
Chair: Paula Elosua
5 July, 08.15-10.00
Equating with small samples
Stefan Bondorowicz
On designing data-sampling for Rasch model
calibrating an achievement test
Klaus D. Kubinger
DieterRasch
TakuyaYanagida
16. New methods of testing.
Chair: Sverre Nielsen
5 July, 08.15-10.00
Intelligence as assessed by modern
computer-based problem solving
scenarios
PhillippSonnleitner
Martin Brunner
Ulrich Keller
ThibaudLatour
Romain Martin
Webcamtest in educational settings.
Assessment of communication skills
Jeroen Kuntze
M.A. Hommes
P. Van den Heuvel
M. Kiestra
Powerplay at work: A multi-media tool
for (really) measuring integrity
Alec W.Serlie
Christof van Nimwegen
Joost Modderman
Herre van Oostendorp
Development in an academic
environment of a web-based testmanager with non-copyrighted
psychological tests
Mark Schittekatte
Johnny Fontaine
Marc Covents
Interactive, tablet-based clinical
assessment: Equivalency findings
Dustin Wahlstrom
Xiaobin Zhou
Mark Daniel
Simulation tests for aviation security
staff: Incremental validity over
conventional psychometric tests
Yehuda Esformes
Moshe Cohen
17. The consequences of feedback,
coaching and (re)testing on test
scores.
Chair: Jacques Grégoire
5 July, 10.30-12.00
Determinants of performance in
intercultural online assessment
MaikeWehrmaker
AchimPreuss
Katharina Lochner
18. Test fairness.
Chair: Cheryl Foxcroft
5 July, 10.30-12.00
Evaluating coaching practices in
admission test performance: Who does
it benefit the most?
Anely Ramirez
MladenKoljatic
Monica Silva
Reviewing the scientific evidence on
the negative consequences of high
stakes testing
Nicholas Marosszeky
John Bennett
Fidelity of implementation in an emerging
international assessment system: A Boolean analytical
approach
Magda Chia
Retesting in employee selection:
Lessons learned from multi-country
unproctored Internet-based tests
Yongwei Yang
Tzu-Yun Chin
Nikki Blacksmith
Anna Truscott-Smith
Retest effects and retest reliabilities of
cognitive ability tests
Dorothea Klinck
The design and use of test illustrations in international
test comparisons
Guillermo Solano-Flores
Chao Wang
How field test results can be used to maximize crosslanguage comparability of main data collection
instruments
Steve Dept
Andrea Ferrari
Juliette Mendelovits
Using cognitive diagnostic modeling to investigate
effects of linguistic simplifications of a source
assessment
Alvaro J. Arce-Ferrer
Rense Lange
Failures of measurement invariance of cognitive ability
tests are more than a nuisance
Jelte M. Wicherts