Gives prompt feedback - studynet.herts.ac.uk

The use of a computerized
automated feedback system
Trevor Barker
Dept. Computer Science
Contents
• Feedback considerations
• Approaches to feedback
• Automated feedback
– Previous research
• Examples
• Discussion
Chickering and Gamson’s Seven Principles:
Good practice in higher education…
1. Encourages contact between students and
lecturers
2. Develops reciprocity and cooperation among
students
3. Encourages active learning
4. Gives prompt feedback
5. Emphasises time on task
6. Communicates high expectations
7. Respects diverse talents and ways of learning
Gives prompt feedback
• Feedback must be prompt but it must also be
good i.e.
– Appropriate
– Useful
– Accurate
– Individual
– Fast
– Facilitate feed forward
Reasons for automated approaches to testing
and learning
• Vast investment in infrastructure
• Availability of MLE systems such as UH
Studynet
• Changes in nature of Higher Education
• Online and distance education
• Increase in student numbers (SSR)
• Increasing pressures on time and cost
5
Previous research
Computer-Adaptive Test
• Based on Item Response Theory (IRT)
• If a student answers a question correctly, the
estimate of his/her ability is raised and a more
difficult question is presented
• If a student answers a question incorrectly, the
estimate of his/her ability is lowered and an easier
question follows
6
Previous research
Computer Adaptive Testing
• Computer-Based Tests (CBTs) mimic aspects of a
paper-and-pencil test
– Accuracy and speed of marking
– Predefined set of questions presented to all participants
and thus questions are not tailored for each individual
student
• Computer-Adaptive Tests (CATs) mimic aspects of an
oral interview
– Accuracy and speed of marking
– Questions are dynamically selected according to student
performance
7
Benefits of the adaptive approach
• Questions that are too easy or too difficult
are likely to
– Be de-motivating
– Provide little or no valuable information about
student knowledge
• The CAT level identifies a unique boundary
between what the student knows and what
he or she does not know
8
Providing individual feedback based on
CAT.
• An application of the CAT approach is in the
provision of automated individual feedback
• This approach has been in operation for several
years at the University of Hertfordshire in two
BSc. Computer Science modules
• Recently this model has been extended to make
it easier to use on other modules
9
About the Feedback
• Learners received feedback on:
–
–
–
–
Overall proficiency level;
Performance in each topic;
Recommended topics for revision
Cognitive level (Bloom)
• Feedback on assessment performance was
initially made available to learners via a webbased application
10
Bloom’s taxonomy
11
Example questions
12
13
Performance Summary
14
Points for Revision
15
Results: tutors’ opinions
• Tutors consider that the fast feedback provided by
a CAT is as good as or better than that currently
provided in many cases.
• The link to Bloom’s levels was positive
• The approach was considered to be efficient,
possibly freeing time for other activities
• CAT considered to be best as a formative tool,
rather than for summative assessment
• Some tutors were concerned that the approach
was ‘impersonal’
• There is a need for a monitoring role for tutors,
for practical and ethical reasons
16
Recent research
• The CAT automated feedback system has been
extended from objective testing to include
written and practical tests
• Testing and evaluation of the new system with
approximately
– 350 first yearBSc (1 final practical test),
– 120 second year BSc(2 written and practical tests)
and
– 80 final year BSc (2 final practical tests)
– 70 MSc students ( 2 written tests)
First prototype
Detailed marking scheme for one
question showing feedback
Converted manually into simple
database file
Output from system - email
Output from system - email
Later prototype
Final summary screen
Added features
• Markers able to comment on the completeness of the
hand-in
– In this version, the hand-in information is presented to the
marker who may then make additional comments on the
completeness or nature of the hand-in.
• Feedback was determined by the system based on the
mark awarded in each section of a question, reading it from
the database file for the assignment.
• After all the question sections had been marked, the
system presented a final summary screen so that the
marker could check that the marks had been awarded
accurately.
• The marker can add additional feedback at the end
Results
• Student attitude to feedback was good
irrespective of score on test
–
–
–
–
–
Useful
Fair
Convenient
Quantity
Quality
• Internal moderator happy with feedback
• Suggestions from moderator were
included in the next prototype
Latest version
Modifications
• Easy to set up feedback database
automatically
• Tutors can modify and add to feedback for
each question
• Additions to feedback saved for re-use later
In summary
• Larger class sizes, greater use of online and
distance assessment ensures that feedback is
often too slow and too general to be of any real
use to learners.
• Personalised automated feedback is likely to
become increasingly important in the future. It is
being used in four modules currently at UH.
• Learners and tutors accept the need for
automated feedback and most appreciate the
benefits of such systems.
• The system is being further developed to make it
simpler for general use
33