Committee on Instruction Recommendations: Part I

Review of SUNY Oneonta
Course Evaluation Form
Report and Recommendations from
The Committee on Instruction: Part I
March 16, 2009
Literature Review
Research Data
• One concern frequently expressed in
published studies is the response
rate
Oneonta’s Response Rates
100%
81%
82%
82%
84%
82%
83%
85%
Response Rate
80%
60%
40%
20%
0%
Spring Fall Spring Fall
2005 2005 2006 2006
Fall Spring Fall
2007 2008 2008
Layne et al. (1999)
100%
Response Rate
80%
61%
60%
48%
40%
20%
0%
Paper Evaluations
(Enrollment = 1,246)
Electronic Evaluations
(Enrollment = 1,207)
Thorpe (2002)
100%
Response Rate
80%
60%
50%
46%
40%
20%
0%
Paper Evaluations
(Enrollment = 414)
Electronic Evaluations
(Enrollment = 430)
Dommeyer et al. (2003)
100%
Incentives
for electronic evaluations
(randomly assigned):
80%
75%
Response Rate
1) modest grade incentive
=> 60%
87% response rate
2) in-class demonstration
=> 40%
53% response rate
3) early grade notification
20%
=> 51% response rate
4) no incentive
0%
Evaluations
=> 29% Paper
response
rate (with
no incentives)
43%
Electronic Evaluations
(some with incentives)
Dommeyer et al. (2003)
100%
Response Rate
80%
70%
60%
40%
29%
20%
0%
Paper Evaluations (with
no incentives)
Electronic Evaluations
(with no incentives)
Kulik (2005), Study 1
100%
Response Rate
80%
75%
74%
Paper Evaluations (18
classes)
Electronic Evaluations
(18 classes)
60%
40%
20%
0%
Kulik (2005), Study 2
100%
80%
Response Rate
80%
65%
60%
40%
20%
0%
Paper Evaluations (57
graduate classes)
Electronic Evaluations
(70 graduate classes)
Donovan et al. (2006)
100%
83%
76%
Response Rate
80%
60%
40%
20%
0%
Paper Evaluations
(Enrollment = 258)
Electronic Evaluations
(Enrollment = 261)
Avery et al. (2006)
100%
Response Rate
80%
68%
60%
46%
40%
20%
0%
Paper Evaluations
(Enrollment = 1,957)
Electronic Evaluations
(Enrollment = 1,080)
Heath et al. (2007)
100%
82%
Response Rate
80%
72%
60%
40%
20%
0%
Paper Evaluations
(Enrollment = 162)
Electronic Evaluations
(Enrollment = 180)
Ardalan et al. (2007)
100%
Response Rate
80%
69%
60%
?
40%
31%
20%
0%
Paper Evaluations
(Enrollment = 1,415)
Electronic Evaluations
(Enrollment = 1,276)
Whose voice is not heard?
Is there a non-response bias?
Richardson (2005)
“It is therefore reasonable to assume
that students who respond to feedback
questionnaires will be systematically
different from those who do not
respond in their attitudes and
experience of higher education.”
(p. 406, emphasis added)
Layne et al. (1999)
• Statistically significant predictors of
responding to electronic course
evaluations:
– GPA
– class
– subject area
Dommeyer (2002)
• Statistically significant predictors of
responding to electronic course
evaluations:
– none!
• Variables examined:
– gender
– expected grade
– rating of professor’s teaching
Thorpe (2002)
• Statistically significant predictors of
responding to electronic course
evaluations:
– final grade
– gender
– GPA
Avery et al. (2006)
• Statistically significant predictors of
responding to electronic course
evaluations:
– anticipated final grade
– gender
– race/ethnicity
– class size
Conclusion
• There is a fairly consistent,
documented history of bias in
response rates, resulting in some
groups being under-represented
Are paper forms biased?
Perhaps, but the response rates are much
higher, so whatever bias exists is not as
problematic as with electronic forms that
yield much lower response rates
Are the averages different
with fewer responses?
Does an electronic format result in higher
or lower overall average ratings?
Conclusion
• Some studies show that electronic
evaluations result in higher overall
averages, some lower, and some not
statistically different than paperbased forms
Responses from
Survey of Teaching Faculty
February 4 - 13, 2009
Procedure
Wednesday, February 4:
Survey opened; e-mail invitation sent to
all teaching faculty
Monday, February 9:
Reminder announcement in Senate
Wednesday, February 11:
E-mail sent to all department chairs
Friday, February 13:
Survey closed
Survey Responses
• Number of respondents: 178
Respondents’ Division
Behavioral
and Applied
Science, 45%
Science and
Social
Science, 55%
Faculty Rank of Respondents
Adjunct
20%
Lecturer
13%
Assistant
34%
Associate
22%
Full or
Distinguished
10%
0%
10%
20%
30%
40%
Respondents’ Length of Service
< 1 year
16%
1-2 years
5%
23%
3-5 years
6-10 years
33%
11-20 years
12%
> 20 years
11%
0%
10%
20%
30%
40%
1. Are you in favor or opposed to the
College conducting all course
evaluations online?
60%
49%
50%
40%
35%
30%
20%
15%
10%
0%
In Favor
Opposed
No Preference
1. Are you in favor or opposed to the
College conducting all course
evaluations online?
70%
58%
60%
50%
42%
40%
30%
20%
10%
0%
In Favor
Opposed
2. How strongly do you feel about the
College conducting all course
evaluations online?
Wholeheartedly
in favor
21%
Conditionally in
favor
19%
Neutral
12%
Opposed with
reservations
21%
Wholeheartedly
opposed
27%
0%
10%
20%
30%
2. How strongly do you feel about the
College conducting all course
evaluations online?
In favor
45%
Opposed
55%
0%
10%
20%
30%
40%
50%
60%
Summary of Written Responses
• Faculty (even some who are in favor
of online evaluations) say they are
“worried” about the following:
– low response rates
– lack of security
– non-discrimination (all instructors get
rated the same)
– biased sample (because of who might
not respond)
Summary of Written Responses,
cont.
• One person reported previous
positive experience with online
evaluations at another institution
Summary of Written Responses,
cont.
• Some faculty who oppose online
evaluations have had experience
with either the pilot project last
summer, online course evaluations
at previous institutions, or other
online aspects of their courses
• Faculty speaking from first-hand
experience explicitly mentioned their
concern about low response rates
Summary of Written Responses,
cont.
• Faculty are concerned about the
emotional/mental state of students
when completing evaluations online
• They also worry about whether
students might be influenced by
others around them at the time
Summary of Written Responses,
cont.
• Overall, the language and tone of
faculty opposed to online evaluations
was far more strongly and
emphatically voiced than the (rather
muffled) approval of those in favor
Summer 2008 Pilot
Response Rates and
Overall Experience
• No summary data available
• Anecdotal data (from the survey and
personal conversations):
Percentage of faculty who participated
in the pilot who are now in favor of
online evaluations: 0%
Percentage of faculty who participated
in the pilot who are now opposed to
online evaluations: 100%
Student Feedback
Committee Conclusions
Data Sources
• Survey of teaching faculty
• Published, peer-reviewed literature
• Consultation with Patty Francis and
Steve Johnson
• Anecdotal evidence from other
institutions
• Local campus experience
Conclusions: Paper Forms
Advantages:
– higher response rate, less likely for bias
in results
– more faculty are confident about
obtaining valid results through this
method
– controlled setting for administration
– students are familiar with the format
Conclusions: Paper Forms
Disadvantages:
– time required to process forms
– delay in receiving results
– use of paper resources
=> Note that none of these
disadvantages is related to the
validity or accuracy of the data
Conclusions: Digital Forms
Advantages:
– results could be delivered to faculty
more quickly
– saves paper and some processing time
Conclusions: Digital Forms
Disadvantages:
– lower response rate
– no good options for incentives
– more likely for bias in results, concerns
about validity
– a majority of faculty have significant
reservations
– concerns among both faculty and
students about security/privacy
Conclusions: Digital Forms
Disadvantages, cont.:
– questions about faculty being able to
opt out
– questions about students being able to
opt out
– student responses can be posted online
for others to see
One Final Consideration
SPI data are currently used to evaluate faculty for:
–
–
–
–
–
merit pay
contract renewal
tenure/continuing appointment
promotion
performance awards
=> If faculty lack confidence in the integrity and
accuracy of course evaluation data, any
decisions that are made on the basis of these
data are likely to be questioned in a way that we
believe is unhealthy for our institution.
Recommendation #1
All course evaluations
should be administered
using paper forms.
We believe the current consensus among
faculty and students will shift at some
point toward favoring an electronic
format. But we are not nearly there yet.
Recommendation #2
Electronic course
evaluations should not
even be an option.
Aggregated results cannot be interpreted
meaningfully (especially if differential
incentives are offered).
EXCEPTION: Distance-learning courses
Recommendation #3
Since significant man-hours are
needed to process course evaluation
forms for our campus, the College
Senate should advocate strongly for
allocating additional (seasonal) help for
processing these forms.
Stay tuned...
... for Part II of our recommendations
regarding changes to the form used for
course evaluation.