Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University Paper presented at the Learning Communities and Assessment Cultures Conference organised by the EARLI Special Interest Group on Assessment and Evaluation, University of Northumbria, 28-30 August 2002 Abstract In a University programme concerned with learning and teaching in higher education, one assessment component on one of the modules is in the form of a group presentation. The presentation is self, peer and tutor assessed using criteria devised by the presenting group and agreed by the principal module tutor (module convenor). In written feedback given to the presenting group, participants are required to comment on the match with the assessment criteria and the learning outcomes for the module (as identified by the presenting group) as well as providing other more holistic comments. Through an investigation of the last two cohorts of participants, this research considers the outcomes from using this assessment method and the impact it has had on the participants’ thinking and practice. Initially, the study compares the assessments made by the various constituent groups (self, peer and tutor) in terms of both the formative feedback and the grade recommended. It was found that self-assessment grades are sometimes noticeably higher than the peer and tutors grades, which themselves demonstrate a greater degree of convergence. A comparison of the formative comments made by the different constituent groups also explores what was seen as relevant in determining the overall recommended mark and, especially, how participants make the connection between the assessment criteria, the use of learning outcomes in the assessment and feedback process. The prevailing wisdom is that having explicit criteria is good for both students (knowing expectations) and to staff (clarity and consistency in marking etc.). However, there can be a wide latitude in the interpretation of such criteria for markers. A key issue is thus ensuring an understanding of agreed criteria before marking commences. Through follow-up questions and interviews with participating University staff, views are sought on the impact of the process on their educational thinking and own practice. A key question is how self and peer assessment contributes to an understanding of the assumed connectivity between the learning outcomes and the assessment criteria agreed, and the demands of the assessment itself. Also investigated is how the negotiation and agreement on such criteria and outcomes assist both the presenting group and the recipients (acting as peer assessors) in making their judgements. The study postulates that staff new to teaching, and therefore assessment, in a University context require opportunities to “try on” assessment methods in order that their significance and potential may be realised. It also explores the value of direct experience in the use of assessment criteria and of matching assessment with learning outcomes. Indications are that engagement with the issues within an environment that “matters”, i.e. one that is linked to their success on the award, provides useful insights for new lecturers, which may be transferred into their own assessment of students. Introduction Much recent literature on assessment in higher education (HE) gives strong support to the use of both peer and self-assessment (Boud 1995, Brown and Knight 1994, Gibbs 2001, Brown Bull & Pendlebury 1997, Brown and Glasner 1998, Brown & Dove 1990). For Boud, self-assessment is a transferable skill and is a principal part of the student learning experience. “Students will be expected to practice self evaluation in every area of their lives on graduation and it is a good exercise in self-development to ensure that their abilities are extended.” (Brown and Knight 1994) Peer assessment is the assessment of the work of others with equal status and usually has an element of mutuality. Underpinning a peer-assessment process is giving and receiving feedback from which continued reflection and perhaps dialogue may continue. Brown, Bull & Pendlebury (1997) draw a distinction between ‘peer assessment’ and ‘peer marking’, the later being the process by which someone makes an estimate of another’s work, and also of ‘peer feedback marking’ which involves students deriving criteria, developing a peer assessment form, providing anonymous feedback and assigning a global mark. In the Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 1 context of this study, the term “peer-assessment” is more closely allied to Brown et al’s definition of “peer feedback marking”. The use of peer and self-assessment carries a number of perceived advantages: Students have more ownership of the assessment process (i.e. it is not just being “done” to them); It can involve students in devising and understanding assessment criteria and in making judgements; It encourages formative assessment – learning through feedback; It encourages the reflective student (autonomous learner); It has validity - it measures what it is supposed to measure; It can emphasise the process not just the product; It is expected in working situations; It encourages intrinsic rather than extrinsic motivation; It challenges the role of the tutor as the sole arbiter of assessment. Toohey (1996) argues that self (and peer) assessment by students is generally trustworthy provided that the criteria have been made explicit, students have had the opportunity to practice assessment skills, the rating instruments are simple and a second marker is used to moderate the assessment. In essence, if a student understands the learning requirement, the process is managed appropriately with opportunities for giving and receiving feedback, then it is likely to be a positive and constructive process for all concerned. The purpose of this study is to involve the students (called “participants” on this programme) in the assessment process using parallel processes of self and peer-assessment and to support them in understanding what is involved and involving them in the determination of the criteria by which their work will be assessed. Background The University operates a postgraduate course in learning and teaching for its teaching staff. For staff with limited experience of teaching in a higher education environment, participation on the programme may be a condition of their probation. Experienced staff also voluntarily participate in the programme for professional development purposes, as may other professionals teaching in a higher education environment. The Postgraduate Certificate, which is part of a full Masters scheme and is accredited by the Institute for Learning and Teaching (ILT), contains three modules; one centred on teaching and facilitating learning, one on assessment, the third focused on curriculum evaluation and development. One of the programme aims is for participants to experience and examine a variety teaching, learning and assessment methods. The module entitled ‘Managing the Assessment Process’ focuses on the role that assessment plays within the overall learning process by examining assessment purposes and practices within generic and subject disciplines and evaluating them in terms of supporting and enhancing the learning experience of students. Of its two assessment components, one is in the form of a group presentation. This assessment contains a number of features: i) The group is required to identify the assessment criteria by which the presentation will be assessed. ii) The module has six learning outcomes. The group must identify which of the learning outcomes are applicable to their presentation. iii) The 15-minute presentation (with 5 minutes for questions) is tutor, peer and selfassessed using the assessment criteria set by the group and against the identified learning outcomes. Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 2 iv) Assessors (tutor, peer and self) recommend an overall mark (percentage) using the grade descriptors provided in the scheme. v) Assessors provide written feedback to the presenting group in which comments refer to the match with the assessment criteria and learning outcomes, as well as providing general remarks relating to the presentation. vi) The final percentage is the average of all the recommended marks unless the overall average is sufficiently different to the average of the tutor marks i.e. falls into a different grade category. If this is the case, a further moderation by tutors will determine the final mark. Given that the majority of assessors will be peer participants, this variation occurs only when there is a significant difference between the tutor and peer averages. The purposes underpinning what may seem like a fairly complex and elaborate process are specific. Accepting that “assessment is at the heart of students’ experience” (Brown and Knight, 1995) and “If you want to change student learning, then change the method of assessment.” (Brown, Bull & Pendlebury, 1997:7), creating greater student involvement and ownership of the assessment process is intended to sharpen their awareness of the key issues, particularly as participants are involved in assessment design themselves. Allied to this is the importance of learning outcomes and assessment criteria. In recent years, the use of ‘learning outcomes’ have replaced ‘objectives’ as the means of identifying the learning that needs to be demonstrated for credit to be awarded. At the same time, if appropriate judgements are to be made, then the criteria by which this occurs need to be “explicit, available, open to interrogation and shared” (Brown & Glasner 1999:11). Involving participants using the mechanisms described above seeks to provide insights in how the assessment process works (most important on a programme which is preparing them to manage assessment processes themselves) and also to provide a process that has a measure of reliability, validity and is transparent. Organisation of the Presenting Groups Each of the two cohorts had twenty-five participants, the majority of whom have little (<3 years) teaching experience in an HE context. Presenting groups were self-selecting (group size varied from 2 to 5, most commonly 4) and were formed around a declaration of interest in specific topic areas associated with assessment. Participants were encouraged to form groups, which crossed subject disciplines although this was not a requirement. Guidance to assist presentation preparation was provided, in particular, in devising appropriate assessment criteria and the selection of learning outcomes. Presenting groups were required to submit presentation outlines early in the module as well as producing final drafts which identified the topic, a brief description, the assessment criteria (with weightings, if deemed applicable), and the appropriate learning outcomes. The first cohort contained six groups, which presented over two sessions (one group was unable to present on the predetermined day). The second cohort had seven groups, which presented over one session. Research Methodology The process for collecting and analysing data in the study was undertaken in two parts: 1. A comparison of the marks and written feedback produced by each assessing group (tutors, peers, self). Each assessor had anonymously completed a feedback form, which outlined the project title, assessment criteria and learning outcomes for the presentation. 2. A follow-up questionnaire sent to all participants four months later to determine: how the presenting group negotiated and agreed the assessment criteria; how appropriate learning outcomes were identified; the ease (or otherwise) of undertaking the self-assessment using the assessment criteria and learning outcomes agreed by the group; Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 3 the ease (or otherwise) of assessing other groups and using their assessment criteria and match to the identified learning outcomes; the ease (or otherwise) of deciding on an appropriate mark for a presentation, whether their own or others; the usefulness of the written feedback received from all assessing groups; the impact that undertaking the process has had on their own thinking and subsequent practice; Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 4 Their current thinking on the value of: - self-assessment - peer assessment - the use of assessment criteria, learning outcomes and their relationship to the assessment process. Analysis of the Marks awarded Tables 1 & 2 show the award of marks by each assessing group in each of the two cohorts. Group Tutor marks Peer marks Selfassessments Averages (all marks) 1 Average: 58% No: 2 Min/Max: 55-60 Range: 15 Average: 49% No: 3 Min/Max: 46-50 Range: Average: 68% No: 3 Min/Max: 62-75 Range: Average: 48% No: 2 Min/Max: 47-48 Range: 1 Average: 61% No: 4 Min/Max: 55-66 Range: 11 Average: 72% No: 2 Min/Max: 72-72 Range: 0 7.3 Average: 61% No: 18 Min/Max:: 55-80 Range: 25 Average: 59% No: 3 Min/Max: 58-60 Range: 2 Average: 55% No: 17 Min/Max: 48-70 Range: 12 Average: 56% No: 17 Min/Max: 44-64 Range: 20 Average: 68% No: 4 Min/Max: 65-70 Range: 15 Average: 71% No: 19 Min/Max: 55-83 Range: 28 18.7 Average: 65% No: 4 Min/Max: 60-70 Range: 10 Average: 60% No: 4 Min/Max: 55-65 Range: 10 Average: 54% No: 4 Min/Max: 47-60 Range: 13 Average: 56% No: 5 Min/Max: 51-64 Range: 13 Average: 65% No: 3 Min/Max: 64-66 Range: 2 With peer marks 64% Percentage awarded (postmoderation) 64% 56% 53% 57% 62% 55% 52% 64% 64% 71% 71% 2 3 4 5 6 Mean range 9.6 Table 1:Cohort 2000-01 Group Tutor marks Peer marks 1 Average: 55% No: 4 Min/Max: 45-65 Range: 20 Average: 53% No: 4 Min/Max: 42-58 Range: 16 Average: 62% No: 4 Min/Max: 57-71 Range: 14 Average: 69% No: 4 Min/Max: 60-81 Range: 21 Average: 72% No: 4 Min/Max: 64-85 Range: 21 Average: 63% No: 4 Min/Max: 60-66 Range: 6 Average: 71% No: 4 Min/Max: 65-75 Range: 10 15.4 Average: 57% No: 19 Min/Max: 50-65 Range: 15 Average: 58% No: 19 Min/Max: 45-66 Range: 21 Average: 59% No: 19 Min/Max: 38-72 Range: 34 Average: 65% No: 19 Min/Max: 58-91 Range: 33 Average: 71% No: 19 Min/Max: 58-85 Range: 25 Average: 64% No: 19 Min/Max: 50-87 Range: 37 Average: 71% No: 19 Min/Max: 58-100 Range: 42 29.9 2 3 4 5 6 7 Mean range Selfassessments Average: 75% No: 3 Min/Max: 70-79 Range: 9 Average: 65% No: 3 Min/Max: 65-65 Range: 0 Average: 85% No: 4 Min/Max: 78-94 Range: 16 Average: 67% No: 4 Min/Max: 65-70 Range: 5 Average: 70% No: 2 Min/Max: 70-70 Range: 0 Average: 61% No: 4 Min/Max: 57-65 Range: 8 Average: 65% No: 3 Min/Max: 60-70 Range: 10 6.9 Averages (all marks) 59% Percentage awarded (post moderation) 59% 57% 57% 63% 63% 66% 66% 71% 71% 64% 64% 71% 71% Table 2:Cohort 2001-02 Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 5 Tutor averages differ from the overall average mark more noticeably in cohort 1. In this cohort, tutors more often scored lower than peers and self-assessments (except for one noticeable exception where the average tutor mark was much higher). At that time, the results had led to a review amongst tutors as to how they had arrived at their marks and to a revision for the next cohort. In the second cohort, no average tutor marks much differed from the overall average mark so none were altered at moderation. However, tutor marks for cohort 2 display a significantly greater range of marks. This is in some part due to the addition of two new programme tutors who had not previously undertaken assessment on the programme. Differences between tutor and peer assessment is again more evident in the first cohort with tutors marking lower than peers in all but one case (see table 3). Tutor, Peer & Self Marks - Cohort 1 100 90 Percentage 80 70 Max Mean Min 60 50 40 Group 1 Group 2 Group 3 Group 4 Self Tutor Self Peer Tutor Peer Self Tutor Self Peer Tutor Self Peer Tutor Self Peer Tutor 30 Group 5 Group 6 Table 3 Comparisons in cohort 2 show an equal number of tutor averages higher than peer averages as the other way round (see table 4). It is interesting that where the average tutor mark has the greatest difference compared with the average peer mark, it contain s a high tutor score. The reverse is also the case when the tutor score is low. Tutor, Peer & Self Marks - Cohort 2 100 90 Percentage 80 70 Max Mean Min 60 50 40 Group 1 Group 2 Group 3 Group 4 Group 5 Group 6 Self Tutor Peer Peer Self Tutor Peer Self Tutor Self Peer Tutor Self Peer Tutor Self Peer Tutor Self Peer Tutor 30 Group 7 Table 4 Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 6 The peer averages may well reflect a regression to the mean that can occur in a larger group size. However, closer scrutiny of the range of marks demonstrates a significantly increased range in the recommended marks awarded by peers. It is greater than the average range for either tutors or in self-assessment. Furthermore, the overall mean range increases in cohort 2 from 18.7 to 29.9 (see tables 1 & 2). Participants were not asked to justify their marks at the time as this would have been inappropriate and possible detrimental to a legitimate peer-assessment process. However, the preparation of cohort 2 included greater attention to the grade descriptors for the scheme. Although few peer marks fall below the pass mark (40%), peer-assessors seemed more comfortable in using the full marking range. It is perhaps understandable that the range of marks from self-assessors is less. Logically, the number of assessors undertaking self-assessment in any one presentation was only a small subset of all assessors. However, questionnaire and interview responses also indicated that the process of self-assessment was less problematic than peer-assessment. Self-assessors often commented on how they had “run out of time” or not presented points clearly or explicitly enough. But because considerable time and thought had been invested in the preparation of presentations, the presenting group already had an “inner feeling” of the worth of the presentation. It was just a matter of delivering the material and “releasing the potential”. This may provide an indication why average self-assessment marks were consistently higher than average tutor marks, and often the average peer mark, particularly when the tutor or peer marks were in the lower bands (40s & 50s). The presenting group knew what they wanted to say, whereas others just received what was delivered. Apart from one group, the difference between the tutor and self-assessment in cohort 2 is significantly less. Again, additional attention to grade criteria that occurred with this cohort may be a factor. As is demonstrated in table 3, comparison of the average peer and self-assessment marks reveals some unanimity in cohort 1. However, there are noticeable differences in cohort 2 (see table 4) where the self-assessment marks are higher. Responses in questionnaires and interviews highlighted the challenge that peer assessments had to be undertaken “at short notice” and using assessment criteria derived by the presenting group. Interpreting them within the time-pressured context of the presentation was problematic, a comment also made by tutors. Some presenting groups decided to provide a weighting for each of the assessment criteria, which proved helpful for some peer assessors. However, it also proved to be a double-edged sword when criteria given a high weighting were less fulfilled in the presentation or others weighted lower which were strongly evident and could not be given any more credit. Analysis of responses from the questionnaires and interviews In order to elicit feedback to determine impact on thinking and practice, a questionnaire was sent to all participants on the two cohorts. The questionnaires were anonymous but where respondents had given their names, they could be further interviewed if necessary. Responses were received from 30% of participants and in one third of these further issues were discussed in short interviews. Analysis of the responses revealed the following: i) Negotiating and agreeing assessment criteria Responses from both cohorts referred to the increased appreciation of the importance of the role and use of assessment criteria during the assessment. Some responses commented that deriving criteria was crucial to their project. Initially, it had seemed an unnecessary chore, but grew in importance both during preparation as well as in determining how the presentation would be assessed. Identifying priority and weighting was deemed important to some groups, and less so for others. Assessment criteria were deemed crucial for assessing presentations objectively and fairly. Setting one’s own criteria, with guidance and tutor approval, did increase a sense of validity. Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 7 “With hindsight, this task was very helpful in that it made us realise that we needed to think through clearly the objectives of our presentation as the assessment criteria would reflect these objectives.” (Cohort 1 participant) “…It was also a useful way of ensuring that we focussed on what we were trying to do. I don’t think we would have done it if we had not had to be specific about the assessment criteria.” (Cohort 2 participant) Assuring reliability was more problematic with some participants. Interpreting the assessment criteria, making judgements using the grade criteria (used throughout the programme), doing so “on the spot” and to “co-colleagues” were factors identified. However, reliability was also felt to be assured in many responses through use of the guidance provided within the module and in the requirement for tutor ratification prior to the group presentation. This supports arguments made by Toohey (1996) as outlined earlier. Some responses referred to difficulties experienced in using assessment criteria, particularly when weightings were employed. They gave an additional specificity, which may help the presenter(s) and assist with inter-marker reliability; however, they also were felt to be somewhat confining. However, as is testified below, this can have a beneficial effect on educational practice. “In marking, they were useful but perhaps rather limiting. The marks on each criterion did not necessarily seem to add up to the overall mark which a presentation deserved. I have since prepared a marking sheet and found that it took a great deal of thought to produce a weighting of criteria which was appropriate. Even then, the overall mark in a few cases was different to that which I would have given without marking so carefully against the criteria and their individual weightings. The process has therefore brought home to me the amount of effort and thought needed to weight criteria appropriately, which certainly ensured that I was aware of the pitfalls when preparing mark scheme and I believe that the scheme was much better and more carefully considered than it would have been had I not had the relevant experience in MAP.” (Cohort 2 participant) ii) Identifying different learning outcomes Responses indicate that the determination of appropriate learning outcomes was undertaken in group discussions, making use of tutor and written guidance available. However, debates in the groups often centred on issues of coverage and sufficiency. Because the learning outcomes are pre-determined by the module, group discussions often focused, not on whether they were addressed in the presentation, but whether they were addressed sufficiently. Consequently, the presenting group omitted some learning outcomes because they were deemed to only be addressed tangentially. However, evidence also indicates that presentations were further modified to with the intention of better matching the learning outcomes. It may be argued that attention to learning outcomes at this juncture had a beneficial effect on the planning and delivery of the presentation. This is borne out in questionnaire and interview feedback. Many respondents noted how, in hindsight, it had helped them to focus on what was key about the module. “I found this to be valuable in that assessing other groups and our own group in the extent to which they/we met the learning outcomes provided me with a greater awareness and understanding of the learning outcomes. In turn this helped our group to focus more on trying to meet the learning outcomes when doing our group presentation.” (Cohort 1 participant) Matching the presentation to the learning outcomes did cause some difficulties. For some participants, learning outcomes were too vague and required greater clarity before they could be effectively matched. This provides a timely reminder to ensure that learning outcomes are produced with the intended students in mind rather than other readers e.g. QAA, programme approval boards etc. iii) Usefulness of feedback In conventional tutor-led assessment, feedback, however comprehensive, mostly is in the hands of one or two people. This tutor, peer and self-assessment process provided in excess of twenty five sets of feedback. Given that assessors are asked to provide separate comments which relate to the match with assessment criteria, learning outcomes and other Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 8 more holistic remarks (as well as giving a recommended mark), each presenting group may receive over a hundred pieces of feedback. The sheer volume of response is considerably more than a student might normally expect to receive. Furthermore, comments reflect the range of recommended marks submitted by the various assessors. Questionnaire and interview respondents were overwhelmingly struck by the range of feedback received and of the informative opportunities it provided. Only one participant felt the self-assessment comments to be of little use. For a few, the discrepancy and apparent contradictions in some feedback made it difficult to determine what was appropriate. However, for the majority this variation, combining a mix of summative comment and formative advice, was welcomed. “I found the feedback from each of the tutors, peers and self-assessments very useful and informative. Whilst I have been used to getting tutor feedback in assessments I have done before, this was the first assessment I had undertaken where there was peer and selfassessment feedback as well. It was interesting to compare the feedback from each of the three groups and there appeared to be a great deal of similarity in the feedback which I felt increased my confidence in the use of self and peer assessment.” (Cohort 1 participant) “Overall, the feedback was very useful, and the range of comments interesting. It provided as much insight into different approaches to the marking process, and different perceptions of the same work, as it did into the strengths and weaknesses of the presentation itself. I was particularly interested to note the different approaches taken to the various criteria – from using them as a checklist, with comments doing no more than ‘ticking the boxes’, through to being virtually disregarded in the comments, or used more as a touchstone for assessment than a strict marking guide. (I suspect that the latter is closer to how most of us generally mark work!) (Cohort 2 participant) iv) Continuing views on peer and self-assessment It is evident that peer and self-assessment are two assessment sources that, hitherto, had been little used by most participants and certainly in the formal assessment of their own work. Creating an opportunity to “try on” these ideas in an arena “that counts” has, for many, provided a genuine reflective opportunity for their own understanding and to apply it in their own teaching. The following participant comments exemplify this: “In the past I had not really recognised the value of self-assessment and I had thought that self assessment would be biased resulting in students awarding the maximum marks possible in their own assessment. However my experience of self-assessment has helped me appreciate its value”. (Cohort 1 participant) “There is a lot a value in self-assessment. Initially, it is a sort of introverted experience but becomes one where you start to see yourself as others might see you. …….Self-assessment teaches one to be more critical of oneself.” (Cohort 2 participant) “I have particularly found peer assessment very valuable. I have now implemented peer assessment (moderated by tutor assessment) in one of the modules I teach on and have found that peer assessment certainly enhances student learning in module. By having peer assessment, students seem clearer and more focused on understanding the assessment criteria in the module. In turn this helps them to try and focus on meeting the assessment criteria when they themselves are being assessed.” (Cohort 1 participant) The last of these adds further evidence to the view that modelling practice can influence understanding and practice in transferable contexts. It provides the opportunity to experience what is really involved and to appreciate the considerations that need to be taken into account. At the same time, some concerns were also expressed. The preparation time required to induct and prepare participants was noted. Other research (Brown & Knight 1994, Jordan 1999, Roach 1999) also comment on this. However, quantity of time needed should not be confused with the quality of the experience. If students (participants) are to engage in the process of peer & self-assessment and to fully appreciate the interrelationship of assessment criteria and learning outcomes in the assessment process, then adequate time and preparation is needed for what should be highly transferable learning. Participant concerns also focused on any potential bias or lack of objectivity that may occur through a peer or self-assessment process. Similar arguments have been made in other Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 9 research (Brown. Bull and Pendlebury 1997, Falchikov 1986, 1996, Boud et al 1997). It is acknowledged that these participant groups were mostly co-colleagues working in a supportive environment. Questionnaire responses referred to “giving the benefit of the doubt”. Other commented on being influenced more by the quality of the presentation rather than by its content. As has previously been mentioned, peer and self-assessment were overall higher than the tutor averages. However, there were some notable exceptions and overall differences were not always significant. Students can be highly critical of themselves and their peers and this was reflected in some quite frank feedback (both positive and negative) in written feedback. Drawing some conclusions Much is postulated about the advantages and disadvantages of using self and peerassessment either instead of or as an adjunct to tutor assessment. For Race (2001:5): Students are already self-assessing and peer-assessing quite naturally; Tutor assessment is not sufficiently valid, reliable or transparent; Peer and self-assessment lets students into the assessment culture; The process deepens students’ learning experiences; Students gain much more feedback than would otherwise be possible; It helps students become autonomous learners. Outcomes from this study lend support to many of these claims. Participants initially entered into the assessment process because it was required of them. Experiencing different assessment mechanisms was an intended part of the programme design. Attention to the match with the learning outcomes was an intentional feature to highlight their role in curriculum design and delivery. Requiring groups to devise their own assessment criteria sought to raise their profile and also to explore their role within the assessment process. Analysis of results and feedback from questionnaires and interviews supports the view that these experiences has provided much “food for thought” and is having an impact which extends past the module and into continued thinking and practice. Many participants refer to “a useful learning exercise”, “a positive exercise”, “providing valuable insights” etc. More specific benefits identified include: becoming more aware of the value of using marking criteria; the crucial role of assessment criteria, not just in guiding peer and self-assessment, but also for inter-marker reliability amongst tutors; harnessing peer feedback to augment that offered by tutors; the value of receiving feedback from a range of legitimate sources (including more than one tutor); an increased understanding of the dynamic interface of assessment criteria and learning outcomes in the assessment process. Other more cautionary points were also identified. The pressure placed on presenting groups to negotiate and agree, not only the presentation content, but also the other requirements e.g. setting assessment criteria, identifying appropriate learning outcomes. Evidence suggests that engagement with them enhances the experiences; however it has to be factored into the whole process. The ability for assessors (in particular peer assessors) to be able interpret, at short notice, the criteria selected by a group and be able to relate it to the presentation. The range of marks awarded and comments received from participants indicate that the process of bringing all the necessary considerations together is a complex one requiring careful preparation. Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 10 The legitimacy of peer feedback particularly when it contradicts comments received from other peers, or differs from that of the tutors. For some, the variation is clearly deemed beneficial. However, for others is serves to confuse or even question the reliability of the process, particularly when tutor comments may be seen to conflict. The importance of clearly understanding learning outcomes, particularly when placed alongside the particular focus of the presentation. Learning outcomes are seen as means of making visible what needs to be demonstrated for a ‘pass’ to be awarded. Sometimes, the language used or lack of specificity in their construction is too inaccessible. The sheer time required to undertake all that is required. However, findings to date support the view that engagement with these issues in an environment that “matters”, i.e. one that is linked to their success on the award, provides useful insights for new lecturers, which may be transferred into their own assessment of students. This may best be summarised in the following comments from programme participants: “The process was very useful. It made me focus on what I was doing and make sure it was applicable to the university setting. I would have been less rigorous in my thinking without this.” (Cohort 2 participant) “My positive experience of peer assessment and self-assessment in the MAP module has encouraged me to experiment with the use peer assessment in my own teaching. I have implemented peer assessment as part of the assessment process in a third year undergraduate module I teach. Student feedback suggests that students, in the majority, have also found this to be a positive and valuable experience in enhancing their own learning.” (Cohort 1 participant) “It has provided an opportunity to “feel” an important learning process before introducing it into ones teaching.” (Cohort 2 participant) Bibliography Biggs, J. (1999) Teaching for Quality Learning at University. Buckingham: Open University Press/SRHE. Boud, D. (1995) Enhancing Learning through Self- Assessment. London: Kogan Page. Boud D, Anderson G, Cohen R, and Sampson J (1997) Developing Assessment for Peer Learning, Research and Development on Higher Education, 20, p. 117-125 Brew A (1999) Towards Autonomous Assessment: Using Self-Assessment and Peer Assessment, in Brown, S. and Glasner, A. (eds.) (1999) Assessment Matters in Higher Education - Choosing and Using Diverse Approaches. Buckingham: OUP. Brown, G., Bull, J. and Pendlebury, M. (1997) Assessing Student Learning in Higher Education. London: Routledge. Brown, S and Dove, P (Ed) (1990) Self- and Peer-Assessment SCED Brown, S and Knight P (1994) Assessing Learners in Higher Education Kogan Page Brown, S. and Glasner, A. (eds.) (1999) Assessment Matters in Higher Education - Choosing and Using Diverse Approaches. Buckingham: OUP. Brown, S., Rust, C. and Gibbs, G. (1994) Strategies for Diversifying Assessment in Higher Education. Oxford Brookes University: Oxford Centre for Staff Development. Falchikov, N (1996) Involving Students in Feedback and Assessment in Hounsell D (ed) (1996) Rethinking Assessment, Kogan Page. Falchikov, N (1986) Product comparisons and process benefits of collaborative peer group and self assessments, Assessment and Evaluation in Higher Education Vol 11 No 1, p146-165 Fry, H., Ketteridge, S. and Marshall, S. (1999) A Handbook for Teaching and Learning in Higher Education - Enhancing Academic Practice. London: Kogan Page. Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 11 Gibbs, G. (1992) Assessing more students. Oxford Brookes University: Oxford Centre for Staff Development. Gibbs, G. (1999) Using assessment strategically to change the way students learn. In: S. Brown and A. Glasner (eds.) Assessment matters in higher education-choosing and using diverse approaches. Buckingham: Open University Press. Heywood (2000), Assessment in Higher Education: Student Learning, Teaching, Programmes and Institutions, Jessica Kingsley Publishers, London Jordan, S (1999), Self-Assessment and Peer Assessment, in Brown, S. & Glasner, A. (eds.) Assessment Matters in Higher Education - Choosing and Using Diverse Approaches, Buckingham: OUP Miller A, Imrie, B & Cox K, (1998), Student Assessment in Higher Education, A Handbook for Assessing Performance, Kogan Page Race, P (2001) LTSN Assessment Series No 9: A Briefing on Self, Peer and Group Assessment, LTSN Generic Centre Roach, P (1999), Using Peer assessment and Self-Assessment for the First Time, in Brown, S. & Glasner, A. (eds.) Assessment Matters in Higher Education - Choosing and Using Diverse Approaches, Buckingham: OUP Comparing Peer, Self and Tutor Assessment in a course for University Teaching staff Steve Wilson, London Metropolitan University 12
© Copyright 2026 Paperzz