1 / 15

5 Marking , Grading & Reporting

RELANG Relating language examinations to the common European reference levels of language proficiency: promoting quality assurance in education and facilitating mobility. 5 Marking , Grading & Reporting

bob
Download Presentation

5 Marking , Grading & Reporting

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. RELANGRelating language examinations to the common European reference levels of language proficiency: promoting quality assurance in education and facilitating mobility 5 Marking, Grading & Reporting European Centre for Modern Languages and European Commission cooperation on INNOVATIVE METHODOLOGIES AND ASSESSMENT IN LANGUAGE LEARNING

  2. The Process The three steps at this stage are represented below:

  3. Marking • While the expression marking covers all activities by which marks are assigned to test responses, a distinction is often made between the marker, indicating a less skilled role, and the rater, which is a role requiring professional training. • We also distinguish clerical (i.e. human) and machine marking.

  4. Clerical Marking • Clerical markers do not need to be testing experts. • Clerical markers need training, guidance and unambiguous answer keys to do a good job. • The process of marking must be managed so that procedures are carried out according to plan and results are ready when required, but also so that the workload for each marker is never so high as to threaten reliability or accuracy.

  5. Machine Marking • Optical mark reading/optical mark recognition (OMR) technology. OMR is most useful when large numbers of papers are to be marked, and when items are of types that do not require any human judgement. OMR technology can also be used for items which require clerical marking. The clerical marker records marks on the OMR sheet and it is then scanned. • Scanners speed up the process of marking, and reduce human error, but the scanning process may fail to read marks on the paper.

  6. Rating • This is marking where the exercise of trained judgement is necessary, to a much greater degree than in clerical marking. • When judgement is used, a single ‘correct answer’ cannot be clearly prescribed by the exam provider before rating. For this reason, there is more scope for disagreement between judgements than in other kinds of marking, and thus a greater danger of inconsistency, between raters, or in the work of an individual rater.

  7. Rating Scales • This is a set of descriptors which describe performances at different levels, showing which mark or grade each performance level should receive. • Rating scales reduce the variation inherent in the subjectivity of human judgements.

  8. Types of Rating Scales (1) • Holistic or analytic scales: a single mark for a performance can be given using a single scale describing each level of performance. • Relative or absolute scales: scales may be worded in relative, evaluative terms (e.g. ‘poor’, ‘adequate’, ‘good’), or may aim to define performance levels in positive, definite terms. • Checklists: marks based on a list of yes/no judgements as to whether a performance fulfils specific requirements or not.

  9. Types of Rating Scales(2) • Generic or task-specific scales: An exam may use a generic scale or set of scales for all tasks, or provide rating criteria which are specific to each task. A combination of both is also possible. • Comparative or absolute judgement: It is possible to define a scale through exemplar performances; the rater’s task is to say whether a performance is lower, higher or the same in relation to one or more exemplars. A mark is thus a ranking on a scale, e.g. in terms of CEFR levels.

  10. Rating Process Ratersmust have a shared understanding of the standard. The basis of this shared understanding is shared examples of performance. For small-scale exams a group of raters may arrive at a shared understanding through free and equal discussion. For large-scale exams the standard must be stable and meaningful: experienced examiners with authority to communicate the standard to newcomers.

  11. Rater Training Training should proceed through a series of steps from more open discussion towards independent rating, where the samples used relate to the exam being marked: • guided discussion of a sample, through which markers come to understand the level • independent marking of a sample followed by comparison with the pre-assigned mark and full discussion of reasons for discrepancies • independent marking of several samples to show how close markers are to the pre-assigned marks.

  12. Grading • In language tests that report results in terms of CEFR levels, grading needs to be criterion-referenced: performance is evaluated with respect to some fixed, absolute criterion or standard. • An exam may be designed to report over several CEFR levels, or just one. In the latter case, those test takers who achieve the level may be said to have ‘passed’, and the others to have ‘failed. • Identifying the score which corresponds to achieving a certain level is called standard setting.

  13. Standard Setting • Different approaches to standard setting apply to performance skills (speaking, writing) and to receptive skills (reading, listening) which are often objectively marked. • Where a test comprises several different skill components a standard must be set for each skill separately, leaving the problem of how to summarise the results.

  14. Reporting of Results • A single result can be reported for each test taker, or a results profile for each test taker, showing performance on each test component. • The first of these is most common: most end users of exam results seem to prefer a simple answer. The second provides a more informative report which might be very useful for some purposes. • A third option is to provide both. The CEFR stresses the importance of reporting profiled scores where possible.

More Related