Rabu, 20 Februari 2013

SCORING RUBRIC / GRADING


                                     

A.    Definition

1.      Scoring
‘A score is a number derived from a measure...’ (Anderson in Sulistyo, 2007). In testing context, this follows then that scoring is a process of utilizing a number to represent the responses made by the test taker.
2.      Grading
Grading rubrics are an effective tool for providing students with your expectations on a given task as well as providing you with a standardized grading mechanism. Rubrics identify the important traits (those to which you assign a score or deduct points) of a task and articulate performance levels. Rubrics attempt to make explicit the implicit criteria by which you grade or assess student performance.
Typically, grading rubrics are created to correspond to an instructor’s student learning objectives. In addition to providing students with clear expectations, a rubric is a valuable assessment tool for faculty wishing to demonstrate direct evidence of students learning.
A grading rubric should be created with the following items:
ð  Specific qualities (or traits) to be assessed
ð  Standards or levels of performance
ð  Specific indicators describing what the various levels of performance look like for each of the qualities being assessed
Be cautious not to make the indicators so broad that they are non-descriptive; they should be as focused as a student learning objective.
Barbara Walvoord and Virginia Anderson’s book, Effective Grading: A Tool for Learning and Assessment has a wealth of information on using rubrics to facilitate the grading process and as a formative assessment tool for both the faculty member and the student.

3.      Rubric
A rubric is a scoring guide that’s intended to help those who must score students’ responses to constructed-response items. The most important component in any rubric is the rubric’s evaluative criteria, which are the factors a scorer considers when determining the quality of a student’s response. For example, one evaluative criterion in an essay-scoring rubric might be the organization of the essay. Evaluative criteria are truly the guts of any rubric, because they lay out what it is that distinguishes students’ winning responses from their weak ones. A second key ingredient in a rubric is a set of quality definitions that accompany each evaluative criterion. These quality definitions spell out what is needed, with respect to each evaluative criterion, for a student’s response to receive a high rating versus a low rating on that criterion. A quality definition for an evaluative criterion like “essay organization” might identify the types of organizational structures that are acceptable and those that aren’t. The mission of a rubric’s quality definitions is to reduce the likelihood that the rubric’s evaluative criteria will be misunderstood.
Finally, a rubric should set forth a scoring strategy that indicates how a scorer should use the evaluative criteria and their quality definitions. There are two main contenders, as far as scoring strategy is concerned. A holistic scoring strategy signifies that the scorer must attend to how well a student’s response satisfies all the evaluative criteria in the interest of forming a general, overall evaluation of the response based on all criteria considered in concert. In contrast, an analytic approach to scoring requires a scorer to make a criterion-by criterion judgment for each of the evaluative criteria, and then amalgamate those per-criterion ratings into a final score (this is often done via a set of predetermined, possibly numerical, rules).

B.     Types
1.      Dichotomous
Dichotomous scoring entails viewing and treating the response as either one of two distinct, exclusive categories. The number utilized in this kind of scoring is 0 (zero) and 1 (one).
2.      Continuous
Continuous scoring views and treat the test taker’s response as being grade in nature. The test taker’s response is considered as having a gradation or degree in it.

C.   Strategy in Scoring
1.         Holistic
Holistic scoring considers the test takers’ response as a whole totality rather than as consisting of fragmented parts. In holistic scoring, the focus is laid on the global quality of the test taker’s response. The benefits of this scoring are quick marking, relatively high inter-rater consistency, easy interpretation due to use of standards, showing the strength of the test takers, and a wide spectrum of use among disciplines. The challenges are absence of diagnostic power, concealing differences across sub-skills within each score, genre-limited application, extensive training on the part of raters.
2.         Primary Trait
Primary trait scoring focuses on one specific type of features or traits that test takers need to demonstrate. The key of primary trait is specificity of discourse to be exhibited by the test takers.
3.         Analytic
The analytic scoring emphasizes on individual points or components of the test taker’s response. In an analytic scoring plan, linguistic and non linguistic features considered important in a discourse become the building blocks of the scheme, and these features are commonly evaluated separately as individual components.

D.      Component of Rubric
A scale of points to be assigned in scoring a piece of work on a continuum of quality. High numbers are typically assigned to the best work.
1.      Descriptors for each level of performance that contain criteria and standards by which the performance will be judged
2.      Indicators are often used in descriptors to provide examples or signs of performance in each level
3.      Criteria that describe the conditions that any performance must meet to be successful
Five categories to consider:
ð  Impact – the success of performance, given the purposes, goals and desired results
ð  Work quality/Craftsmanship – the overall polish, organization, and rigor of the work
ð  Methods – the quality of the procedures and manner of presentation, prior to and during performance
ð  Content – the correctness of the ideas, skills or materials used
ð  Sophistication of the performance – the relative complexity or maturity of the knowledge used. Should describe both strengths and errors (errors should be described particularly in lower levels of performance)
4.      Standards that specify how well criteria must be met

E.     Steps in Designing Rubric
A step-by-step process for designing scoring rubrics for classroom use is presented below.
Information for these procedures was compiled from various sources (Airasian, 2000 & 2001;
Mertler, 2001; Montgomery, 2001; Nitko, 2001; Tombari &Borich, 1999). The steps will be
summarized and discussed, followed by presentations of two sample scoring rubrics.

Step 1: Re-examine the learning objectives to be addressed by the task. This allows you tomatch your scoring guide with your objectives and actual instruction.
Step 2: Identify specific observable attributes that you want to see (as well as those youdon’t want to see) your students demonstrate in their product, process, orperformance. Specify the characteristics, skills, or behaviors that you will belooking for, as well as common mistakes you do not want to see.
Step 3: Brainstorm characteristics that describe each attribute. Identify ways todescribe above average, average, and below average performance for eachobservable attribute identified in Step 2.
Step 4a: For holistic rubrics, write thorough narrative descriptions for excellent workand poor work incorporating each attribute into the description. Describe thehighest and lowest levels of performance combining the descriptors for allattributes.
Step 4b: For analytic rubrics, write thorough narrative descriptions for excellent workand poor work for each individual attribute. Describe the highest and lowestlevels of performance using the descriptors for each attribute separately.
Step 5a: For holistic rubrics, complete the rubric by describing other levels on thecontinuum that ranges from excellent to poor work for the collectiveattributes. Write descriptions for all intermediate levels of performance.
Step 5b: For analytic rubrics, complete the rubric by describing other levels on the continuum that ranges from excellent to poor work for each attribute. Write descriptions for all intermediate levels of performance for each attribute separately.
Step 6: Collect samples of student work that exemplify each level. These will help you in designing scoring rubrics for your classroom.
Step 7: Revise the rubric, as necessary. Be prepared to reflect on the effectiveness of the rubric and revise it prior to its next implementation.


F.     Criteria of A Good Rubric
1.      Are sufficiently generic to relate to general goals beyond an individual performance task but specific enough to enable useful and sound inferences on the task. The word “rubric” derives from the Latin word for “red.” It was once used to signify the highlights of a legal decision as well as the directions for conducting religious services, found in the margins of liturgical books—both written in red.
2.      Discriminate among performances validly, not arbitrarily—by the central features of performance, not by the easiest to see, count, or score.
3.      Do not combine independent criteria in one rubric.
4.      Are based on analysis of many work samples, and based on the widest possible range of work samples—including valid exemplars.
5.      Rely on descriptive language—what quality, or its absence, looks like—as opposed to relying heavily on mere comparatives or value language (e.g. “not as thorough as,” or “excellent product”) to make the discrimination.
6.      Provide useful and apt discrimination to enable sufficiently fine judgments—but not using so many points on the scale as to threaten reliability (typically involving, therefore, 6-12 points on a scale).
7.      Use descriptors that are sufficiently rich to enable student performers to verify their score, accurately self-assess, and self-correct.
8.      Highlight the judging of the “impact” of performance—the effect, given the purpose—as opposed to over-rewarding merely the processes, the formats, or the content used; and/or the good faith effort made.

References:
Airasian, P. W. (2000). Assessment in the classroom: A concise approach (2nd ed.). Boston: McGraw-Hill.
Airasian, P. W. (2001). Classroom assessment: Concepts and applications (4th ed.). Boston: McGraw-Hill.
Mertler, C. A. (2001). Using performance assessment in your classroom. Unpublished manuscript, Bowling Green State University.
Montgomery, K. (2001). Authentic assessment: A guide for elementary teachers. New York: Longman.
Nitko, A. J. (2001). Educational assessment of students (3rd ed.). Upper Saddle River, NJ: Merrill.
Tombari, M. & Borich, G. (1999).Authentic assessment in the classroom: Applications and practice. Upper Saddle River, NJ: Merrill.


Tidak ada komentar:

Posting Komentar