Rubrics

Nicole Messier, CATE Instructional Designer
June 28th, 2022

WHAT? Heading link

An instructor holds up a grid with marks denoting student test scores.

Rubrics usually consist of a table, grid, or matrix.

Rubrics are criterion-referenced grading tools that describe qualitative differences in student performance for evaluating and scoring assessments. Criterion-referenced grading refers to students being evaluated based on their performance against a set of criteria. Whereas norm-referenced grading refers to students being assessed through the comparison of student performances.

Rubrics usually consist of a table, grid, or matrix that contain information on how students’ learning and performance will be measured. Rubrics can be designed for a specific assessment. For example, a rubric can be used to grade a written assignment in Week 1 of a course. Or rubrics can be designed for a general purpose, like the grading of all the discussion posts or journal entries in an entire course.

Elements of a Rubric

Most rubrics will contain the following elements:

  • Grading criteria
  • Performance levels
  • Weight and scoring
  • Description of grading criteria

These elements along with the number of rows or columns will vary based on the type of rubric you chose to design. Please see the Types of Rubrics section below for more information and examples of these elements in different types of rubrics.

This is an example of an analytic rubric with two criteria and five columns for performance levels.
Grading Criteria Outstanding Proficient Satisfactory Developing Needs Revision
Criterion I — Row for Description of Criterion I Description at highest level of performance Description at B level of performance Description at C level of performance Description at D level of performance Description at lowest level of performance
Criterion Total Points or Percentage — Row for Weight and Scoring Points or Percentage for the Criterion I 100% to A percentage or points for highest level of performance of criterion I B percentage or points for criterion I C percentage or points for criterion I D percentage or points for criterion I 0% to F percentage or points for lowest level of performance of criterion I
Criterion II — Row for Description of Criterion II Description at highest level of performance Description at B level of performance Description at C level of performance Description at D level of performance Description at lowest level of performance
Criterion II Total Points or Percentage — Row for Weight and Scoring - Points or Percentage for the Criterion II 100% to A percentage or points for highest level of performance of criterion II B percentage or points for criterion II C percentage or points for criterion II D percentage or points for criterion II 0% to F percentage or points for lowest level of performance of criterion II

Grading criteria refer to what students will do (performance) and what instructors will measure and score. Grading criteria should have a direct alignment with the learning objectives. This alignment will improve the validity and reliability of the assessment (see the WHY section of this teaching guide for more information on improving validity and reliability). There are two main types of grading criteria: concrete and abstract grading criteria.

Concrete Grading Criteria

Concrete grading criteria are criteria that can be viewed and assessed with less interpretation and subjectivity. Examples include:

  • Content knowledge or declarative knowledge (about a topic or learning objective)
  • Procedural knowledge (knowledge about how to do a task or action)
  • Conditional knowledge (knowledge about why or when to do an action)
  • Art composition
  • Argument with justification or defense
  • Accuracy or correctness
  • Information literacy (supporting ideas with research and creating new information from research)
  • Writing mechanics (spelling, grammar, punctuation, capitalization)

For example, you might develop a rubric or checklist for weekly math assignments that includes grading criteria for procedural knowledge (showing work), conditional knowledge (explaining why they used a formula or operation), and accuracy (correctness of answer).

Abstract Grading Criteria 

Abstract grading criteria are grading criteria that are interpreted and are considered more subjective than concrete grading criteria. Examples include:

  • Critical thinking
  • Creativity
  • Problem-solving skills
  • Decision-making or reasoning skills
  • Communication or expression of ideas
  • Development of new ideas
  • Organization or cohesion of writing

For example, you might develop a rubric for a piece of art that includes concrete grading criteria for procedural knowledge (demonstration of specific technique), composition of the piece, as well as abstract grading criteria for creativity and decision-making skills.

It is important to note that abstract grading criteria can be difficult for students to know what the expectations are and how to demonstrate those expectations. Abstract grading criteria can be described in a rubric to help students understand the expectations.

Rubric performance levels are usually labeled with a column heading that can be a numeric point value, percentage, letter grade, or heading title. For example:

  • 100% – A level of performance could use any of the following terms as a heading: Exemplary, outstanding, distinguished, exceptional, excellent, expert, etc.
  • 80% – B level of performance could use any of the following terms as a heading: Proficient, above average, accomplished, etc.
  • 70% – C level of performance could use any of the following terms as a heading: Satisfactory, competent, average, acceptable, etc.
  • 60% – D level of performance could use any of the following terms as a heading: Developing, emerging, approaching, novice, etc.
  • 50% – F level of performance could use any of the following terms as a heading: Beginning, rudimentary, needs revision, no evidence, etc.

The above terms can be used as headings for your rubric columns or as adjectives to describe grading criteria at that performance level. It is recommended to utilize the same column headings for all the rubrics developed for a specific course. For example, if you select “Outstanding” for an A level of performance column heading then you should utilize the same column heading for the A level of performance in all your rubrics.

Rubrics contain descriptions of grading criteria. These descriptions should be aligned to the learning objectives being assessed and will support students’ understanding of the assessment expectations. For example, you have the learning objective: Synthesize information and ideas from multiple texts and sources. You label the grading criteria as “Information Literacy” and you describe the grading criteria in an analytic rubric at five performance levels as follows:

  • 100% – A level – Outstanding synthesis of information and ideas from multiple credible sources with exceptional cohesion of information presented.
  • 80% – B level – Concise synthesis of information and ideas from multiple credible sources with cohesion of information presented.
  • 70% – C level – Adequate synthesis of information and ideas from multiple credible courses.
  • 60% – D level – Attempted synthesis of information and ideas and/or missing multiple or credible sources.
  • 50% – F level – Submission did not demonstrate synthesis of information or ideas, missing multiple and/or credible sources, please revise and resubmit.

See the HOW section of this teaching guide to learn tips for writing criteria descriptions.

Types of Rubrics

There are several types of rubrics to choose from based on what you want to measure, how much feedback you want to provide, and how you want to assess performance, including:

  • Single-point rubric
  • Analytic rubric
  • Holistic rubric
  • Checklist

Single-Point Rubric

Single-point rubrics are used to measure learning based on one level of performance for the grading criteria and provide an opportunity for discussions about the strengths and weaknesses of students performance. The single-point rubric has only one column describing a passing level of performance and rows for each grading criterion.

The instructor grades each criterion as either “does not meet the criterion,” “meets the criterion,” or “exceeds the criterion.” And the instructor provides individualized feedback on any criterion that is graded as “does not meet the criterion” or “exceeds the criterion” for students to understand their scores.

This grading tool is called a single-point rubric because it only describes a single level of performance. The single-point rubric provides the opportunity for instructors to give individualized feedback on how students exceeded the criterion or did not meet the criterion.
Grading Criteria Exceeds Criterion Meets Criterion Does Not Meet Criterion
Criterion I Description of criterion I
Total percentage or points for criterion I Less than total percentage or points
Criterion II Description of criterion II
Total percentage or points for criterion II Less than total percentage or points
Criterion III Description of criterion III
Total percentage or points for criterion III Less than total percentage or points

Single-point rubrics will have a total number of points or a percentage for the assessment. And each grading criteria in a single-point rubric will have a point or percentage value. Typically, the “meets the criterion” column will be awarded the total points or an A or B value. For example, the assessment is worth 25 points and contains three criteria.

The total points need to be distributed to each of the criteria (criterion I is worth 5 points, criterion II is worth 10 points, and criterion III is worth 10 points). Students who meet all three criteria will be awarded 25 points.

  • Small class sizes (under 25 students)
  • Involves less time to develop
  • Requires more time to grade and score because students need more personalized feedback to understand their performance and score
  • Supports conversations about performance
  • Can be used for formative and summative assessments
  • Appropriate for on-campus or hybrid course modalities
  • If using video or audio feedback, it can be adapted for online course modalities
  • Best suited for a single user (one instructor)

Grading the Making a Peanut Butter and Jelly Sandwich with a Single-Point Rubric

The following single-point rubric will be used to evaluate the making of a peanut butter and jelly sandwich. Please view the other rubric types to see how they would assess the making of a peanut butter and jelly sandwich. And please visit the Examples and Template section below for more examples.
Grading Criteria Exceed Criterion Meets Criterion Does Not Meet Criterion
Bread Bread is placed together symmetrically and is cut in half diagonally without smashing the bread and appears fluffy.
10 points 10 points 8 or fewer points
Peanut butter Peanut butter is evenly spread on both sides of the sandwich, ensuring that every bite experiences the flavor of the peanut butter.
10 points 10 points 8 or fewer points
Jelly Jelly is evenly spread on one side of the sandwich, ensuring that every bite experiences the flavor of the jelly.
10 points 10 points 8 or fewer points

Analytic Rubric

 

Analytic rubrics are used to evaluate grading criteria separately to provide students with detailed feedback on their performance. The analytic rubric typically has three to five columns to describe performance levels and rows for each grading criterion to be described separately. The instructor grades each criterion at varying levels of performance, and students can read the description to understand their performance and scores.

This assessment tool is called an analytic rubric because it analyzes every grading criterion separately and describes it at each performance level.
Grading Criteria Exemplary Proficient Satisfactory Developing Needs Revision or No Submission
Criterion I Description at highest level of performance Description at B level of performance Description at C level of performance Description at D level of performance Description at lowest level of performance
Criterion I Total Points or Percentage 100% to A percentage or points for highest level of performance of criterion I B percentage or points for criterion I C percentage or points for criterion I D percentage or points for criterion I 0% to F percentage or points for lowest level of performance of criterion I
Criterion II Description at highest level of performance Description at B level of performance Description at C level of performance Description at D level of performance Description at lowest level of performance
Criterion II Total Points or Percentage 100% to A percentage or points for highest level of performance of criterion II B percentage or points for criterion II C percentage or points for criterion II D percentage or points for criterion II 0% to F percentage or points for lowest level of performance of criterion II
Criterion III Description at highest level of performance Description at B level of performance Description at C level of performance Description at D level of performance Description at lowest level of performance
Criterion III Total Points or Percentage 100% to A percentage or points for highest level of performance of criterion III B percentage or points for criterion III C percentage or points for criterion III D percentage or points for criterion III 0% to F percentage or points for lowest level of performance of criterion III

Analytic rubrics will have a total number of points or a percentage for the assessment. And each grading criteria will have a point or percentage value. For example, the assessment is worth 25 points and contains three criteria. The total points need to be distributed to each of the criteria (criterion I is worth 5 points, criterion II is worth 10 points, and criterion III is worth 10 points). Next, the grading criteria points are broken down further by performance level in an analytic rubric.

  • Criterion I is worth 5 points – the highest level is worth 5 points (100%), the next level is worth 4 points (80%), and the last level is worth 3 points (60%).
  • Criterion II is worth 10 points – the highest level is worth 10 points (100%), the next level is worth 8 points (80%), and the last level is worth 6 points (60%).
  • Criterion III is worth 10 points – the highest level is worth 10 points, the next level is worth 8 points (80%), and the last level is worth 6 points (60%).
  • All class sizes
  • Involves more time to develop
  • Requires less time to grade and score (if the scorer is familiar with the rubric)
  • Can be used for formative and summative assessments
  • Provides more descriptive feedback in a formative assessment to help students improve performance
  • Appropriate for any course modality
  • Should be used in online asynchronous course modalities to support student understanding of expectations
  • Utilized by multiple instructors and/or TAs

Grading the Making a Peanut Butter and Jelly Sandwich with An Analytic Rubric

The following analytic rubric will be used to evaluate the making of a peanut butter and jelly sandwich. Please view the other rubric types to see how they would assess the making of a peanut butter and jelly sandwich. And please visit the Examples and Template section below for more examples.
Grading Criteria Exemplary Proficient Satisfactory Developing Needs Revision or No Submission
Bread Bread is placed symmetrically together, cut in half along the diagonal, and shows no evidence of the bread being smashed and appears fluffy. Bread is placed symmetrically together, cut in half, and shows no evidence of the bread being smashed. Bread is placed symmetrically together, cut in half, and shows minor evidence of the bread being smashed. Bread is placed symmetrically together, cut in half, and shows evidence of the bread being smashed. Bread is placed together (one top to one bottom piece), and/or missing cut in half, and significant evidence of the bread being smashed.
10 Points 9 to 10 points 8 points 7 points 6 points 5 or fewer points
Peanut Butter Peanut butter is evenly spread on both sides of the sandwich, ensuring that every bite experiences the flavor of the peanut butter. Peanut butter is evenly spread on both sides of the sandwich. Peanut butter is spread on both sides of the sandwich with minor inconsistencies in coverage and thickness of spread. Peanut butter is spread on both sides of the sandwich with inconsistencies in coverage and thickness of spread or is spread on only one side of the bread. Peanut butter is spread on both sides of the sandwich with numerous inconsistencies in coverage and thickness of spread or is spread on only one side of the bread.
10 Points 9 to 10 points 8 points 7 points 6 points 5 or fewer points
Jelly Jelly is evenly spread on one side of the sandwich, ensuring that every bite experiences the flavor of the jelly. Jelly is evenly spread on one side of the sandwich. Jelly is spread on one side of the sandwich with minor inconsistencies in coverage and thickness of spread. Jelly is spread on one side of the sandwich with inconsistencies in coverage and thickness of spread. Peanut butter is spread on one side of the sandwich with numerous inconsistencies in coverage and thickness of spread.
10 Points 9 to 10 points 8 points 7 points 6 points 5 or fewer points

Holistic Rubric

 

Holistic rubrics are used to evaluate overall competence or ability when grading criteria can’t be separated, or when you want a holistic view of student progress. The holistic rubric typically has around three to five columns to describe performance levels and one row for all the criteria to be described together. The instructor grades the entire assessment at one level of performance and provides the student with individualized feedback identifying what criteria caused their performance to be scored at that level.

This grading tool is called a holistic rubric because it describes all the grading criteria together at each of the performance levels.
Grading Criteria Exemplary Proficient Satisfactory Developing Needs Revision or No Submission
Assessment Description of entire assessment at highest level of performance Description of entire assessment at B level of performance Description of entire assessment at C level of performance Description of entire assessment at D level of performance Description of entire assessment at lowest level of performance
Total Points or Percentage for Assessment 100% to A percentage or points for highest level of performance for the entire assessment B percentage or points for the entire assessment C percentage or points for the entire assessment D percentage or points for the entire assessment 0% to F percentage or points for lowest level of performance for the entire assessment

In a holistic rubric, the grading criteria are not broken down and the weighting occurs in the performance levels. For example, the assessment is worth 25 points and contains five levels of performance (the highest level is worth 25 points (100%), the next level is worth 20 points (80%), the third level is worth 15 points (60%), and the fourth level is worth 10 points (40%), and the last level is worth 9 or less points (0 to 39%).

  • All class sizes
  • Involves less time to develop
  • Requires more time to grade and score because students need more personalized feedback to understand their performance and score
  • Can be used for formative and summative assessments
  • Best suited for summative assessments to measure overall competence or quality of students’ work.
  • Appropriate for any course modality
  • Best suited for a single user (one instructor)

Grading the Making a Peanut Butter and Jelly Sandwich with a Holistic Rubric

The following holistic rubric will be used to evaluate the making of a peanut butter and jelly sandwich. Please view the other rubric types to see how they would assess the making of a peanut butter and jelly sandwich. And please visit the Examples and Templates section below for more examples.
Grading Criteria Exemplary Proficient Satisfactory Developing Needs Revision or No Submission
Peanut Butter and Jelly Sandwich Bread is placed together symmetrically, cut in half diagonally, and shows no evidence of the bread being smashed and appears fluffy. Peanut butter is evenly spread on both sides of the sandwich and jelly is evenly spread between the pb on a piece of bread, ensuring that every bite experiences both flavors. Bread is placed together symmetrically, cut in half diagonally, and shows no evidence of the bread being smashed. Peanut butter is evenly spread on both sides of the sandwich and jelly is evenly spread between the pb on a piece of bread. Bread is placed together symmetrically, cut in half diagonally, and shows minor evidence of the bread being smashed. Peanut butter is spread on both sides of the sandwich and jelly is spread between the pb on a piece of bread with minor inconsistencies in coverage and thickness of spread. Bread is placed together symmetrically, cut in half diagonally, and shows evidence of the bread being smashed. Peanut butter is spread on both sides and jelly is spread on a piece of bread with inconsistencies in coverage and thickness of spread. Bread is placed together (one top to one bottom piece), and/or missing cut in half, and shows significant evidence of the bread being smashed. Peanut butter is spread on one side and jelly on the other allowing the jelly to make the bread soggy, with numerous inconsistencies in coverage and thickness of spread.
Total: 30 Points 30 points 24 points 21 points 18 points 15 points

Checklist

 

Checklists are used to measure criteria that have a correct answer or evidence of correctness or completion (e.g., math, engineering, programming, etc.). The checklist has two columns for performance levels and rows for each grading criterion. Checklist columns are typically labeled with “Yes or No” or “Correct or Incorrect.”

This grading tool is called a checklist because it lists grading criteria separately but does not describe the criteria at different levels of performance. Instructors can simply check whether there is evidence of the grading criteria or no evidence.
Grading Criteria Yes No
Description of Criterion I
Total points or percentage for Criterion I 0 points or percentage
Description of Criterion II
Total points or percentage for Criterion II 0 points or percentage
Description of Criterion III
Total points or percentage for Criterion III 0 points or percentage
Description of Criterion IV
Total points or percentage for Criterion IV 0 points or percentage

Checklists will have a total number of points or a percentage for the assessment. And each grading criteria in a checklist will have a point or percentage value.

For example, the assessment is worth 25 points and contains three criteria. The total points need to be distributed to each of the criteria (criterion I is worth 5 points, criterion II is worth 10 points, and criterion III is worth 10 points).

  • All class sizes
  • Involves less time to develop and grade
  • Provides a breakdown of grading criteria
  • Used for “Yes or No” or “Correct or Incorrect” performance levels
  • Best suited for criteria where there is a correct answer or evidence of correctness or completion: math, engineering, programming, etc.
  • Can be used for formative and summative assessments
  • Appropriate for any course modality
  • Utilized by multiple instructors and/or TAs

Grading the Making a Peanut Butter and Jelly Sandwich with a Checklist

The following checklist will be used to evaluate the making of a peanut butter and jelly sandwich. Please view the other rubric types to see how they would assess the making of a peanut butter and jelly sandwich. And please visit the Examples and Template section below for more examples.
Grading Criteria Yes No
Bread is placed together symmetrically and is cut in half diagonally without smashing the bread and appears fluffy.
10 points 0 points
Peanut butter is evenly spread on both sides of the sandwich, ensuring that every bite experiences the flavor of the peanut butter.
10 points 0 points
Jelly is evenly spread between the pb, ensuring the bread does not become soggy and that every bite experiences the flavor of the jelly.
10 points 0 points

See the HOW section of this teaching guide to learn more about designing rubrics and review examples of rubric types.

WHY? Heading link

Impact of Rubric Use

Research has shown that the use of rubrics has a positive impact on instruction and learning for students and instructors.

Rubrics impact student performance and learning positively (Abdel-Magid, 2020; Hazels, 2020; Nkhoma, 2020) by:

  • Informing students of the expectations for an assignment, including explaining the grading criteria, alignment to learning objectives, and how to meet the performance standards.
  • Improving student motivation, self-efficacy, engagement, and satisfaction.
  • Promoting self-regulation of learning (time and effort) to reach instructors’ expectations.
  • Influencing students’ cognitive and metacognitive performance in the assessment, including the ability to identify strengths and weaknesses in their performance.
  • Providing qualitative feedback to support students’ future learning and performance.

Rubrics also impact instructors’ grading, scoring, and assessment practices positively (Abdel-Magid, 2020; Hazels, 2020; Nkhoma, 2020) by:

  • Providing improved alignment of instructions, expectations, and grading practices, as well as clarity and transparency of the course learning objectives. The rubric design process provides instructors with opportunities to reflect and review the course and learning objectives’ alignment to the assessments and grading criteria.
  • Reducing grading time and overall faculty workload by utilizing the clickable rubrics built in the Blackboard LMS. This reduced workload will allow for more planning of formative assessment and practice opportunities with feedback to improve student outcomes.
  • Improving the consistency, accuracy, and objectivity of grading and scoring will help to prevent or reduce bias in grading by making judgments based on students’ actual performance of the grading criteria. And this consistency, accuracy, and objectivity can potentially reduce students’ questions and arguments about grading, scoring, and fairness.
  • Collecting reliable and valid data for decision-making and continuous quality improvements (see the next section for information on validity and reliability). The consistent use of rubrics will collect data on student performance based on grading criteria aligned to the course and learning objectives for the course.

Improving Validity and Reliability of Assessments

 

Research has shown that the validity and reliability of assessments can be improved through the development and utilization of rubrics.

The validity of an assessment can be described as how well the assessment measures what it was designed to measure. This type of validity is often called face validity or logical validity; in other words, the assessment appears to do what it claims to do (based on face value).

Rubric design improves the alignment of the course and learning objectives with the assessment, and this helps increase the validity of the assessments (Jescovitch, et.al, 2019). Rubric development also improves the alignment of cognitive levels or complexity of the assessment with the course and learning objectives, again improving the validity. Also, the validity of an assessment can be improved by avoiding construct underrepresentation and construct-irrelevant variance through the designing of a rubric.

Construct underrepresentation refers to when an assessment is too narrow and doesn’t include elements of the construct (course or learning objective). The data collected will not have face, content, or construct validity because the assessment omitted aspects (e.g., the assessment doesn’t capture key aspects of the learning objective it was designed to measure). Content validity refers to how well an assessment measures all facets of an item, and how well it represents or gauges the entire domain (e.g., how well the assessment measures the entirety of the learning objectives). Construct validity refers to how well the assessment collects evidence to support interpretations, appropriateness of inferences, and what the data reflects (e.g., does the data collected allow you to make sound decisions about current instruction or continuous quality improvements).

For example, the evaluation of a piece of art might exclude the composition of the artwork or the grading of an oral presentation might miss the communication of the content (Lin, 2020). The rubric design process helps to ensure that no elements are missing, and all aspects of the construct are being evaluated to improve content and construct validity.

Construct-irrelevant variance refers to when an assessment contains excess or uncontrollable variables that distort the data collected (e.g., the assessment contains grading criteria that are not aligned to the task or learning objectives, or assesses skills and knowledge not taught in the course).

For example, an assessment for an oral presentation has grading criteria for costumes or props. This grading criterion isn’t aligned to the assessment and might cause an assessment bias, a grading criterion that unfairly penalizes students because of personal characteristics (Lin, 2020). In the case of the costume or props criteria, more affluent students could afford better costumes or props and may receive a better grade. This bias would cause an unfairness in grading, and data collected wouldn’t have face, content, or construct validity.

It is essential to review your rubrics to ensure that your grading will be focused on the construct (learning objectives) and that it isn’t missing any elements of the construct or adding any excessive or uncontrollable variables that might distort data or cause an assessment bias.

The reliability of an assessment can be described as how well the evaluation and measurement of student performance are consistent and repeatable.

In other words, the consistency of grading and scoring practices from student to student and term to term will influence the reliability of data collected. Rubrics can improve the internal consistency reliability and rater reliability of an assessment.

Internal consistency reliability refers to the interrelatedness of the assessment items and the accuracy of what is measured (e.g., assessments that are directly aligned to the learning objectives would have questions that measure the same construct). Rubric development can enhance the internal consistency reliability of an assessment through the analysis and alignment of learning objectives.

Rater reliability can be described in two sub-categories: intra-rater reliability and inter-rater reliability. Intra-rater reliability refers to how an instructor might grade and score differently based on external circumstances (e.g., one day the instructor is healthy and feeling good and the next day the instructor has a migraine while grading).

Inter-rater reliability refers to how two different instructors might grade and score differently based on what they value (e.g., one instructor might score the organization and technical language in a paper with more weight than another instructor who scores formatting and mechanics with more weight).

Rubric utilization can provide consistent grading criteria that can be repeated under different conditions improving intra-rater reliability (sick instructor) and inter-rater reliability (multiple instructors or TAs). It is important to note that there can still be discrepancies and inconsistencies among multiple instructors or TAs while utilizing a rubric. Please review the HOW section of this teaching guide to learn ways to reduce grading and scoring discrepancies and inconsistencies in order to improve inter-rater reliability.

HOW? Heading link

Selecting the right rubric type for your assessment (and course) is the first step in rubric design. After you decide what type of rubric you want to design, you will need to determine how you will design the rubric.

As an instructor, you can design a rubric, or you can co-construct a rubric. See the below sections for steps on either designing a rubric or co-constructing a rubric with your students.

The following steps will support you as you design a rubric:

  1. Title the rubric using the title of the assessment you want to grade and score.
  2. Identify the grading criteria that you want to measure. Remember your grading criteria should be directly aligned to the course and learning objectives.
  3. Determine how the grading criteria should be assessed: holistically, separately, with a yes/no, etc.
  4. Determine how many levels of performance you will measure students’ learning and performance for the identified criteria and what type of feedback you want to provide students.
    1. Single-point rubric – has one column describing a passing performance (typically an A value) and rows for each grading criterion.
    2. Analytic rubric – has between three to five columns to describe performance levels and rows for each grading criterion separately.
    3. Holistic rubric – has between three to five columns to describe performance levels but with only one row as the criteria are described together.
    4. Checklist – has two columns (for yes and no) and rows for each grading criterion.
  1. Describe the grading criteria (please see Writing Criteria Descriptions below for more information).
  2. Assign points or percentages for each grading criterion (single-point rubric, analytic rubric, or checklist).
  3. Describe levels of performance for each criterion and assign points or percentages for each level of performance (analytic rubric or holistic rubric).
  4. Review your rubric for mutually exclusive language for levels of performance and student-centered language to ensure student understanding of expectations.
  5. Utilize the rubric tool in Blackboard to build a clickable rubric for grading and a viewable rubric for students.
  6. Implement the rubric (without making changes) for the entire term. Reflect on the use of the rubric and identify areas of improvement to make adjustments to criteria, descriptions, or weight for the next term.

For more information on building rubrics in your course site visit the Blackboard Grading and Assessments page on the CATE website to view the Getting Started with Rubrics section.

You can co-construct rubrics with students by first sharing a work sample with them. This work sample could be an exemplar (exemplary work sample) or could be an average work sample (B performance level).

The following steps will support you as you co-construct an analytic rubric with your students:

  1. Share the course and learning objectives that will be measured by the rubric with students.
  2. Share the exemplar (exemplary work sample) or average work sample with students.
  3. Break students into groups either synchronously or asynchronously (using a collaborative tool like Jamboard, Google slides, Padlet, Trello, etc.) and ask them to identify what the potential grading criteria might be.
  4. Bring students back together and remove any redundancies in the grading criteria.
  5. Once you have the grading criteria, you can choose to continue the rubric development with your students or without them.
  1. If you choose to continue with students, then ask students to determine the weight of each criterion for the assessment.
  2. Next, you can break students into groups and have each group describe a different grading criterion at a set number of performance levels (e.g., A, B, C, D).
  3. Collect all the descriptions and create one analytic rubric from each group’s descriptions.
  4. Ask students to review, check for mutually exclusive language, and discuss any changes needed as a class.
  5. Utilize the rubric tool in Blackboard to build a clickable rubric for grading and a viewable rubric for students.
  6. Implement the rubric (without making changes) for the entire term. Reflect on the use of the rubric and identify areas of improvement to make adjustments to criteria, descriptions, or weight for the next term.

For more information on building rubrics in your course site visit the Blackboard Grading and Assessments page on the CATE website to view the Getting Started with Rubrics section.

Tips for Writing Criteria Descriptions

 

You will need to describe the grading criteria, regardless of the type of rubric or checklist you select. Consider the following tips for writing descriptions of the grading criteria.

Criterion descriptions should not contain duplication of criteria within the description; in other words, you should not have two grading criteria that assess the same attribute or element (e.g., critical thinking or formatting, etc.). Each criterion should be specific without duplications in grading.

For example, you have created a checklist that has one criterion for showing work and another criterion for the correct answer. The correct answer criterion should only assess the correctness of the final answer, not the demonstration of the correct problem-solving in the work; this element should be assessed in the showing work criterion.

Adjectives and adverbs can be used to help describe the grading criteria at different performance levels but should be mutually exclusive. Mutually exclusive language means that an adjective used to describe the performance at the highest level shouldn’t be used to describe the performance at the next level. For example, you have used the adjective “thorough” to describe the level of details provided at the exemplary level. So, you should not use the same adjective to describe the proficient level of performance or the subsequent level.

Please note that the following list is not all-encompassing and should be viewed as a starting point for describing grading criteria.

  • 100% – A level of performance could be described with any of the following terms: Exemplary, outstanding, distinguished, exceptional, well-developed, excellent, comprehensive, thorough, robust, expert, extensive, etc.
  • 80% –  B level of performance could be described with any of the following terms: Proficient, above average, accurate, complete, skillful, accomplished, clear, concise, consistent, etc.
  • 70% – C level of performance could be described with any of the following terms: Satisfactory, competent, average, adequate, reasonable, acceptable, basic, sufficient, etc.
  • 60% – D level of performance could be described with any of the following terms: Developing, attempted, emerging, approaching, novice, partial, etc.
  • 50% – F level of performance could be described with any of the following terms: Beginning, rudimentary, rarely, seldom, needs revision, no evidence, etc.

It is important to be consistent with the use of adjectives when developing a rubric or checklist. This consistency will help support student understanding of expectations as well as improve inter-rater reliability if more than one instructor or TA is utilizing the rubric for grading and scoring.

As you begin describing criteria, make sure to focus on the tangible items that can be more objectively measured. For example, if there is a grading criterion for the overall quality of the work, avoid adding subjective elements like “effort.” A student who has not developed the skills yet to perform highly on the assessment might have put in a lot of effort but may have still performed poorly.

Try to use supportive language when describing criteria that help instill a growth mindset in students. And try to avoid negative language that may demotivate students. For example, instead of describing a criterion as “lacking” an element, use the word “missing, developing, or beginning.” Also, consider using terminology like “attempted” at the C or D level; this helps recognize students’ efforts.

Lastly, when describing the grading criteria focus on the quality of the work. Utilize descriptions that help highlight the work’s quality and focus less on quantifying the students’ work. For example, if you have a grading criterion for the mechanics of writing, you can describe it without counting errors in a paper.

  • The Exemplary level of performance could be described as “professional language used in a 2-page report with minimal to no errors in spelling, grammar, punctuation, and capitalization.”
  • The Proficient level of performance could be described as “professional language used in a 2-page report with minor errors in spelling, grammar, punctuation, and capitalization (e.g., misplaced punctuation, homophone errors – to, too, two).”
  • The Satisfactory level of performance could be described as “professional language used in a 2-page report with errors in spelling, grammar, punctuation, and capitalization (e.g., capitalization errors, missing punctuation, grammar, etc.) but still able to understand the point of view.”
  • The Developing level of performance could be described as “attempted professional language in a 2-page report with numerous errors in spelling, grammar, punctuation, and capitalization that distract and cause unreadability.”

If creating an analytic rubric or a holistic rubric, it is recommended to start with the description for the highest level of performance (exemplary level). This level would typically receive an A percentage or point value.

Once you describe the highest level of performance then you can focus on the next level and then the next level, etc. Once you have described all the criteria for the rubric, make sure to check that you are not duplicating criteria and have mutually exclusive language.

Using Exemplars with Rubrics

 

Just as you can use an exemplar (exemplary work sample) to co-construct a rubric with your students; you can also use exemplars with your instructor-designed rubrics. These exemplars help to improve student understanding of the rubric and increase the inter-rater reliability of rubrics when multiple graders are using them (see the WHY section of this guide for more information on reliability).

Not all students will understand the criterion descriptions in your rubric, so by providing an exemplar students can compare the descriptions in the rubric with the work sample. Providing an exemplar will also help other instructors or TAs to understand what the rubric descriptions mean, which will, in turn, improve their consistency in grading and scoring and will positively influence the inter-rater reliability of the assessment.

  • Exemplars can be former students’ work (with permission), published work (with permission), or instructor work.
  • Present features of the exemplar during a class session and deconstruct the rubric using the exemplar to illustrate what the rubric descriptions mean.
  • Think of exemplars as a guide for students to know how to start. Students will understand the expectations of the structure, style, layout, content, etc.
  • Have students use the exemplar and rubric to self-assess their own work. Students will develop the ability to analyze their work to determine strengths and weaknesses and the ability to know how to make it better.

Consider the following questions to improve the validity and reliability of the assessment as you develop and review your rubrics (Lin, 2020):

  • Does the rubric measure the learning objective(s) adequately?
  • Does the rubric include aspects that are irrelevant to the learning objective(s) and/or task?
  • Do the descriptions for grading criteria contain tangible, supportive, and qualitative terms?
  • Does the rubric include any aspects that could potentially reflect assessment biases?
  • Are the grading criteria distinct from one another and use mutually exclusive language?
  • Are the grading criteria weighted appropriately?
  • Are the levels of performance weighted appropriately?
  • Is the rubric paired with an exemplar (exemplary work sample) to support students and multiple instructors’ understanding of expectations?

HOW TO USE/CITE THIS GUIDE Heading link

Creative Commons Attribution-Noncommercial 4.0 international logo of three circles. Circle to the left has
  • This work is licensed under Creative Commons Attribution-NonCommercial 4.0 International.
  • This license requires that reusers give credit to the creator. It allows reusers to distribute, remix, adapt, and build upon the material in any medium or format, for noncommercial purposes only.

Please use the following citation to cite this guide:

Messier, N. (2022). “Rubrics.” Center for the Advancement of Teaching Excellence at the University of Illinois Chicago. Retrieved [today’s date] from https://teaching.uic.edu/resources/teaching-guides/assessment-grading-practices/rubrics/

ADDITIONAL RESOURCES Heading link

Eberly Center. (n.d.). Grading and performance rubrics. Carnegie Mellon University.

Gonzalez, J. (2014). Know your terms: Holistic, analytic, and single-point rubrics. Cult of Pedagogy

Poorvu Center for Teaching and Learning. (n.d.). Creating and using rubrics. Yale University.

Teaching Commons. (n.d.). Rubrics. DePaul University.

REFERENCES Heading link

Abdel-Magid, T., Abdel-Magid, I. (2020). Grading of an assessment rubric. 10.13140/RG.2.2.16717.38887.

Al-Ghazo, A., Ta’amneh, I. (2021). Evaluation and grading of students’ writing: Holistic and analytic scoring rubrics. Journal for the Study of English Linguistics. 9. 77. 10.5296/jsel.v9i1.19060.

Al-Salmani, F., Thacker, B. (2021). Rubric for assessing thinking skills in free-response exam problems. Physical Review Physics Education Research. 17. 10.1103/PhysRevPhysEducRes.17.010135.

Hazels, T., Schutte, K., McVay, S. (2020). Case study in using integrated rubrics in assessment. Journal of Education and Culture Studies. 4. p81. Doi: 10.22158/jecs.v4n3p81. http://dx.doi.org/10.22158/jecs.v4n3p81

Jescovitch, L., Scott, E., Cerchiara, J., Doherty, J., Wenderoth, M., Merrill, J., Urban-Lurain, M., Haudek, K. (2019). Deconstruction of holistic rubrics into analytic rubrics for large-scale assessments of students’ reasoning of complex science concepts. 

Lin, R. (2020). Rubrics for scoring, interpretations and decision-making. 10.4324/9780429022081-5.

Nkhoma, C., Nkhoma, M., Thomas, S., Le, N. (2020). The role of rubrics in learning and implementation of authentic assessment: A literature review. 237-276. 10.28945/4606.

Smyth, P., To, J., Carless, D. (2020). The interplay between exemplars and rubrics.

Tomas, C., Whitt, E., Lavelle-Hill, R., Severn, K. (2019). Modeling holistic marks with analytic rubrics.