Write Grading Rubrics with AI
Create clear, consistent grading rubrics for any assignment in minutes. AI helps you define performance levels, write concrete descriptors, and assign point values so students know exactly what's expected and you can grade faster and more fairly.
Tools You'll Need
MCP Servers for This Scenario
Browse all MCP servers →- 1
Define the Assignment and Criteria
Establish what you're grading and what success looks like before writing a single descriptor. The criteria you choose determine everything else.
I'm a [subject] teacher for [grade level] students. I need to create a grading rubric for the following assignment: Assignment description: [e.g., 'A 5-paragraph persuasive essay arguing for or against school uniforms, minimum 600 words, must include at least 3 pieces of evidence'] Help me design the foundation for this rubric: 1. **Core Criteria**: Identify 4-6 distinct criteria (categories) that should be evaluated in this assignment. For each criterion, explain in one sentence what aspect of quality it measures and why it matters for this assignment type. Don't overlap criteria — each should measure something distinctly different. 2. **Criteria Weighting**: Suggest a point weighting for each criterion that reflects its importance. For example, in a persuasive essay, argument quality should probably weigh more than formatting. Show me two weighting options: one where all criteria are equal, and one that reflects the true priorities of this assignment type. 3. **Performance Levels**: Should this rubric use 4 levels (e.g., Excellent/Proficient/Developing/Beginning) or 3 levels? Make a recommendation based on the assignment type and age group, and explain why. 4. **What 'Excellent' Looks Like**: For each criterion, give me one concrete example of what an excellent student response or product would look like. Be specific — avoid vague phrases like 'shows deep understanding.' 5. **Grader Consistency Issues**: What aspects of this assignment are hardest to grade consistently across different teachers or between students? Flag these so I can write especially precise descriptors for them.
Tip: 4-6 criteria is the sweet spot. Fewer than 4 feels oversimplified; more than 6 makes the rubric hard to use during grading and hard for students to internalize. If you find yourself wanting 8 criteria, look for two that are actually measuring the same thing and merge them.
- 2
Write Performance Level Descriptors
Generate concrete, observable descriptors for each level of each criterion. This is where most rubrics fail — vague language that two teachers interpret differently.
Write detailed performance level descriptors for my [assignment type] rubric for [grade level] [subject] students. Criteria and weights: [paste from Step 1] Performance levels: [e.g., Excellent (4), Proficient (3), Developing (2), Beginning (1)] For EACH criterion, write descriptors at ALL performance levels. Follow these rules: 1. **Be Observable**: Descriptors must describe what you can actually see in the work, not what the student 'understands' or 'knows.' Bad: 'Student demonstrates understanding of thesis statements.' Good: 'Thesis statement clearly states the argument and three supporting reasons in a single sentence.' 2. **Show Gradation**: Each level should clearly differ from the adjacent level. A student and parent should be able to read their work and self-identify which level it falls in without your help. 3. **Avoid Negatives in Lower Levels**: Instead of 'thesis statement is missing or unclear,' write 'thesis statement attempts to state a position but does not specify supporting reasons.' Describe what IS there, not what ISN'T. 4. **Avoid Judgment Words**: Remove subjective terms like 'excellent,' 'poor,' 'good,' and 'bad' from descriptors — these are the level NAMES, not the descriptions themselves. Descriptors should be factual and observable. 5. **Use Parallel Structure**: Each descriptor at the same level across different criteria should feel like it belongs to the same student performance. They should be internally consistent. Criteria to write descriptors for: [List each criterion from Step 1]
Tip: Write the 'Excellent' descriptor first, then the 'Beginning' descriptor, then fill in the middle. It's much easier to describe the extremes than to work level-by-level. Once you have both ends, the middle levels almost write themselves as the midpoint between them.
- 3
Format and Finalize the Rubric
Compile descriptors into a clean rubric table, add point values, and create both a teacher copy and a student-facing version.
Format all the components into a complete grading rubric. I need two versions: **VERSION 1: TEACHER GRADING RUBRIC** Assignment: [assignment name] Grade Level: [grade level] Subject: [subject] Total Points: [total] Due Date: [date] Format as a table: | Criteria | Weight | Excellent ([X] pts) | Proficient ([X] pts) | Developing ([X] pts) | Beginning ([X] pts) | Score | Comments | - Include all criteria rows - Add a 'Teacher Notes' column for each criterion (common errors to watch for, examples of edge cases) - Add a 'Total Score' row at the bottom - Add a 'General Comments' text box **VERSION 2: STUDENT-FACING CHECKLIST** Convert the rubric into a self-assessment checklist students complete BEFORE submitting: - For each criterion, list 3-5 specific things students should check their work for - Write in second person ('Your thesis statement...' not 'The thesis statement...') - Add a 'Self-Score' column students fill in - Add a brief reflection prompt at the bottom: 'One thing I would revise if I had more time: ___' - Keep language appropriate for [grade level] — no jargon **BONUS**: Write a 3-sentence rubric explanation I can read aloud to students when introducing the assignment that captures the most important grading priorities without reading the whole rubric.Tip: Give students the rubric when you assign the work, not when you return it. Research consistently shows that students who see the rubric before starting produce higher-quality work — not because they 'game' the rubric, but because they understand what quality looks like. The rubric is a teaching tool, not just a grading tool.
- 4
Test the Rubric on Sample Work
Validate your rubric by running it against real or hypothetical student samples to catch ambiguities before grading 30 papers.
I want to test my rubric before I start grading. Help me identify problems with it. Here is my complete rubric: [paste your rubric] Please do the following: 1. **Ambiguity Test**: Read each descriptor and flag any language that two reasonable teachers might interpret differently. For each flagged item, suggest a more precise rewrite. 2. **Edge Case Analysis**: Describe 3-4 hypothetical student submissions that would be genuinely difficult to score with this rubric — situations where the descriptors don't clearly map to one level. For each edge case, suggest how the rubric should be updated to handle it. 3. **Consistency Check**: Are the performance levels graduated consistently across all criteria? Is it possible to score Excellent on one criterion but Beginning on another in a way that's realistic? If so, does the rubric handle that combination fairly in the final score? 4. **Student Comprehension Check**: Is any language in the student-facing version likely to confuse [grade level] students? Flag terms that need simplification or explanation. 5. **Grade a Sample**: I'll describe a hypothetical student submission. You score it against my rubric and explain your reasoning for each criterion. This lets me verify the rubric produces the grades I'd intuitively give. Sample submission description: [describe a mid-level student submission in detail]
Tip: Before your first real grading session with a new rubric, grade 5 papers without recording scores. This calibration pass shows you where the rubric is unclear and what the real distribution of work looks like. Then finalize your interpretation of each level and grade everything again consistently.
Recommended Tools for This Scenario
ChatGPT
The AI assistant that started the generative AI revolution
- GPT-4o multimodal model with text, vision, and audio
- DALL-E 3 image generation
- Code Interpreter for data analysis and visualization
Claude
Anthropic's AI assistant built for thoughtful analysis and safe, nuanced conversations
- 200K token context window for massive document processing
- Artifacts — interactive side-panel for code, docs, and visualizations
- Projects with persistent context and custom instructions
Frequently Asked Questions
Should I use a holistic rubric or an analytic rubric?
How do I handle student work that falls between two levels?
How long does it take to grade with a rubric versus without one?
Related Articles
Agent Skills for This Workflow
Get More Scenarios Like This
New AI guides, top MCP servers, and the best tools — curated weekly.