The Writing Center • University of North Carolina at Chapel Hill

Comparing and Contrasting

What this handout is about.

This handout will help you first to determine whether a particular assignment is asking for comparison/contrast and then to generate a list of similarities and differences, decide which similarities and differences to focus on, and organize your paper so that it will be clear and effective. It will also explain how you can (and why you should) develop a thesis that goes beyond “Thing A and Thing B are similar in many ways but different in others.”

Introduction

In your career as a student, you’ll encounter many different kinds of writing assignments, each with its own requirements. One of the most common is the comparison/contrast essay, in which you focus on the ways in which certain things or ideas—usually two of them—are similar to (this is the comparison) and/or different from (this is the contrast) one another. By assigning such essays, your instructors are encouraging you to make connections between texts or ideas, engage in critical thinking, and go beyond mere description or summary to generate interesting analysis: when you reflect on similarities and differences, you gain a deeper understanding of the items you are comparing, their relationship to each other, and what is most important about them.

Recognizing comparison/contrast in assignments

Some assignments use words—like compare, contrast, similarities, and differences—that make it easy for you to see that they are asking you to compare and/or contrast. Here are a few hypothetical examples:

  • Compare and contrast Frye’s and Bartky’s accounts of oppression.
  • Compare WWI to WWII, identifying similarities in the causes, development, and outcomes of the wars.
  • Contrast Wordsworth and Coleridge; what are the major differences in their poetry?

Notice that some topics ask only for comparison, others only for contrast, and others for both.

But it’s not always so easy to tell whether an assignment is asking you to include comparison/contrast. And in some cases, comparison/contrast is only part of the essay—you begin by comparing and/or contrasting two or more things and then use what you’ve learned to construct an argument or evaluation. Consider these examples, noticing the language that is used to ask for the comparison/contrast and whether the comparison/contrast is only one part of a larger assignment:

  • Choose a particular idea or theme, such as romantic love, death, or nature, and consider how it is treated in two Romantic poems.
  • How do the different authors we have studied so far define and describe oppression?
  • Compare Frye’s and Bartky’s accounts of oppression. What does each imply about women’s collusion in their own oppression? Which is more accurate?
  • In the texts we’ve studied, soldiers who served in different wars offer differing accounts of their experiences and feelings both during and after the fighting. What commonalities are there in these accounts? What factors do you think are responsible for their differences?

You may want to check out our handout on understanding assignments for additional tips.

Using comparison/contrast for all kinds of writing projects

Sometimes you may want to use comparison/contrast techniques in your own pre-writing work to get ideas that you can later use for an argument, even if comparison/contrast isn’t an official requirement for the paper you’re writing. For example, if you wanted to argue that Frye’s account of oppression is better than both de Beauvoir’s and Bartky’s, comparing and contrasting the main arguments of those three authors might help you construct your evaluation—even though the topic may not have asked for comparison/contrast and the lists of similarities and differences you generate may not appear anywhere in the final draft of your paper.

Discovering similarities and differences

Making a Venn diagram or a chart can help you quickly and efficiently compare and contrast two or more things or ideas. To make a Venn diagram, simply draw some overlapping circles, one circle for each item you’re considering. In the central area where they overlap, list the traits the two items have in common. Assign each one of the areas that doesn’t overlap; in those areas, you can list the traits that make the things different. Here’s a very simple example, using two pizza places:

Venn diagram indicating that both Pepper's and Amante serve pizza with unusual ingredients at moderate prices, despite differences in location, wait times, and delivery options

To make a chart, figure out what criteria you want to focus on in comparing the items. Along the left side of the page, list each of the criteria. Across the top, list the names of the items. You should then have a box per item for each criterion; you can fill the boxes in and then survey what you’ve discovered.

Here’s an example, this time using three pizza places:

As you generate points of comparison, consider the purpose and content of the assignment and the focus of the class. What do you think the professor wants you to learn by doing this comparison/contrast? How does it fit with what you have been studying so far and with the other assignments in the course? Are there any clues about what to focus on in the assignment itself?

Here are some general questions about different types of things you might have to compare. These are by no means complete or definitive lists; they’re just here to give you some ideas—you can generate your own questions for these and other types of comparison. You may want to begin by using the questions reporters traditionally ask: Who? What? Where? When? Why? How? If you’re talking about objects, you might also consider general properties like size, shape, color, sound, weight, taste, texture, smell, number, duration, and location.

Two historical periods or events

  • When did they occur—do you know the date(s) and duration? What happened or changed during each? Why are they significant?
  • What kinds of work did people do? What kinds of relationships did they have? What did they value?
  • What kinds of governments were there? Who were important people involved?
  • What caused events in these periods, and what consequences did they have later on?

Two ideas or theories

  • What are they about?
  • Did they originate at some particular time?
  • Who created them? Who uses or defends them?
  • What is the central focus, claim, or goal of each? What conclusions do they offer?
  • How are they applied to situations/people/things/etc.?
  • Which seems more plausible to you, and why? How broad is their scope?
  • What kind of evidence is usually offered for them?

Two pieces of writing or art

  • What are their titles? What do they describe or depict?
  • What is their tone or mood? What is their form?
  • Who created them? When were they created? Why do you think they were created as they were? What themes do they address?
  • Do you think one is of higher quality or greater merit than the other(s)—and if so, why?
  • For writing: what plot, characterization, setting, theme, tone, and type of narration are used?
  • Where are they from? How old are they? What is the gender, race, class, etc. of each?
  • What, if anything, are they known for? Do they have any relationship to each other?
  • What are they like? What did/do they do? What do they believe? Why are they interesting?
  • What stands out most about each of them?

Deciding what to focus on

By now you have probably generated a huge list of similarities and differences—congratulations! Next you must decide which of them are interesting, important, and relevant enough to be included in your paper. Ask yourself these questions:

  • What’s relevant to the assignment?
  • What’s relevant to the course?
  • What’s interesting and informative?
  • What matters to the argument you are going to make?
  • What’s basic or central (and needs to be mentioned even if obvious)?
  • Overall, what’s more important—the similarities or the differences?

Suppose that you are writing a paper comparing two novels. For most literature classes, the fact that they both use Caslon type (a kind of typeface, like the fonts you may use in your writing) is not going to be relevant, nor is the fact that one of them has a few illustrations and the other has none; literature classes are more likely to focus on subjects like characterization, plot, setting, the writer’s style and intentions, language, central themes, and so forth. However, if you were writing a paper for a class on typesetting or on how illustrations are used to enhance novels, the typeface and presence or absence of illustrations might be absolutely critical to include in your final paper.

Sometimes a particular point of comparison or contrast might be relevant but not terribly revealing or interesting. For example, if you are writing a paper about Wordsworth’s “Tintern Abbey” and Coleridge’s “Frost at Midnight,” pointing out that they both have nature as a central theme is relevant (comparisons of poetry often talk about themes) but not terribly interesting; your class has probably already had many discussions about the Romantic poets’ fondness for nature. Talking about the different ways nature is depicted or the different aspects of nature that are emphasized might be more interesting and show a more sophisticated understanding of the poems.

Your thesis

The thesis of your comparison/contrast paper is very important: it can help you create a focused argument and give your reader a road map so they don’t get lost in the sea of points you are about to make. As in any paper, you will want to replace vague reports of your general topic (for example, “This paper will compare and contrast two pizza places,” or “Pepper’s and Amante are similar in some ways and different in others,” or “Pepper’s and Amante are similar in many ways, but they have one major difference”) with something more detailed and specific. For example, you might say, “Pepper’s and Amante have similar prices and ingredients, but their atmospheres and willingness to deliver set them apart.”

Be careful, though—although this thesis is fairly specific and does propose a simple argument (that atmosphere and delivery make the two pizza places different), your instructor will often be looking for a bit more analysis. In this case, the obvious question is “So what? Why should anyone care that Pepper’s and Amante are different in this way?” One might also wonder why the writer chose those two particular pizza places to compare—why not Papa John’s, Dominos, or Pizza Hut? Again, thinking about the context the class provides may help you answer such questions and make a stronger argument. Here’s a revision of the thesis mentioned earlier:

Pepper’s and Amante both offer a greater variety of ingredients than other Chapel Hill/Carrboro pizza places (and than any of the national chains), but the funky, lively atmosphere at Pepper’s makes it a better place to give visiting friends and family a taste of local culture.

You may find our handout on constructing thesis statements useful at this stage.

Organizing your paper

There are many different ways to organize a comparison/contrast essay. Here are two:

Subject-by-subject

Begin by saying everything you have to say about the first subject you are discussing, then move on and make all the points you want to make about the second subject (and after that, the third, and so on, if you’re comparing/contrasting more than two things). If the paper is short, you might be able to fit all of your points about each item into a single paragraph, but it’s more likely that you’d have several paragraphs per item. Using our pizza place comparison/contrast as an example, after the introduction, you might have a paragraph about the ingredients available at Pepper’s, a paragraph about its location, and a paragraph about its ambience. Then you’d have three similar paragraphs about Amante, followed by your conclusion.

The danger of this subject-by-subject organization is that your paper will simply be a list of points: a certain number of points (in my example, three) about one subject, then a certain number of points about another. This is usually not what college instructors are looking for in a paper—generally they want you to compare or contrast two or more things very directly, rather than just listing the traits the things have and leaving it up to the reader to reflect on how those traits are similar or different and why those similarities or differences matter. Thus, if you use the subject-by-subject form, you will probably want to have a very strong, analytical thesis and at least one body paragraph that ties all of your different points together.

A subject-by-subject structure can be a logical choice if you are writing what is sometimes called a “lens” comparison, in which you use one subject or item (which isn’t really your main topic) to better understand another item (which is). For example, you might be asked to compare a poem you’ve already covered thoroughly in class with one you are reading on your own. It might make sense to give a brief summary of your main ideas about the first poem (this would be your first subject, the “lens”), and then spend most of your paper discussing how those points are similar to or different from your ideas about the second.

Point-by-point

Rather than addressing things one subject at a time, you may wish to talk about one point of comparison at a time. There are two main ways this might play out, depending on how much you have to say about each of the things you are comparing. If you have just a little, you might, in a single paragraph, discuss how a certain point of comparison/contrast relates to all the items you are discussing. For example, I might describe, in one paragraph, what the prices are like at both Pepper’s and Amante; in the next paragraph, I might compare the ingredients available; in a third, I might contrast the atmospheres of the two restaurants.

If I had a bit more to say about the items I was comparing/contrasting, I might devote a whole paragraph to how each point relates to each item. For example, I might have a whole paragraph about the clientele at Pepper’s, followed by a whole paragraph about the clientele at Amante; then I would move on and do two more paragraphs discussing my next point of comparison/contrast—like the ingredients available at each restaurant.

There are no hard and fast rules about organizing a comparison/contrast paper, of course. Just be sure that your reader can easily tell what’s going on! Be aware, too, of the placement of your different points. If you are writing a comparison/contrast in service of an argument, keep in mind that the last point you make is the one you are leaving your reader with. For example, if I am trying to argue that Amante is better than Pepper’s, I should end with a contrast that leaves Amante sounding good, rather than with a point of comparison that I have to admit makes Pepper’s look better. If you’ve decided that the differences between the items you’re comparing/contrasting are most important, you’ll want to end with the differences—and vice versa, if the similarities seem most important to you.

Our handout on organization can help you write good topic sentences and transitions and make sure that you have a good overall structure in place for your paper.

Cue words and other tips

To help your reader keep track of where you are in the comparison/contrast, you’ll want to be sure that your transitions and topic sentences are especially strong. Your thesis should already have given the reader an idea of the points you’ll be making and the organization you’ll be using, but you can help them out with some extra cues. The following words may be helpful to you in signaling your intentions:

  • like, similar to, also, unlike, similarly, in the same way, likewise, again, compared to, in contrast, in like manner, contrasted with, on the contrary, however, although, yet, even though, still, but, nevertheless, conversely, at the same time, regardless, despite, while, on the one hand … on the other hand.

For example, you might have a topic sentence like one of these:

  • Compared to Pepper’s, Amante is quiet.
  • Like Amante, Pepper’s offers fresh garlic as a topping.
  • Despite their different locations (downtown Chapel Hill and downtown Carrboro), Pepper’s and Amante are both fairly easy to get to.

You may reproduce it for non-commercial use if you use the entire handout and attribute the source: The Writing Center, University of North Carolina at Chapel Hill

Make a Gift

How to Teach the Compare and Contrast Essay

Rewards and Resources to Help Students Write Excellent Essays

  • Teaching Resources
  • An Introduction to Teaching
  • Tips & Strategies
  • Policies & Discipline
  • Community Involvement
  • School Administration
  • Technology in the Classroom
  • Teaching Adult Learners
  • Issues In Education
  • Becoming A Teacher
  • Assessments & Tests
  • Elementary Education
  • Secondary Education
  • Special Education
  • Homeschooling
  • M.Ed., Curriculum and Instruction, University of Florida
  • B.A., History, University of Florida

The compare/contrast essay is easy and rewarding to teach for several reasons:

  • It's easy to convince students there is a reason for learning it.
  • You can teach it effectively in a few steps.
  • You can see students' critical thinking skills improve as they learn to write the essay.
  • Once mastered, students feel proud of their ability to systematically compare and contrast two subjects.

Below are the steps you can use to teach the compare/contrast essay. They have been used in regular high school classes where reading levels ranged from fourth to twelfth grade.

  • Discuss practical reasons for comparing and contrasting.
  • Discuss reasons for learning to write about similarities and differences.

Selecting subjects that matter to students is critical for this step. For example, one might be to compare two models of cars and then write a letter to a benefactor who might buy them one. Another would be a store manager writing to a buyer about two products. Academic topics such as comparing two organisms, two wars, two approaches to solving a math problem may also be useful.

  • Show a model compare/contrast essay.

Explain that there are two ways to write the essay but don't go into any detail on how to do it just yet.

  • Explain compare/contrast cue words.

Explain that when comparing, students should mention differences but focus on similarities. Conversely, when contrasting they should mention similarities but focus on differences.

  • Teach students how to use compare/contrast charts.

You should plan to spend a few classes on this. Although it seems simple, students doing it for the first time perform better if they aren't rushed through this step. Working in teams, with a partner, or in a group is helpful.

  • List and model the Writing Den's  cue words  to show similarities and differences.

Many tenth graders have difficulty thinking of these words if this step is skipped. Provide model sentences with these words which they can use until they become comfortable with them.

  • Explain charts showing how to organize compare/contrast  paragraphs  and essays.

Have students write the block style first since it is easier. Students should be told that the block is better to show similarities and the feature-by-feature is better to show differences.

  • Provide guided practice in writing the  first draft .

Guide students through their first essay providing help with an introduction and transition sentences. It is helpful to allow students to use a chart they have completed as a class or one that they have done independently and that you have checked. Do not assume they understand the chart until they have done one correctly.

  • Provide in-class writing time.

By giving in-class writing time, many more students will work on the assignment. Without it, students with little motivation may not write the essay. Walk around asking who needs a little help to get more participation from reluctant learners.

  • Review the steps in the  writing process .
  • Review editing suggestions and give time for  revision .

Explain that after writing their essay, students should edit and revise. They should continue the cycle of editing and revising until they are satisfied with the quality of their essay. Explain the advantages of revising on the computer.

For  editing  tips, check these suggestions for revising drafts  from the University of North Carolina Writing Center.

  • Review the  SWAPS Proofreading Guide  and give students time to proofread their essays.
  • Have students evaluate their peers' essays using a Compare/Contrast Rubric.

Staple a rubric to each essay and have students evaluate them. Be sure to check off on a roster the names of students who turn in essays because they could be stolen during the peer evaluation activity. Consider requiring students who have not finished to submit their essay for peer evaluation after writing " Not Finished"  at the top of their papers. This helps peers recognize that the essay is incomplete. More importantly, taking their paper forces them to participate in the evaluation activity rather than trying to finish the essay in class. Consider giving 25 points each for evaluating three essays and another 25 points for quiet participation.

  • Review the proofreading guide briefly and then devote half a period to proofread one another's essays.

Tell students to read their essay aloud or to have someone else read it to them to catch any errors. Have students proofread several essays and sign their names at the top of the paper: "Proofread by ________."

  • Beef Up Critical Thinking and Writing Skills: Comparison Essays
  • Organizing Compare-Contrast Paragraphs
  • Write a Compare and Contrast Essay
  • Tips to Cut Writing Assignment Grading Time
  • 101 Compare and Contrast Essay Topics
  • Venn Diagrams to Plan Essays and More
  • Writing a Lesson Plan: Closure and Context
  • Compare-Contrast Prewriting Chart
  • How to Teach Topic Sentences Using Models
  • Miss Nelson Is Missing Lesson Plan
  • The Whys and How-tos for Group Writing in All Content Areas
  • Here's What You Need to Know About Lesson Plans
  • Explore and Evaluate Your Writing Process
  • Comparative Words Lesson Plan
  • 501 Topic Suggestions for Writing Essays and Speeches
  • Predictions to Support Reading Comprehension

Compare and Contrast Rubric

Compare and Contrast Rubric

About this printout

Students and teachers can use this rubric when doing writing that compares and contrasts two things, as well as when assessing the writing. 

Teaching with this printout

More ideas to try.

When assigning a compare and contrast writing assignment, students need to be aware of what makes an outstanding written work.  This rubric is a great tool to show students what is expected of them in a concrete way.  Additionally, this rubric will help teachers assess this student writing and inform further instruction.  This Comparison and Contrast Rubric is also a great way to introduce different rubrics that are used to assess student writing on many state tests.

  • Have students work with a partner or a group and grade each others' papers using the rubric (the entire group should grade the same paper at one time).  Students can then compare their scores with those of other students, as well as the author, to see how they are similar and different.  Feedback from group members will help students when editing their own papers and looking to see if they have included what is required of the assignment.
  • Before turning in the writing assignment, have students assess his/her own writing using the rubric.  After the teacher has assessed the assignment, the student and teacher could have a one-on-one conference to discuss how their assessed rubrics differ and what improvements could be made. 
  • Print this resource

Explore Resources by Grade

  • Kindergarten K

Rubric Best Practices, Examples, and Templates

A rubric is a scoring tool that identifies the different criteria relevant to an assignment, assessment, or learning outcome and states the possible levels of achievement in a specific, clear, and objective way. Use rubrics to assess project-based student work including essays, group projects, creative endeavors, and oral presentations.

Rubrics can help instructors communicate expectations to students and assess student work fairly, consistently and efficiently. Rubrics can provide students with informative feedback on their strengths and weaknesses so that they can reflect on their performance and work on areas that need improvement.

How to Get Started

Best practices, moodle how-to guides.

  • Workshop Recording (Fall 2022)
  • Workshop Registration

Step 1: Analyze the assignment

The first step in the rubric creation process is to analyze the assignment or assessment for which you are creating a rubric. To do this, consider the following questions:

  • What is the purpose of the assignment and your feedback? What do you want students to demonstrate through the completion of this assignment (i.e. what are the learning objectives measured by it)? Is it a summative assessment, or will students use the feedback to create an improved product?
  • Does the assignment break down into different or smaller tasks? Are these tasks equally important as the main assignment?
  • What would an “excellent” assignment look like? An “acceptable” assignment? One that still needs major work?
  • How detailed do you want the feedback you give students to be? Do you want/need to give them a grade?

Step 2: Decide what kind of rubric you will use

Types of rubrics: holistic, analytic/descriptive, single-point

Holistic Rubric. A holistic rubric includes all the criteria (such as clarity, organization, mechanics, etc.) to be considered together and included in a single evaluation. With a holistic rubric, the rater or grader assigns a single score based on an overall judgment of the student’s work, using descriptions of each performance level to assign the score.

Advantages of holistic rubrics:

  • Can p lace an emphasis on what learners can demonstrate rather than what they cannot
  • Save grader time by minimizing the number of evaluations to be made for each student
  • Can be used consistently across raters, provided they have all been trained

Disadvantages of holistic rubrics:

  • Provide less specific feedback than analytic/descriptive rubrics
  • Can be difficult to choose a score when a student’s work is at varying levels across the criteria
  • Any weighting of c riteria cannot be indicated in the rubric

Analytic/Descriptive Rubric . An analytic or descriptive rubric often takes the form of a table with the criteria listed in the left column and with levels of performance listed across the top row. Each cell contains a description of what the specified criterion looks like at a given level of performance. Each of the criteria is scored individually.

Advantages of analytic rubrics:

  • Provide detailed feedback on areas of strength or weakness
  • Each criterion can be weighted to reflect its relative importance

Disadvantages of analytic rubrics:

  • More time-consuming to create and use than a holistic rubric
  • May not be used consistently across raters unless the cells are well defined
  • May result in giving less personalized feedback

Single-Point Rubric . A single-point rubric is breaks down the components of an assignment into different criteria, but instead of describing different levels of performance, only the “proficient” level is described. Feedback space is provided for instructors to give individualized comments to help students improve and/or show where they excelled beyond the proficiency descriptors.

Advantages of single-point rubrics:

  • Easier to create than an analytic/descriptive rubric
  • Perhaps more likely that students will read the descriptors
  • Areas of concern and excellence are open-ended
  • May removes a focus on the grade/points
  • May increase student creativity in project-based assignments

Disadvantage of analytic rubrics: Requires more work for instructors writing feedback

Step 3 (Optional): Look for templates and examples.

You might Google, “Rubric for persuasive essay at the college level” and see if there are any publicly available examples to start from. Ask your colleagues if they have used a rubric for a similar assignment. Some examples are also available at the end of this article. These rubrics can be a great starting point for you, but consider steps 3, 4, and 5 below to ensure that the rubric matches your assignment description, learning objectives and expectations.

Step 4: Define the assignment criteria

Make a list of the knowledge and skills are you measuring with the assignment/assessment Refer to your stated learning objectives, the assignment instructions, past examples of student work, etc. for help.

  Helpful strategies for defining grading criteria:

  • Collaborate with co-instructors, teaching assistants, and other colleagues
  • Brainstorm and discuss with students
  • Can they be observed and measured?
  • Are they important and essential?
  • Are they distinct from other criteria?
  • Are they phrased in precise, unambiguous language?
  • Revise the criteria as needed
  • Consider whether some are more important than others, and how you will weight them.

Step 5: Design the rating scale

Most ratings scales include between 3 and 5 levels. Consider the following questions when designing your rating scale:

  • Given what students are able to demonstrate in this assignment/assessment, what are the possible levels of achievement?
  • How many levels would you like to include (more levels means more detailed descriptions)
  • Will you use numbers and/or descriptive labels for each level of performance? (for example 5, 4, 3, 2, 1 and/or Exceeds expectations, Accomplished, Proficient, Developing, Beginning, etc.)
  • Don’t use too many columns, and recognize that some criteria can have more columns that others . The rubric needs to be comprehensible and organized. Pick the right amount of columns so that the criteria flow logically and naturally across levels.

Step 6: Write descriptions for each level of the rating scale

Artificial Intelligence tools like Chat GPT have proven to be useful tools for creating a rubric. You will want to engineer your prompt that you provide the AI assistant to ensure you get what you want. For example, you might provide the assignment description, the criteria you feel are important, and the number of levels of performance you want in your prompt. Use the results as a starting point, and adjust the descriptions as needed.

Building a rubric from scratch

For a single-point rubric , describe what would be considered “proficient,” i.e. B-level work, and provide that description. You might also include suggestions for students outside of the actual rubric about how they might surpass proficient-level work.

For analytic and holistic rubrics , c reate statements of expected performance at each level of the rubric.

  • Consider what descriptor is appropriate for each criteria, e.g., presence vs absence, complete vs incomplete, many vs none, major vs minor, consistent vs inconsistent, always vs never. If you have an indicator described in one level, it will need to be described in each level.
  • You might start with the top/exemplary level. What does it look like when a student has achieved excellence for each/every criterion? Then, look at the “bottom” level. What does it look like when a student has not achieved the learning goals in any way? Then, complete the in-between levels.
  • For an analytic rubric , do this for each particular criterion of the rubric so that every cell in the table is filled. These descriptions help students understand your expectations and their performance in regard to those expectations.

Well-written descriptions:

  • Describe observable and measurable behavior
  • Use parallel language across the scale
  • Indicate the degree to which the standards are met

Step 7: Create your rubric

Create your rubric in a table or spreadsheet in Word, Google Docs, Sheets, etc., and then transfer it by typing it into Moodle. You can also use online tools to create the rubric, but you will still have to type the criteria, indicators, levels, etc., into Moodle. Rubric creators: Rubistar , iRubric

Step 8: Pilot-test your rubric

Prior to implementing your rubric on a live course, obtain feedback from:

  • Teacher assistants

Try out your new rubric on a sample of student work. After you pilot-test your rubric, analyze the results to consider its effectiveness and revise accordingly.

  • Limit the rubric to a single page for reading and grading ease
  • Use parallel language . Use similar language and syntax/wording from column to column. Make sure that the rubric can be easily read from left to right or vice versa.
  • Use student-friendly language . Make sure the language is learning-level appropriate. If you use academic language or concepts, you will need to teach those concepts.
  • Share and discuss the rubric with your students . Students should understand that the rubric is there to help them learn, reflect, and self-assess. If students use a rubric, they will understand the expectations and their relevance to learning.
  • Consider scalability and reusability of rubrics. Create rubric templates that you can alter as needed for multiple assignments.
  • Maximize the descriptiveness of your language. Avoid words like “good” and “excellent.” For example, instead of saying, “uses excellent sources,” you might describe what makes a resource excellent so that students will know. You might also consider reducing the reliance on quantity, such as a number of allowable misspelled words. Focus instead, for example, on how distracting any spelling errors are.

Example of an analytic rubric for a final paper

Example of a holistic rubric for a final paper, single-point rubric, more examples:.

  • Single Point Rubric Template ( variation )
  • Analytic Rubric Template make a copy to edit
  • A Rubric for Rubrics
  • Bank of Online Discussion Rubrics in different formats
  • Mathematical Presentations Descriptive Rubric
  • Math Proof Assessment Rubric
  • Kansas State Sample Rubrics
  • Design Single Point Rubric

Technology Tools: Rubrics in Moodle

  • Moodle Docs: Rubrics
  • Moodle Docs: Grading Guide (use for single-point rubrics)

Tools with rubrics (other than Moodle)

  • Google Assignments
  • Turnitin Assignments: Rubric or Grading Form

Other resources

  • DePaul University (n.d.). Rubrics .
  • Gonzalez, J. (2014). Know your terms: Holistic, Analytic, and Single-Point Rubrics . Cult of Pedagogy.
  • Goodrich, H. (1996). Understanding rubrics . Teaching for Authentic Student Performance, 54 (4), 14-17. Retrieved from   
  • Miller, A. (2012). Tame the beast: tips for designing and using rubrics.
  • Ragupathi, K., Lee, A. (2020). Beyond Fairness and Consistency in Grading: The Role of Rubrics in Higher Education. In: Sanger, C., Gleason, N. (eds) Diversity and Inclusion in Global Higher Education. Palgrave Macmillan, Singapore.

Improving Writing Feedback for Struggling Writers: Generative AI to the Rescue?

  • Original Paper
  • Open access
  • Published: 14 May 2024

Cite this article

You have full access to this open access article

rubric for a comparison essay

  • Anya S. Evmenova 1 ,
  • Kelley Regan 1 ,
  • Reagan Mergen 1 &
  • Roba Hrisseh 1  

41 Accesses

1 Altmetric

Explore all metrics

Generative AI has the potential to support teachers with writing instruction and feedback. The purpose of this study was to explore and compare feedback and data-based instructional suggestions from teachers and those generated by different AI tools. Essays from students with and without disabilities who struggled with writing and needed a technology-based writing intervention were analyzed. The essays were imported into two versions of ChatGPT using four different prompts, whereby eight sets of responses were generated. Inductive thematic analysis was used to explore the data sets. Findings indicated: (a) differences in responses between ChatGPT versions and prompts, (b) AI feedback on student writing did not reflect provided student characteristics (e.g., grade level or needs; disability; ELL status), and (c) ChatGPT’s responses to the essays aligned with teachers’ identified areas of needs and instructional decisions to some degree. Suggestions for increasing educator engagement with AI to enhance teaching writing is discussed.

Similar content being viewed by others

rubric for a comparison essay

Students’ voices on generative AI: perceptions, benefits, and challenges in higher education

rubric for a comparison essay

Artificial Intelligence (AI) Student Assistants in the Classroom: Designing Chatbots to Support Student Success

Examining science education in chatgpt: an exploratory study of generative artificial intelligence.

Avoid common mistakes on your manuscript.

The advances in Generative Artificial Intelligence (generative AI) have transformed the field of education introducing new ways to teach and learn. Its integration is fast growing in all areas of education, including special education (Marino et al., 2023 ). Generative AI has the potential to increase the inclusion of students with disabilities in general education by providing additional assistive supports (Garg and Sharma, 2020 ; Zdravkova, 2022 ). Specifically, large language models like the one used by a popular AI tool, ChatGPT (Chat Generative Pre-trained Transformer) can generate human-like responses to prompts, similar to a conversation. It can facilitate learning for students with and without high-incidence disabilities (e.g., learning disabilities, ADHD) who struggle with writing (Barbetta, 2023 ). While experts continue to investigate the future of writing in the ChatGPT era, it is evident that it will significantly alter writing instruction (Wilson, 2023 ). ChatGPT can support students in choosing a topic, brainstorming, outlining, drafting, soliciting feedback, revising, and proofreading (Trust et al., 2023 ). This tool may also be a helpful resource for teachers in providing feedback on students’ writing. Timely and quality feedback by ChatGPT can encourage the use of higher-level thinking skills while improving the writing process including the planning, writing, and reviewing phases of that process (Golinkoff & Wilson, 2023 ).

Writing Instruction and Feedback for Struggling Writers

The writing process may be challenging for some students for many reasons. For example, planning is the first step of writing, but many students don’t systematically brainstorm. Instead, they move directly into drafting their sentences which may, in turn, be disjointed and not effectively communicated (Evmenova & Regan, 2019 ). Students, particularly those with high-incidence disabilities may not produce text or compose limited text, struggling with content generation, vocabulary, and the organization of ideas (Chung et al., 2020 ). While multilinguism is an asset, we have observed similar challenges with writing among English Language Learners in our research (Hutchison et al., 2024 ). The cognitive demands needed for drafting a response leave many students at no capacity to then edit or revise their work (Graham et al., 2017 ). Therefore, teachers should provide scaffolds to break down the complex process of writing so that it is sequential and manageable, progressing from simple to more complex concepts and skills.

Instruction for struggling writers is typically characterized as systematic and explicit (Archer & Hughes, 2011 ; Hughes et al., 2018 ). In order to provide explicit instruction, teachers should be guided by ongoing student data. Specifically, special and general education teachers of writing should collaboratively, systematically, and continuously monitor and responsively adjust instruction based on student progress (Graham et al., 2014 ). Formative assessments of writing inform feedback that a teacher provides a learner. McLeskey et al., ( 2017 ) describes:

Effective feedback must be strategically delivered, and goal directed; feedback is most effective when the learner has a goal, and the feedback informs the learner regarding areas needing improvement and ways to improve performance… Teachers should provide ongoing feedback until learners reach their established learning goals. (p. 25)

Various formative assessments are available to guide feedback in writing, with rubrics being one frequently used method, which we will explore in the following section.

Supporting Writing by Struggling Writers

School-aged students are required to show progress towards mastery of writing independently in order to be successful at school, future work, and in their personal lives (Graham, 2019 ). Thus, educators continuously look for tools to increase and support learner agency and independence including in writing (Edyburn, 2021 ). Over the past decade, the authors have developed a digital tool to support learner autonomy, access, and independence during essay composition as part of a federally funded, design-based research project referred to as WEGO: Writing Effectively with Graphic Organizers (Evmenova et al., 2018–2023 ). This tool is a technology-based graphic organizer (or TBGO) that embeds numerous evidence-based strategies and universally designed supports for students as well as an analytic rubric for teachers to evaluate student products and providing feedback. A detailed description of the tool can be found elsewhere (students’ features: Evmenova et al., 2020a ; teachers’ features: Regan et al., 2021 ).

The TBGO was developed to support upper elementary and middle school students with and without high-incidence disabilities to compose multiple genres of writing including persuasive (Evmenova et al., 2016 ), argumentative (Boykin et al., 2019 ), and/or personal narrative writing (Rana, 2018 ). The TBGO has also been effectively used by English Language Learners (Day et al., 2023 ; Boykin et al., 2019 ). In addition, it includes a dashboard that allows a teacher or caregiver to personalize instruction: assign prompts and support features embedded in the TBGO. After the student has an opportunity to write independently, the teacher can engage in what we refer to as data-driven decision making (or DDDM; Park & Datnow, 2017 ; Reeves and Chiang, 2018 ).

Teachers’ DDDM

A common formative assessment of writing used in classrooms is a rubric. In order to facilitate the DDDM process within the TBGO, various data are collected by the tool and provided to teachers including final writing product, total time spent actively using the tool, video views and duration, text-to-speech use and duration, audio comments use and duration, transition words use, total number of words, number of attempts to finish. A teacher first evaluates those data as well as student’s writing using a 5-point rubric embedded in the teacher dashboard of the TBGO (a specific rubric is available at  https://wego.gmu.edu ). Based on the rubric, a teacher identifies an area of need organized by phases of the writing process: Planning (select a prompt; select essay goal; select personal writing goal; brainstorm); Writing (identify your opinion, determine reasons, explain why or say more, add transition words, summarize, check your work); and Reviewing: Revise and Edit (word choice, grammar/spelling, punctuation, capitalization, evaluate). Then, a teacher provides specific instructional suggestions when the students’ score does not meet a threshold (e.g., content video models, modeling, specific practice activities). Once teachers select a targeted instructional move that is responsive to the identified area on the writing rubric, they record their instructional decision in the TBGO dashboard. The student’s work, the completed rubric, and the instructional decision is stored within the teacher dashboard. Recent investigations report that teachers positively perceive the ease and usability of the integrated digital rubric in the TBGO (see Regan et al., 2023a ; b ). Although promising, the teachers in those studies used DDDM with only a few students in their inclusive classes.

Efficient and Effective DDDM

The current version of the TBGO relies on teachers or caregivers to score student writing using an embedded rubric and to subsequently provide the student(s) with instructional feedback. In a classroom of twenty or more students, scoring individual essays and personalizing the next instructional move for each student is time consuming, and teachers may not regularly assess or interpret students’ writing abilities in the upper grades, especially (Graham et al., 2014 ; Kiuhara et al., 2009 ). Generative AI or chatbots are arguably leading candidates to consider when providing students with instructional feedback in a more time efficient manner (Office of Educational Technology, 2023 ). For example, automated essay scoring (AES) provides a holistic and analytic writing quality score of students’ writing and a description as to how the student can improve their writing. Recent research on classroom-based implementation of AES suggests its potential; but questions have been raised as to how teachers and students perceive the scores, and how it is used in classroom contexts (Li et al., 2015 ; Wilson et al., 2022 ). Other investigations remark on the efficiency and reliability among AES systems (Wilson & Andrada, 2016 ) and the consistency of scores with human raters (Shermis, 2014 ). More recently, a large-language model (specifically, GPT-3.5 version of ChatGPT) was prompted to rate secondary students’ argumentative essays and chatbot’s responses were compared to humans across five measures of feedback quality (see Steiss et al., 2023 ). Although GPT-3.5 included some inaccuracies in the feedback and the authors concluded that humans performed better than ChatGPT, the comparisons were remarkably close.

A greater understanding of what generative AI tools can do to support classroom teachers is needed. First, leveraging technology, with the use of automated systems, or logistical tools, can potentially improve working conditions for both general and special education teachers (Billingsley & Bettini, 2017 ; Johnson et al., 2012 ). Also, although educators see the benefits of AI and how it can be used to enhance educational services, there is urgent concern about the policies needed around its use and how it is ever evolving. For example, when writing this manuscript, GPT-4 evolved, but at a cost, this latter version may not be widely accessible for educators or students. With the fast adoption of AI, the Office of Educational Technology states that “it is imperative to address AI in education now to realize and mitigate emergent risks and tackle unintended consequences” (U.S. Department of Education, 2023 , p. 3). A first step in addressing AI in education is to understand what AI can do, and how its use supports or hinders student learning and teacher instruction. In this case, we focus on teachers’ writing instruction and feedback.

As we learn more about AI tools, it becomes obvious that AI literacy skills will need to be developed as part of digital skills by both teachers and students (Cohen, 2023 ). The importance of how we use chatbots, how we prompt them, and what parameters we use to direct the responses of chatbots becomes paramount.

Thus, the purpose of this study was to explore feedback and instructional suggestions generated by different AI tools when using prompts providing varying specificity (e.g., a generic 0–4 rating vs. analytic rubric provided) to help guide teachers of writing in their use of these tools. The purpose of including two versions of ChatGPT was not to criticize one and promote the other; but rather to understand and leverage their similarities and differences, given the same prompt. The research questions were:

RQ1: What is the difference between responses generated by GPT-3.5 and GPT-4 given prompts which provide varying specificity about students’ essays?

RQ2: What is the nature of the instructional suggestions provided by ChatGPT for students with and without disabilities and/or ELLs (aka struggling writers)?

RQ3: How does the formative feedback provided by GPT-3.5 and GPT-4 compare to the feedback provided by teachers when given the same rubric?

Data for this study were selected from a large intervention research study (led by the same authors) for a secondary data analysis. Specifically, while previous studies focused on the improvements in students’ writing outcomes (e.g., both quantity and quality of written essays) as well as explored how teachers provide feedback on students’ writing, the unique focus of this paper was on the use of AI to provide writing feedback (something we have not done before). The data included 34 persuasive student essays, a teacher’s completed analytic rubric evaluating the essay, and a teacher’s data-driven decisions with instructional feedback in the area of Writing and Reviewing (essays with the teachers’ DDDM in the area of Planning were excluded). We purposefully selected essays completed by students with various abilities and needs in different grade levels who struggled with writing and needed the TBGO intervention.

Participants

The 34 essays used in this study were written by 21 girls and 13 boys. Students ranged in age 8–13 and were in grades 3–7. The majority (59%) were White, 21% were Hispanic, 3% were African American, and 17% were other. Among the students, 41% were identified with high-incidence disabilities (learning disabilities, ADHD); 24% were English language learners (with a variety of primary languages); and 35% were struggling writers as reported by teachers. Teachers identified struggling writers as those who consistently demonstrated writing performance below grade level expectations (e.g., needing extra support with writing mechanics, cohesive and well-organized ideas).

Study Context

The data used in this study were collected in two separate settings: two inclusive classrooms in a suburban, private day school and an after-school program in a community center serving economically disadvantaged families. The same essay writing procedures were used in both settings. All students were first asked to write a persuasive opinion-based essay in response to one of two prompts validated by previous research (Regan et al., 2023b ). Examples of the prompts included:

Some students go to school on Saturday. Write an essay on whether or not students should go to school on Saturdays.

Some people believe kids your age should not have cell phones. Using specific details and examples to persuade someone of your opinion, argue whether or not kids your age should have cell phones.

After the pretest, students were introduced to the technology-based graphic organizer (TBGO) with embedded evidence-based strategies and supports. The instruction lasted 5–6 lessons. Then students were asked to use the TBGO to practice independent essay writing without any help from the teachers. As the TBGO is a Chrome-based web application and works on any device with a Chrome browser installed, each student used their own device/laptop and individual login credentials to access the TBGO. After completing the independent writing, teachers reviewed students’ products and completed the analytic rubric built into the TBGO’s teacher dashboard. They identified one primary area of need and determined an instructional decision that should take place in order to address the existing area of need. The instructional decisions included whole- and small-group activities (especially in those cases when multiple students demonstrated the same area of need); independent activities (including watching video models embedded within the TBGO); as well as individual teacher-student check-ins to discuss the area of need and future steps. A posttest with the TBGO and a delayed posttest without the TBGO were later administered. The essays used in the current study were from an independent writing phase since those included teachers’ DDDM. On average, essays had 133.44 ( SD  = 57.21; range 32–224) total words written. The vast majority included such important persuasive essay elements such as a topic sentence introducing the opinion, distinct reasons, examples to explain the reasons, summary sentence, and transition words. While this provides some important context, the quantity and quality of students’ writing products is not the focus of the current study and is reported elsewhere (Boykin et al., 2019 ; Day et al., 2023 ; Evmenova et al., 2016 , 2020b ; Regan et al., 2018 , 2023b ).

Data Sources

The existing 34 essays were imported into two different versions of the ChatGPT generative AI: GPT-3.5 version of ChatGPT (free version) and GPT-4 (subscription version). Four different prompts were used in both ChatGPT versions (see Table  1 ). As can be seen in Table  1 , the different prompts included (1) using a specific analytic rubric (when a rubric from the TBGO was uploaded to ChatGPT); (2) asking for a generic 0–4 rating (without any additional specifics regarding scoring); (3) no rubric (asking to identify the area of need without any rubric); (4) no information (asking to provide generic feedback without any information about the student in the prompt). Each prompt type constituted its own GPT chat. Thus, eight sets of responses (or eight different chats) were generated by ChatGPT. A prompt tailored to include the student’s essay as well as the specific student characteristics and the essay topic when applicable (according to the prompt samples presented in Table  1 ) was pasted into the chat. After GPT had a chance to react and provide feedback, the next prompt was pasted into the same chat. Thus, each chat included a total of 34 prompts and 34 GPT outputs. Each chat was then saved and analyzed.

Data Analysis and Credibility

Inductive thematic analysis was used to explore how generative AI can be used to provide writing feedback and guide writing instruction for struggling writers (Guest et al., 2011 ). First, each set of ChatGPT responses (or each GPT chat) was analyzed individually, and reoccurring codes across responses were grouped into categories. The four members of the research team were randomly assigned to analyze two GPT sets each. Each member generated a list of codes and categories within a chat that were the shared with the team and discussed. During those discussions, the patterns within categories were compared across different sets to develop overarching themes in response to RQ1 and RQ2. The trustworthiness of findings was established by data triangulation across 34 writing samples and eight sets of feedback. Also, peer debriefing was used throughout the data analysis (Brantlinger et al., 2005 ).

To answer RQ3, frequencies were used to compare teachers’ and ChatGPT scores on the analytic rubric and suggested instructional decisions. First, two researchers independently compared teachers’ and ChatGPT scores and suggestions. Since the same language from the rubric was used to identify the area of need, the comparisons were rated as 0 = no match; 1 = match. For instructional suggestions, the scale was 0 = no match; 1 = match in concept, but not in specifics; and 2 = perfect match. Over 50% of comparisons were completed by two independent researchers. Interrater reliability was established using point-by-point agreement formula dividing the number of agreements by the total number of agreements plus disagreements and yielding 100% agreement.

RQ1: Differences in AI Responses

In effort to answer RQ1 and explore the differences between responses generated by different ChatGPT versions when given prompts with varying specificity, we analyzed eight sets of responses. While the purpose was not to compare the sets in effort to find which one is better, several patterns have been observed that can guide teachers in using ChatGPT as the starting point for generating writing feedback to their struggling writers. The following are the six overarching themes that emerged from this analysis.

Predictable Pattern of Response

As can be seen in Table  2 , all sets generated excessive amounts of feedback (average length: M  = 383; SD  = 109.7; range 258–581 words) and followed a consistent, formulaic, and predictable pattern of responses across all the writing samples. While the layout and headers used to organize the responses differed across different ChatGPT versions and prompts, the layout and headers were consistent within each set. That said, it was also observed in all ChatGPT sets that the organization and headings found in a response changed slightly towards the end of the run for the 34 writing samples. It is unclear whether this pattern change may happen after a certain number of entries (or writing samples in our case) were entered into the ChatGPT run or if this shift in pattern occurs randomly. Similarly, we also observed that the later responses seemed to be more concise and lacked details which were observed earlier in the same set.

Specific Analytic Rubric

Both GPT-3.5 and GPT-4 provided responses organized into nine categories matching those included in the uploaded rubric. Each category included 1–2 sentences of feedback along with a numerical rating on a 0–4 scale. An overall holistic score was also calculated at the end along with a summary of the student’s overall strengths and weaknesses.

Generic 0–4 Rating

For each writing sample, GPT-3.5 consistently included an evaluation of student writing using four criteria-based categories: Content, Organization, Language Use (punctuation, spelling, and grammar), and Development of Ideas. Two to three bullet points of feedback were listed under each category along with a numeric rating on a 0–4 scale for each. The scale was not defined or explained. An overall holistic score was totaled at the end along with a summary of feedback presented in a bulleted list.

GPT-4’s response to the first writing sample included a definition of what each point on the scale meant (e.g., 4 = writing is clear, well-organized, well-developed, with effectively chosen details and examples presented logically, and few to no errors in conventions). In all consecutive responses, an introductory paragraph identified an overall bold-faced score (0–4) and an overview of what the student did and did not demonstrate in the writing. The following areas of writing were discussed across essays: Organization, Development, Main Idea, Reasons, Examples, Coherence, and Grammar.

Each response in GPT-3.5 began with “One area of need is…” followed by two sentences including how to address the need. Areas of need for instruction identified by ChatGPT included a high frequency of subject-verb agreement as parts of sentence structure (topic sentence and supporting details), followed by transition words or phrases, spelling and grammar conventions, spelling and word choice, capitalization, and punctuation. The second part of the response, titled Instructional Suggestion, provided an instructional strategy for a teacher to use, followed by a model of a ‘revised’ essay using ideas from the student’s response.

GPT-4 provided four consistent parts. First, the response opened with a statement about what the student wrote, a positive affirmation, and an instructional area of writing that could be improved upon. Next, under a header of Instructional Suggestion was a brief description as to what the teacher should do. The third part was a bold-faced, numbered list of steps for implementing that suggestion with bulleted cues underneath. The final part of the response was a ‘revised’ paragraph using the student’s initial writing and addressing the area of need.

GPT-3.5 provided feedback organized in 9 to 11 bolded categories. The sections that were identical for every writing sample included Proofreading; Revising and Editing; Encourage Creativity; and Positive Reinforcement. The sections that were consistent but individualized for each writing sample were Clarity and Organization (including a topic/introductory sentence); Supporting Details; Sentence Structure and Grammar (primarily focus on sentence fragments, punctuation, and capitalization); Conclusion; Vocabulary and Word Choice. Feedback on spelling and transition words/phrases was offered either as separate categories or subsumed under others.

GPT-4’s response could be organized in 3 overarching groups: Positive Reinforcement (including specific praise, affirmation, and creativity); Areas for Improvement (content feedback including idea development; details; coherence; clarity and focus; concluding sentence; and technical feedback including sentence structure; punctuation; grammar; word choice); as well as Instructional Suggestions. A sample revised paragraph was offered at the end with an explanation as to how it showcased the offered suggestions.

Using Specific Language from the Rubric

Both Specific Analytic Rubric sets (using GPT-3.5 and GPT-4) referred exclusively to the uploaded rubric and provided feedback using specific language from the rubric. This included feedback across the nine categories built into the rubric (e.g., the writer clearly identified an opinion, the writer has determined three reasons that support his/her opinion, etc.). Also, both ChatGPT versions used descriptors from the rubric (0 = Try again; 1 = Keep trying; 2 = Almost there; 3 = Good job; 4 = Got it). However, GPT-3.5 did not use any explicit examples from the student’s writing within the feedback and used broad and general statements. GPT-4 referred to the specific content from the students’ writing samples and was more tailored, or individualized (e.g., There are some grammatical and spelling errors present, e.g., "are" instead of "our").

Identifying General, Broad Areas of Need

Feedback in all GPT-3.5 sets (regardless of the prompt) was characterized as using common phrases representing broad areas of need. These phrases were not specifically targeted or explicit. For example, the Generic Rating GPT-3.5 set included such common phrases as “The essay presents ideas and supports them with reasonable detail, but there's room for more depth and elaboration.” or “The content is well-structured and effectively conveys the main points.” Similarly, the No Rubric GPT-3.5 set identified instructional areas of need that were only broadly relevant to the students’ writing. For example, in several instances, our review questioned the prioritization of the writing area identified and if ChatGPT was overgeneralizing areas in need of improvement. Specifically, does two instances of using lowercase when it should be uppercase mean that capitalization should be prioritized over other essential features of writing? Finally, the No Info GPT-3.5 set also used common phrases to describe areas for improvement regardless of the writing sample. For example, there were no difference in ChatGPT’s feedback for a writing essay with eight complete, robust, well-written sentences vs. an incomplete paragraph with just two sentences indicating the lack of targeted and specific feedback.

No Rubric GPT-4 set would start with identifying a broad area of need (e.g., coherence, grammar, development, organization/development of ideas, attention to detail) followed by a more individualized and specific instructional suggestion (as discussed below). The authors acknowledge that this might be explained by the prompt language to identify one area of need.

Focusing on an Individualized, Specific Areas of Need

Like the Specific Analytic Rubric GPT-4 set, the Generic 0–4 Rating GPT-4 set and the No Info GPT-4 sets were observed to include more guidance for the student, drawing on specific areas of an essay to provide corrective feedback. For example, Generic Rating GPT-4 feedback noted, “We should also try to provide more specific examples or explanations for each reason. For example, you mentioned that students get tired – maybe you can explain more about how having some recess can help them feel less tired.” In turn, No Info GPT-4 included detailed feedback focused on specific areas of need such as encouraging more details and clarifications, cohesion and flow, capitalization, spelling, homophones, and punctuation (including avoiding run-on sentences and properly using commas). Word choice, contractions, and conjunctions were often mentioned offering specific revisions. Varying the length and structure of sentences was sometimes suggested for making the writing more engaging and readable.

Misaligned Feedback

While there were some occasional discrepancies in GPT-4 sets, all GPT-3.5 sets appeared to generate feedback that was more misaligned with writing samples. For example, in the Specific Analytic Rubric GPT-3.5 set, a “Good Job” score of 3 was given for the Summary sentence that read, “Moreover, …” and was not a complete sentence. Also, the Generic Rating GPT-3.5 set did not mention any misuse of capitalization despite numerous cases of such misuse. Subject-verb agreement was erroneously mentioned as an area of need for some writing samples for the No Rubric GPT-3.5 set, and then, not mentioned for those students’ writing in which this feedback would be relevant. In the No Info GPT-3.5 set, the topic or introductory sentence was always noted as a suggested area of improvement and a revised sentence was always provided. This was true for cases when a student:

was missing an opinion that aligned with the prompt

had an opinion but did not start it with words “I believe …” (e.g., “Kids should get more recess time.”); and

already had a strong introductory sentence (e.g., “I believe that school starts too early and should begin later in the morning.”).

Starting with Specific Praise/Positive Affirmation

While most ChatGPT feedback included some general praise and affirmation, Generic Rating GPT-4, No Rubric GPT-4, and No Info GPT-4 sets always started with specific positive reinforcement. Unique elements in each essay were praised including conveying personal experiences, having a clear stance or position, and including a variety of reasons, etc.

RQ2: Instructional Suggestions

Instructional suggestions based on the evaluation of student writing was a focus of RQ2. Although we expected the responses from prompts that included specific student characteristics to differentiate the instructional suggestions in some way, this was not the case. In fact, none of the sets provided explicit instructional suggestions aligned with students’ characteristics (e.g., grade, disability, ELL). First, the suggestions for improving the writing of a 3rd grader’s essay were not distinct from those suggestions provided in response to a 7th grader’s writing (in Generic Rating GPT-3.5 and No Rubric GPT-3.5 sets). Also, there were no remarkable differences in the vocabulary used in the feedback for a 3rd grader vs. a 7th grader (in Generic Rating GPT-4 set). Only one set (Generic Rating GPT-4) offered a personalized message in a student-friendly format (without any additional prompting to do so).

Second, student characteristics were merely acknowledged in some sets. For example, Specific Analytic Rubric GPT-3.5 and GPT-4 only noted those characteristics in the summary section at the end of the feedback (e.g., “This is a well-written persuasive essay by your 7th-grade student with ADHD”). This was also observed in responses from the Generic Rating GPT-4 set, as well. For example, “This feedback emphasizes both the strengths of the student’s writing and the areas where improvement can be made, offering encouragement and guidance that is particularly important for a student with ADHD.” Finally, the No Rubric GPT-4 set also gave a mere nod to the additional context (e.g., Given student characteristics…). Although rare, connecting student characteristics with instruction was observed here: “Students with ADHD often struggle with organizing their thoughts in a coherent manner, and the flow of ideas in this student’s paragraph seems a bit disjointed….” Students’ characteristics were not mentioned in any other sets in which student information was included in the prompt (Generic Rating GPT-3.5 and No Rubric GPT-3.5).

Below is the description of how specific, broad, or no instructional suggestions were included in the ChatGPT sets (see Table  2 ).

Specific Suggestions

Specific instructional suggestions were mentioned in Generic Rating GPT-4, No Rubric GPT-4, and No Info GPT-4 sets. At the end of responses for the Generic Rating GPT-4 set, ChatGPT encouraged the teacher to use self-regulatory instructional strategies with students, such as goal setting or self-evaluation. For example, “By involving the student in the refinement of their work and setting goals, you empower them to take ownership of their learning and progression.”

No Rubric GPT-4 responses used such headings as modeling, guided practice, feedback, and independent practice with bulleted ideas under each. The specific suggestions included practice, mini-instructional lessons, engaging activities, peer review, explicit instruction, sentence-building activities, peer review sentence starters, technology such as word processing and online games, the five W’s and How strategy (i.e., a writing strategy that helps students remember to include the answers to “who,” “what,” “where,” “when,” “why,” and “how” in their writing to make their writing complete and clear), a mnemonic referred to as PEE (i.e., Point, Explain, Elaborate; this mnemonic helps students ensure their writing is focused, well-supported, and thoroughly developed), a personal dictionary, interactive editing, and a graphic organizer or outline. When the latter was suggested to support the “coherence” or “development of ideas,” ChatGPT’s response sometimes provided a backwards planning model of what the student’s ideas would look like in an outline format.

Responses of the No Info GPT-4 set included specific and varied instructional suggestions organized by categories: Writing Exercises; Focused Practice; and Revision Work. Suggestions included mini lessons on sentence structure, transition workshops, details workshops, personal experience illustrations, developing ideas workshops, worksheets, grammar lessons, spelling activities, sentence expansion or completion, and editing practice.

Broad Instructional Suggestions

Primarily broad instructional suggestions were offered in the Generic Rating GPT-3.5 and No Rubric GPT-3.5 sets. For example, Generic Rating GPT-3.5 responses had a section with a bulleted list of actionable, instructional suggestions. Each began with a verb (i.e., Work on…; Encourage the student to…; Practice…). It was also not clear if these suggestions were presented in any order of instructional priority. Also, the items included broad ideas that aligned with the student essays but may or may not have aligned with the lowest rated category of writing. Examples of largely vague and broad instructional suggestions recycled throughout the responses in the No Rubric GPT-3.5 set including: “use different types of sentences,” “teach basic spelling rules,” and “use appropriate punctuation.”

Revised Essay

The following three ChatGPT sets included responses with a revised student essay along with a brief explanation of how it was better (even though a revision was not requested in the prompt): No Rubric GPT-3.5, No Rubric GPT-4, and No Info GPT-4 sets. We considered that a model of writing, revised for improvement, was a broad instructional strategy. This is one of many excellent strategies for teaching writing, however, the revisions were often characterized by sophisticated vocabulary and complex elaborations. For example, a student wrote, “To illustrate, when students are hungry it’s hard for them to listen.” And ChatGPT elevated the sentence with, “To illustrate, when students are hungry, it's hard for them to listen because their minds may be preoccupied with thoughts of food.” Whereas the latter sentence is a well-crafted model for the student, this revision arguably loses the student’s voice and tone.

No Instructional Suggestions

No explicit instructional suggestions were included in the responses for Specific Analytic Rubric GPT-3.5, No Info GPT-3.5, and Specific Analytic Rubric GPT-4 sets. The reader was only reminded to provide feedback in a constructive and supportive manner and encourage students to ask questions and seek clarifications on any offered suggestions. While this is logical for both Specific Analytic rubric sets (not asking for instructional suggestions in the prompt), it is surprising for the No Info GPT-3.5 set (which asked for feedback and instructional suggestions).

RQ3: Comparisons Between Teachers and ChatGPT

In response to RQ3, we compared a real teachers’ data-based decision-making (DDDM), including the score and the instructional decision, to the scores generated in the Specific Analytic Rubric GPT-3.5 and Specific Analytic Rubric GPT-4 sets for students’ essays ( N  = 34). The first rubric category scored with a 2 or below was considered the area of need for writing instruction.

GPT-3.5 matched the teacher’s recommendation for the area of writing need 17.6% of the time. For example, the teacher identified Word Selection as the area of need (e.g., high use of repeated words and lacking sensory words) and GPT-3.5 noted the same area of need (e.g., there is some repetition and awkward phrasing). When comparing teacher versus ChatGPTs instructional decisions, there was no perfect match; however, 26.5% were coded as a partial match. For example, both the teacher and GPT-3.5 suggested an instructional activity of modeling how to write a summary sentence.

GPT-4 matched the teacher’s recommendation for the area of writing need 23.5% of the time. Similarly, when comparing the teacher versus ChatGPT’s instructional decisions, 47.1% were coded as a partial match for instruction.

Discussion and Practical Implications

Since the end of 2022 when it debuted, school leaders and teachers of writing have been grappling with what ChatGPT means for writing instruction. Its ability to generate essays from a simple request or to correct writing samples is making an impact on the classroom experience for students with and without disabilities and it is reshaping how teachers assess student writing (Marino et al., 2023 ; Trust et al., 2023 ; Wilson, 2023 ). However, teachers may have limited knowledge of how AI works and poor self-efficacy for using AI in the classroom to support their pedagogical decision making (Chiu et al., 2023 ). It is imperative to ensure that teachers receive professional development to facilitate the effective and efficient use of AI. There are more questions than answers currently, especially for its application by students struggling with academics.

The purpose of this investigation was to explore the application of ChatGPT chatbot for teachers of writing. Specifically, we used different versions of ChatGPT (GPT-3.5 – free and GPT-4 – subscription) and purposefully different types of prompts, providing limited or more information about the student characteristics and the topic of their writing. Essentially, we asked ChatGPT to evaluate an authentic student’s writing, identify the area(s) of need, and provide instructional suggestion(s) for addressing the problematic area(s) in that individual writing sample. We then compared AI-generated feedback to that completed by humans.

The findings indicate the possibilities and limitations of ChatGPT for evaluating student writing, interpreting a teacher-developed rubric, and providing instructional strategies.

Our finding is that, generally, ChatGPT can follow purposeful prompts, interpret and score using a criterion-based rubric when provided, create its own criteria for evaluating student writing, effectively revise student essay writing, celebrate what students do well in their writing, paraphrase student essay ideas, draft outlines of a student’s completed essay, and provide formative feedback in broad and specific areas along different stages of the writing process. Moreover, the response is immediate. These findings are consistent with previous investigations of ChatGPT and the assessment of student writing (Steiss et al., 2023 ). However, teachers need to consider the following points before relying on ChatGPT to provide feedback to their struggling writers.

In the ChatGPT sets which included no contextual information, the responses included more feedback.

All sets generated excessive amounts of feedback about student writing with no delineation of the next clear instructional move a teacher should attend to. So, ChatGPT may work as a great starting point, but teachers will need to go through the response to prioritize and design their instruction. Sifting through information for relevance can be time consuming and may even warrant a teacher verifying the content further.

Additionally, if students relied directly on ChatGPT, without any vetting from a teacher about the content, they too may be overwhelmed by the amount of feedback given to modify their writing or they may even be provided with erroneous feedback.

All GPT-3.5 sets identified broad areas of writing that needed improvement and frequently used common phrases such as grammar, organization/development of ideas, and attention to detail. In addition, this feedback was more often misaligned with students’ writing. This observation is worrisome since GPT-3.5 version of ChatGPT is free and highly accessible, making it likely the preferred AI tool for classroom educators.

Most GPT-4 sets (except one) generated more specific and individualized feedback about student writing. The specific feedback included in the generated outputs were much lengthier and would take much more time for a teacher to review than GPT-3.5 responses.

All sets identified multiple areas of need and when included in the responses, there were multiple instructional suggestions. Even the No Rubric sets, which explicitly prompted ChatGPT to focus on just one area of instructional need and one suggestion, included much more in the responses. This finding reiterates that we are still learning about AI literacy and the language we need to use to communicate effectively.

Both GPT-3.5 and GPT-4 allowed the upload of a researcher-developed analytic rubric and moreover, interpreted the performance criteria, rating scale, and indicators. ChatGPT also used the rubric’s specific language when providing its evaluation of the student writing.

No tailored feedback or specific suggestions were contextualized when prompts included varying ages, grade levels, or various student abilities and needs. Further research is needed to determine the types of AI literacy prompts or the contextual information that ChatGPT needs to address the particular needs of an individual child. Specially designed instruction, the heart of special education, should be tailored to a particular student (Sayeski et al., 2023 ).

Low agreement reported between the rubric scores and instructional suggestions made by teachers and those generated by ChatGPT does not necessarily mean that ChatGPT’s feedback is incorrect. One explanation for the difference may be that teachers provide targeted and individualized instruction using multiple forms of data and critical information to make instructional decisions. This includes their own professional judgement and knowledge about how each students’ backgrounds, culture, and language may influence student performance (McLeskey et al., 2017 ).

Limitations

This study is an initial exploration. There are several limitations that need to be taken into consideration. First and foremost, the four prompts were designed to present the chatbots with varying levels of details and student information to consider when providing feedback about a student’s writing sample. For example, Specific Analytic Rubric prompt asked the chatbot to assess students’ writing using an uploaded rubric, while No Rubric prompt asked to identify one area of need for the student’s writing and offer one instructional suggestion to address it. In addition to providing the chatbots with varying information, we also used varying language throughout the prompts when seeking feedback and suggestions (e.g., “Identify areas of need for this student’s writing”; “Identify one area of need … and offer one instructional suggestion”; “what feedback and instructional suggestions…”). Chatbots are clearly sensitive to the word choices made; thus, a consistency of the language in prompts should be considered for any future investigations that aim at prompt comparison. The purpose of this work was not to compare the four prompts in effort to find the best possible one. We also were not looking specifically for the feedback that could be shared with students as is (even though some versions generated such feedback without additional prompting). Instead, we were trying to explore how the output might differ depending on the prompts with differing level of detail. So, some of the reported difference are logical. We also did not prompt the ChatGPT any further, which would most likely result in refined feedback and/or suggestions. There is an infinite number of prompts that we could have used in this analysis. In fact, a new field of prompt engineering is emerging right in front of our eyes as we learn to design inputs for generative AI tools that would produce optimal outputs. Further investigations of various prompts to feed ChatGPT are needed. Our hope is that this paper will inspire teacher to spend some time exploring different tools and prompts in effort to find the most appropriate output depending on their context and their students’ needs.

Also, there was a limited numbers of essays from each specific group of learners (e.g., certain age/grade, specific disability categories and other characteristics). While we reported meaningful findings for this initial exploratory analysis, future research should include writing products from more homogeneous groups. Finally, teachers’ DDDM was accomplished by evaluating a completed graphic organizer, while ChatGPT feedback was provided based on the final student essay copied and pasted from the TBGO. Future research should consider new features of generative AI tools (e.g., Chat GPT’s new image analysis feature) where an image of a completed graphic organizer can be uploaded and analyzed.

This study offers examples for how to potentially incorporate AI effectively and efficiently into writing instruction. High quality special education teachers are reflective about their practice, use a variety of assistive and instructional technologies to promote student learning, and regularly monitor student progress with individualized assessment strategies. It seems very likely that teachers will adopt the capabilities of generative AI tools. With ongoing development and enhancements, AI technology is certain to become an integral component of classroom instruction. However, given the limitations of ChatGPT identified in this study, teacher-led instruction and decision making is still needed to personalize and individualize specialized instruction. Engaging with the technology more and building familiarity of what it can do to improve student learning and teacher practice is warranted.

Archer, A. L., & Hughes, C. A. (2011). Explicit instruction: Effective and efficient teaching . Guilford press.

Google Scholar  

Barbetta, P. M. (2023). Remedial and compensatory writing technologies for middle school students with learning disabilities and their classmates in inclusive classrooms. Preventing School Failure: Alternative Education for Children and Youth . https://doi.org/10.1080/1045988X.2023.2259837

Article   Google Scholar  

Boykin, A., Evmenova, A. S., Regan, K., & Mastropieri, M. (2019). The impact of a computer-based graphic organizer with embedded self-regulated learning strategies on the argumentative writing of students in inclusive cross-curricula settings. Computers & Education, 137 , 78–90. https://doi.org/10.1016/j.compedu.2019.03.008

Billingsley, B., & Bettini, E. (2017). Improving special education teacher quality and effectiveness. In J. M. Kauffman, D. P. Hallahan, & P. C. Pullen (Eds.), Handbook of special education (2nd ed., pp. 501-520). Boston: Taylor & Francis.

Brantlinger, E., Jimenez, R., Klinger, J., Pugach, M., & Richardson, V. (2005). Qualitative studies in special education. Exceptional Children, 71 (2), 195–207. https://doi.org/10.1177/001440290507100205

Garg, S., & Sharma, S. (2020). Impact of artificial intelligence in special need education to promote inclusive pedagogy. International Journal of Information and Education Technology, 10 (7), 523–527. https://doi.org/10.18178/ijiet.2020.10.7.1418

Chiu, T. K. F., Xia, Q., Zhou, X., Chai, C. S., & Cheng, M. (2023). Systematic literature review on opportunities, challenges, and future research recommendations. Computers and Education: Artificial Intelligence, 4 , 1–15. https://doi.org/10.1010/j.caeai.2022.100118

Chung, P. J., Patel, D. R., & Nizami, I. (2020). Disorder of written expression and dysgraphia: Definition, diagnosis, and management. Translational Pediatrics, 9 (1), 46–54. https://doi.org/10.21037/tp.2019.11.01

Cohen, Z. (2023). Moving beyond Google: Why ChatGPT is the search engine of the future [Blog Post]. Retrieved from https://thecorecollaborative.com/moving-beyond-google-why-chatgpt-is-the-search-engine-of-the-future/ . Accessed 1 Nov 2023

Day, J., Regan, K., Evmenova, A. S., Verbiest, C., Hutchison, A., & Gafurov, B. (2023). The resilience of students and teachers using a virtual writing intervention during COVID-19. Reading & Writing Quarterly, 39 (5), 390–412. https://doi.org/10.1080/10573569.2022.2124562

Edyburn, D. (2021). Universal usability and Universal Design for Learning. Intervention in School and Clinic, 56 (5), 310–315. https://doi.org/10.1177/1053451220963082

Evmenova, A. S., & Regan, K. (2019). Supporting the writing process with technology for students with disabilities. Intervention in School and Clinic, 55 (2), 78–87. https://doi.org/10.1177/1053451219837636

Evmenova, A. S., Regan, K., Boykin, A., Good, K., Hughes, M. D., MacVittie, N. P., Sacco, D., Ahn, S. Y., & Chirinos, D. S. (2016). Emphasizing planning for essay writing with a computer-based graphic organizer. Exceptional Children, 82 (2), 170–191. https://doi.org/10.1177/0014402915591697

Evmenova, A. S., Regan, K., & Hutchison, A. (2018-2023). W EGO RIITE: Writing ef iciently with graphic organizers – responsive instruction while implementing technology ef ectively (Project No. H327S180004) [Grant]. Technology and media services for individuals with disabilities: Stepping-up technology implementation grant, office of special education.

Evmenova, A. S., Regan, K., & Hutchison, A. (2020a). AT for writing: Technology-based graphic organizers with embedded supports.  TEACHING Exceptional Children , 52 (4), 266–269. https://doi.org/10.1177/0040059920907571

Evmenova, A. S., Regan, K., Ahn, S. Y., & Good, K. (2020b). Teacher implementation of a technology-based intervention for writing. Learning Disabilities: A Contemporary Journal, 18 (1), 27–47. https://www.ldw-ldcj.org/

Golinkoff, R. M., & Wilson, J. (2023). ChatGPT is a wake-up call to revamp how we teach writing. [Opinion]. Retrieved from https://www.inquirer.com/opinion/commentary/chatgpt-ban-ai-education-writing-critical-thinking-20230202.html . Accessed 1 Nov 2023

Graham, S. (2019). Changing how writing is taught. Review of Research in Education, 43 (1), 277–303. https://doi.org/10.3102/0091732X18821125

Graham, S., Capizzi, A., Harris, K. R., Hebert, M., & Morphy, P. (2014). Teaching writing to middle school students: A national survey. Reading and Writing, 27 , 1015–1042. https://doi.org/10.1007/s11145-013-9495-7

Graham, S., Collins, A. A., & Rigby-Wills, H. (2017). Writing characteristics of students with learning disabilities and typically achieving peers: A meta-analysis. Exceptional Children, 83 (2), 199–218. https://doi.org/10.1177/001440291666407

Guest, G., MacQueen, K. M., & Namey, E. E. (2011). Applied thematic analysis . SAGE Publications.

Hughes C. A., Riccomini P. J., & Morris J. R. (2018). Use explicit instruction. In  High leverage practices for inclusive classrooms  (pp. 215–236). Routledge. https://doi.org/10.4324/9781315176093 .

Hutchison, A., Evmenova, A. S., Regan, K., & Gafurov, B. (2024). Click, see, do: Using digital scaffolding to support persuasive writing instruction for emerging bilingual learners. Reading Teacher . https://doi.org/10.1002/trtr.2310

Johnson S. M., Kraft M. A., & Papay J. P. (2012). How context matters in high-need schools: The effects of teachers’ working conditions on their professional satisfaction and their students’ achievement. Teachers College Record , 114 , 1–39.

Kiuhara, S. A., Graham, S., & Hawken, L. S. (2009). Teaching writing to highschool students. Journal of Educational Psychology, 101 (1), 136–160. https://doi.org/10.1037/a0013097

Li, J., Link, S., & Hegelheimer, V. (2015). Rethinking the role of automated writing evaluation (AWE) feedback in ESL writing instruction. Journal of Second Language Writing, 27 , 1–18. https://doi.org/10.1016/j.jslw.2014.10.004

Marino, M. T., Vasquez, E., Dieker, L., Basham, J., & Blackorby, J. (2023). The future of artificial intelligence in special education technology. Journal of Special Education Technology, 38 (3), 404–416. https://doi.org/10.1177/01626434231165977

McLeskey, J., Barringer, M.-D., Billingsley, B., Brownell, M., Jackson, D., Kennedy, M., Lewis, T., Maheady, L., Rodriguez, J., Scheeler, M. C., Winn, J., & Ziegler, D. (2017). High-leverage practices in special education . Council for Exceptional Children & CEEDAR Center.

Office of Educational Technology (2023). Artificial intelligence and the future of teaching and learning: Insights and recommendations. Retrieved from https://tech.ed.gov/files/2023/05/ai-future-of-teaching-and-learning-report.pdf . Accessed 1 Nov 2023

Park, V., & Datnow, A. (2017). Ability grouping and differentiated instruction in an era of data-driven decision making. American Journal of Education, 123 (2), 281–306.

Rana, S. (2018). The impact of a computer-based graphic organizer with embedded technology features on the personal narrative writing of upper elementary students with high-incidence disabilities (Publication No. 13420322) [Doctoral dissertation, George Mason University]. ProQuest Dissertation Publishing.

Reeves, T. D., & Chiang, J.-L. (2018). Online interventions to promote teacher data-driven decision making: Optimizing design to maximize impact. Studies in Educational Evaluation, 59 , 256–269. https://doi.org/10.1016/j.stueduc.2018.09.006

Regan, K., Evmenova, A. S., Good, K., Leggit, A, Ahn, S., Gafurov, G., & Mastropieri, M. (2018). Persuasive writing with mobile-based graphic organizers in inclusive classrooms across the curriculum. Journal of Special Education Technology, 33 (1), 3–14. https://doi.org/10.1177/0162643417727292

Regan, K., Evmenova, A. S., Hutchison, A., Day, J., Stephens, M., Verbiest, C., & Gufarov, B. (2021). Steps for success: Making instructional decisions for students’ essay writing. TEACHING Exceptional Children , 54 (3), 202–212. https://doi.org/10.1177/00400599211001085

Regan, K., Evmenova, A. S., & Hutchison, A. (2023a). Specially designed assessment of writing to individualize instruction for students. In K. L. Write, & T. S. Hodges (Eds.), Assessing disciplinary writing in both research and practice (pp. 29–56). IGI Global. https://doi.org/10.4018/978-1-6684-8262-9

Regan, K., Evmenova, A. S., Mergen, R., Verbiest, C., Hutchison, A., Murnan, R., Field, S., & Gafurov, B. (2023b). Exploring the feasibility of virtual professional development to support teachers in making data-based decisions for improving student writing. Learning Disabilities Research & Practice , 38 (1), 40–56. https://doi.org/10.1111/ldrp.12301

Sayeski, K. L., Reno, E. A., & Thoele, J. M. (2023). Specially designed instruction: Operationalizing the delivery of special education services. Exceptionality, 31 (3), 198–210. https://doi.org/10.1080/09362835.2022.2158087

Shermis, M. D. (2014). State-of-the-art automated essay scoring: Competition, results, and future directions from a United States demonstration. Assessing Writing, 20 , 53–76. https://doi.org/10.1016/j.asw.2013.04.001

Steiss, J., Tate, T., Graham, S., Cruz, J., Hevert, M., Wang, J., Moon, Y., Tseng, W., & Warschauer, M. (2023). Comparing the quality of human and ChatGPT feedback on students’ writing. Retrieved from https://osf.io/preprints/edarxiv/ty3em/ . Accessed 1 Nov 2023

Trust, T., Whalen, J., & Mouza, C. (2023). Editorial: ChatGPT: Challenges, opportunities, and implications for teacher education. Contemporary Issues in Technology and Teacher Education, 23 (1), 1–23.

U.S. Department of Education. (2023). Office of Educational Technology, Artificial Intelligence and Future of Teaching and Learning: Insights and Recommendations , Washington, DC.

Wilson, J., & Andrada, G. N. (2016). Using automated feedback to improve writing quality: Opportunities and challenges. In Y. Rosen, S. Ferrara, & M. Mosharraf (Eds.), Handbook of research on technology tools for real-world skill development (pp. 678–703). IGI Global.

Wilson, J., Myers, M. C., & Potter, A. (2022). Investigating the promise of automated writing evaluation for supporting formative writing assessment at scale. Assessment in Education: Principles, Policy, & Practice, 29 (1), 1–17. https://doi.org/10.1080/0969594X.2022.2025762

Wilson, J. (2023). Writing without thinking? There’s a place for ChatGPT – if used properly [Guest Commentary]. Retrieved from https://www.baltimoresun.com/opinion/op-ed/bs-ed-op-0206-chatgpt-tool-20230203-mydxfitujjegndnjwwen4s4x7m-story.html . Accessed 1 Nov 2023

Zdravkova, K. (2022). The potential of artificial intelligence for assistive technology in education. In M. Ivanović, A. Klašnja-Milićević, L. C. Jain (Eds) Handbook on intelligent techniques in the educational process. Learning and analytics in intelligent systems (vol 29). Springer. https://doi.org/10.1007/978-3-031-04662-9_4 .

Download references

This research was supported by the U.S. Department of Education, Office of Special Education Programs [award number: H327S180004]. The views expressed herein do not necessarily represent the positions or policies of the Department of Education. No official endorsement by the U.S. Department of Education of any product, commodity, service, or enterprise mentioned in this publication is intended or should be inferred.

Author information

Authors and affiliations.

George Mason University, 4400 University Dr. MS 1F2, Fairfax, VA, 22030, USA

Anya S. Evmenova, Kelley Regan, Reagan Mergen & Roba Hrisseh

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Anya S. Evmenova .

Ethics declarations

Research involving human participants.

All the procedures in this study were evaluated and approved by the Institutional Research Board. All authors have complied with the ethical standards in the treatment of our participants.

Informed Consent

Informed parental consent and student assent were obtained for all individual participants in the study.

Conflict of Interest

There is no known conflict of interest to declare.

Additional information

Publisher's note.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .

Reprints and permissions

About this article

Evmenova, A.S., Regan, K., Mergen, R. et al. Improving Writing Feedback for Struggling Writers: Generative AI to the Rescue?. TechTrends (2024). https://doi.org/10.1007/s11528-024-00965-y

Download citation

Accepted : 19 April 2024

Published : 14 May 2024

DOI : https://doi.org/10.1007/s11528-024-00965-y

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Data-based decision-making
  • Generative AI
  • Essay feedback
  • Struggling writers
  • Students with and without high-incidence disabilities
  • Writing instruction
  • Find a journal
  • Publish with us
  • Track your research

IMAGES

  1. Compare and Contrast Essay Rubric

    rubric for a comparison essay

  2. Compare and Contrast Essay Rubric

    rubric for a comparison essay

  3. Wondrous Compare And Contrast Essay Rubric ~ Thatsnotus

    rubric for a comparison essay

  4. Comparative Essay Rubric

    rubric for a comparison essay

  5. Comparative Essay Rubric

    rubric for a comparison essay

  6. How to write a comparison essay with text

    rubric for a comparison essay

VIDEO

  1. Reviewing Writing Essay Rubric Up Dated Sp 2024

  2. Multi-paragraph essay rubric

  3. 17 How to Assign a Rubric to an Essay item within TestWriter

  4. Essay Rubric and CUSS

  5. Great Writing 4

  6. Essay On IPL (Indian Premier League) With Easy Language In English

COMMENTS

  1. PDF Comparison and Contrast Rubric

    The paper points to specific examples to illustrate the comparison. The paper includes only the information relevant to the comparison. The paper compares and contrasts items clearly, but the supporting information is general. The paper includes only the information relevant to the comparison. The paper compares and contrasts items clearly, but ...

  2. PDF Scoring Rubric: Comparison/Contrast

    The introduction lacks a topic sentence (in a paragraph) or thesis sentence (in an essay); the body does not de-velop comparisons or contrasts; the conclu-sion is missing, contra-dictory, or repetitive; the writing lacks transitions. Elements of Comparison/ Contrast Writing. The overall purpose of informing, persuading, evaluating, or enter ...

  3. PDF Name/Period: COMPARISON-CONTRAST ESSAY RUBRIC

    COMPARISON-CONTRAST ESSAY RUBRIC. The paper clearly compares and contrasts points that are sophisticated, offers specific examples to illustrate the comparison, and includes only the information relevant to the comparison. The paper compares and contrasts points clearly, but the supporting information is general, and/or the points are basic.

  4. Comparing and Contrasting

    This handout will help you first to determine whether a particular assignment is asking for comparison/contrast and then to generate a list of similarities and differences, decide which similarities and differences to focus on, and organize your paper so that it will be clear and effective. It will also explain how you can (and why you should ...

  5. Comparing and Contrasting in an Essay

    Making effective comparisons. As the name suggests, comparing and contrasting is about identifying both similarities and differences. You might focus on contrasting quite different subjects or comparing subjects with a lot in common—but there must be some grounds for comparison in the first place. For example, you might contrast French ...

  6. PDF Compare and Contrast Essay Rubric

    Transition sentences are smooth. Essay includes introduction and conclusion. Writing is logically and clearly organized. Three similarities and differences are readily identifiable. Writing is somewhat organized. Points are identifiable with minimal reader confusion. Writing is disorganized and hard to follow. OR.

  7. PDF Compare & Contrast Essay

    Step 1 - Students will encounter different kinds of writing assignments; one of the most common is the comparison/contrast essay where the focus is on the ways in which certain things or ideas - usually two of them - are similar to (comparison) and/or different from (contrast) one another. When writing such essays students make ...

  8. Teaching the Compare and Contrast Essay

    Step 1. Discuss practical reasons for comparing and contrasting. Discuss reasons for learning to write about similarities and differences. Selecting subjects that matter to students is critical for this step. For example, one might be to compare two models of cars and then write a letter to a benefactor who might buy them one.

  9. PDF ESSAY 2 RUBRIC

    Grading Rubric for Essay 2: Comparative Analysis. Criteria. Exemplary. Good. Fair. Poor. Thesis. (30 possible points) Thesis states in clear and direct language: the two texts to be discussed; 2-3 grounds for comparison; and either an evaluative claim that determines one to be better than the other, or an explanatory claim that interprets the ...

  10. PDF Compare and Contrast Rubric

    The paper moves from one idea to the next, but there is little variety. The paper uses comparison and contrast transition words to show relationships between ideas. Some transitions work well; but connections between other ideas are fuzzy. The transitions between ideas are unclear or nonexistent. Grammar &.

  11. iRubric: Compare/Contrast Essay rubric

    20pts. Excellent. Each paragraph contains a clearly focused topic sentence that relates to the thesis statement. Details in the paragraphs are clear and specific, and there are enough details to create vivid images for a thorough comparison or contrast. The details clearly support the author's point for the comparison/contrast.

  12. PDF Compare and Contrast Rubric Category 4 3 2 1

    The paper uses comparison and contrast transition words to show relationships between ideas. The paper uses a variety of sentence structures and transitions. Writer makes no errors in grammar or spelling that distract the reader from the content. 3. The paper compares and contrasts items clearly, but the supporting information is general.

  13. Compare and Contrast Rubric

    Teaching with this printout. When assigning a compare and contrast writing assignment, students need to be aware of what makes an outstanding written work. This rubric is a great tool to show students what is expected of them in a concrete way. Additionally, this rubric will help teachers assess this student writing and inform further instruction.

  14. Compare & Contrast Essay Rubric for Middle School

    The core of the compare/contrast essay for students is, of course, comparing and contrasting. So when setting up the rubric for your students, you need to start with these skills. Whether students ...

  15. PDF Comparative Essay Rubric

    Essay is between 4 and 5 paragraphs in length. Essay is typed, double spaced, has 1-inch margins, and is in Times New Roman (12 pt.) font. Includes the titles and authors of both short story selections. Focus is on one literary element and two short stories, chosen from the lists provided. Focus of the writing is a comparison and contrast of ...

  16. Compare & Contrast Essay Rubric for High School

    Compare & Contrast Essay Rubric for High School. Nora has a Master's degree in teaching, and has taught a variety of elementary grades. Asking your students to write an essay that compares and ...

  17. PDF Writing Assessment and Evaluation Rubrics

    Holistic scoring is a quick method of evaluating a composition based on the reader's general impression of the overall quality of the writing—you can generally read a student's composition and assign a score to it in two or three minutes. Holistic scoring is usually based on a scale of 0-4, 0-5, or 0-6.

  18. DOC Central Bucks School District / Homepage

    ÐÏ à¡± á> þÿ a c þÿÿÿ ...

  19. PDF Five-Paragraph Essay Writing Rubric

    Five-Paragraph Essay Writing Rubric. Thesis statement/topic idea sentence is clear, correctly placed, and restated in the closing sentence. Your three supporting ideas are briefly mentioned. Thesis statement/topic idea sentence is either unclear or incorrectly placed, and it's restated in the closing sentence.

  20. Rubric Best Practices, Examples, and Templates

    Rubric Best Practices, Examples, and Templates. A rubric is a scoring tool that identifies the different criteria relevant to an assignment, assessment, or learning outcome and states the possible levels of achievement in a specific, clear, and objective way. Use rubrics to assess project-based student work including essays, group projects ...

  21. Improving Writing Feedback for Struggling Writers: Generative AI to the

    In response to RQ3, we compared a real teachers' data-based decision-making (DDDM), including the score and the instructional decision, to the scores generated in the Specific Analytic Rubric GPT-3.5 and Specific Analytic Rubric GPT-4 sets for students' essays (N = 34). The first rubric category scored with a 2 or below was considered the ...