• Resources ›
  • For Educators ›
  • Tips & Strategies ›

Creating and Scoring Essay Tests

FatCamera / Getty Images

  • Tips & Strategies
  • An Introduction to Teaching
  • Policies & Discipline
  • Community Involvement
  • School Administration
  • Technology in the Classroom
  • Teaching Adult Learners
  • Issues In Education
  • Teaching Resources
  • Becoming A Teacher
  • Assessments & Tests
  • Elementary Education
  • Secondary Education
  • Special Education
  • Homeschooling
  • M.Ed., Curriculum and Instruction, University of Florida
  • B.A., History, University of Florida

Essay tests are useful for teachers when they want students to select, organize, analyze, synthesize, and/or evaluate information. In other words, they rely on the upper levels of Bloom's Taxonomy . There are two types of essay questions: restricted and extended response.

  • Restricted Response - These essay questions limit what the student will discuss in the essay based on the wording of the question. For example, "State the main differences between John Adams' and Thomas Jefferson's beliefs about federalism," is a restricted response. What the student is to write about has been expressed to them within the question.
  • Extended Response - These allow students to select what they wish to include in order to answer the question. For example, "In Of Mice and Men , was George's killing of Lennie justified? Explain your answer." The student is given the overall topic, but they are free to use their own judgment and integrate outside information to help support their opinion.

Student Skills Required for Essay Tests

Before expecting students to perform well on either type of essay question, we must make sure that they have the required skills to excel. Following are four skills that students should have learned and practiced before taking essay exams:

  • The ability to select appropriate material from the information learned in order to best answer the question.
  • The ability to organize that material in an effective manner.
  • The ability to show how ideas relate and interact in a specific context.
  • The ability to write effectively in both sentences and paragraphs.

Constructing an Effective Essay Question

Following are a few tips to help in the construction of effective essay questions:

  • Begin with the lesson objectives in mind. Make sure to know what you wish the student to show by answering the essay question.
  • Decide if your goal requires a restricted or extended response. In general, if you wish to see if the student can synthesize and organize the information that they learned, then restricted response is the way to go. However, if you wish them to judge or evaluate something using the information taught during class, then you will want to use the extended response.
  • If you are including more than one essay, be cognizant of time constraints. You do not want to punish students because they ran out of time on the test.
  • Write the question in a novel or interesting manner to help motivate the student.
  • State the number of points that the essay is worth. You can also provide them with a time guideline to help them as they work through the exam.
  • If your essay item is part of a larger objective test, make sure that it is the last item on the exam.

Scoring the Essay Item

One of the downfalls of essay tests is that they lack in reliability. Even when teachers grade essays with a well-constructed rubric, subjective decisions are made. Therefore, it is important to try and be as reliable as possible when scoring your essay items. Here are a few tips to help improve reliability in grading:

  • Determine whether you will use a holistic or analytic scoring system before you write your rubric . With the holistic grading system, you evaluate the answer as a whole, rating papers against each other. With the analytic system, you list specific pieces of information and award points for their inclusion.
  • Prepare the essay rubric in advance. Determine what you are looking for and how many points you will be assigning for each aspect of the question.
  • Avoid looking at names. Some teachers have students put numbers on their essays to try and help with this.
  • Score one item at a time. This helps ensure that you use the same thinking and standards for all students.
  • Avoid interruptions when scoring a specific question. Again, consistency will be increased if you grade the same item on all the papers in one sitting.
  • If an important decision like an award or scholarship is based on the score for the essay, obtain two or more independent readers.
  • Beware of negative influences that can affect essay scoring. These include handwriting and writing style bias, the length of the response, and the inclusion of irrelevant material.
  • Review papers that are on the borderline a second time before assigning a final grade.
  • How to Study Using the Basketball Review Game
  • Creating Effective Fill-in-the-Blank Questions
  • Teacher Housekeeping Tasks
  • 4 Tips for Effective Classroom Management
  • Dealing With Trips to the Bathroom During Class
  • Tips to Cut Writing Assignment Grading Time
  • Collecting Homework in the Classroom
  • Why Students Cheat and How to Stop Them
  • Planning Classroom Instruction
  • Using Cloze Tests to Determine Reading Comprehension
  • 10 Strategies to Boost Reading Comprehension
  • Taking Daily Attendance
  • How Scaffolding Instruction Can Improve Comprehension
  • Field Trips: Pros and Cons
  • Assignment Biography: Student Criteria and Rubric for Writing
  • 3 Poetry Activities for Middle School Students
 
 
TOPICS A. Fill-in-the-Blank Items B. Essay Questions C. Scoring Options

Assignments


Essay questions are a more complex version of constructed response assessments. With essay questions, there is one general question or proposition, and the student is asked to respond in writing. This type of assessment is very powerful -- it allows the students to express themselves and demonstrate their reasoning related to a topic. Essay questions often demand the use of higher level thinking skills, such as analysis, synthesis, and evaluation.

Essay questions may appear to be easier to write than multiple choice and other question types, but writing effective essay questions requires a great deal of thought and planning. If an essay question is vague, it will be much more difficult for the students to answer and much more difficult for the instructor to score. Well-written essay questions have the following features:

Essay questions are used both as formative assessments (in classrooms) and summative assessments (on standardized tests). There are 2 major categories of essay questions -- (also referred to as or ) and .

Short response questions are more focused and constrained than extended response questions. For example, a short response might ask a student to "write an example," "list three reasons," or "compare and contrast two techniques." The short response items on the Florida assessment (FCAT) are designed to take about 5 minutes to complete and the student is allowed up to 8 lines for each answer. The short responses are scored using a 2-point scoring rubric. A complete and correct answer is worth 2 points. A partial answer is worth 1 point.


How are the scrub jay and the mockingbird different? Support your answer with details and information from the article.

Extended Response

Extended responses can be much longer and complex then short responses, but students should be encouraged to remain focused and organized. On the FCAT, students have 14 lines for each answer to an extended response item, and they are advised to allow approximately 10-15 minutes to complete each item. The FCAT extended responses are scored using a 4-point scoring rubric. A complete and correct answer is worth 4 points. A partial answer is worth 1, 2, or 3 points.

Robert is designing a demonstration to display at his school’s science fair. He will show how changing the position of a fulcrum on a lever changes the amount of force needed to lift an object. To do this, Robert will use a piece of wood for a lever and a block of wood to act as a fulcrum. He plans to move the fulcrum to different places on the lever to see how its placement affects the force needed to lift an object.

  Identify at least two other actions that would make Robert’s demonstration better.

  Explain why each action would improve the demonstration.

| | | Constructed Response


and the at .

 

AP PGECET

Restricted Response Questions: A Closer Look

  • by Sandra Vargas
  • October 28, 2023

Welcome to the world of academic assessments where questions play a vital role in gauging knowledge and understanding. One such type of question is the “restricted response question .” If you’re a student or an educator looking to brush up on your assessment strategies, or simply a curious mind, you’ve come to the right place.

In this blog post, we’ll explore what exactly a restricted response question is and how it differs from other question types. We’ll also delve into why they are used and offer some tips on how to answer them effectively. So, whether you’re preparing for an exam or just aiming to expand your knowledge, let’s dive in and uncover the essence of restricted response questions together.

What is a restricted response question?

What is a Restricted Response Question

Have you ever taken a test where you felt like you were unraveling the mysteries of the universe, only to be stumped by a question that seemed to come from another dimension? We’ve all been there. One of the most confounding types of questions that can leave us scratching our heads is the restricted response question.

Understanding the Restricted Response Question

So, what exactly is a restricted response question? Well, let me break it down for you in good ol’ plain English. A restricted response question is a type of question that puts some serious boundaries on how you can answer. It’s like being trapped in a cage with only a few options to escape. But fear not, my friend, because there’s a method to this madness.

The Purpose Behind Restricted Response Questions

Restricted response questions have a purpose, and it’s not just to make your brain ache. These questions are designed to measure specific knowledge or skills, kind of like a sniper zeroing in on its target. They focus on assessing your comprehension, analysis, and application of information, rather than just regurgitating facts like a parrot.

The Structure of a Restricted Response Question

Now that you know the why, let’s dive into the how. A restricted response question typically consists of a stem, which sets the stage for what you need to accomplish, and specific guidelines or criteria for your response. It’s like a mini adventure with a clear map and instructions on where to go next. It leaves little room for ambiguity or improvisation, which can be a blessing (or a curse, depending on your perspective).

Examples of Restricted Response Questions

To paint a better picture of what a restricted response question looks like, here are a couple of examples:

  • “Describe three key factors that contribute to climate change and explain their impacts on the environment.”
  • “Create a timeline highlighting the major events leading up to the American Revolutionary War, including dates and brief explanations for each event.”

See how these questions provide clear directives on what you need to do? They don’t leave much room for random musings or philosophical ponderings. It’s all about focusing on the task at hand and demonstrating your knowledge effectively.

The Takeaway

So, the next time you encounter a restricted response question, don’t panic. Remember that it’s designed to assess your critical thinking skills and application of knowledge within a specific framework. Embrace the challenge, follow the guidelines, and show off your expertise like a boss.

Now that we’ve deciphered the code of restricted response questions, let’s move on to our next adventure: exploring the world of extended response questions. Stay tuned, fellow knowledge warriors!

What is a restricted response question?

FAQ: Restricted Response Questions

What is a restricted response question.

A restricted response question is just like that one friend who asks you specific questions, reigning in your response options and not giving you a chance to ramble on about your entire life story. In simpler terms, it is a question that requires a concise and focused answer within a limited range.

How do you agree with a statement in an essay

Ah, the art of agreement! It’s like finding the perfect wingman or wingwoman to back up your every move. In an essay, when you come across a statement you want to agree with, you’ll want to follow a few simple steps:

Start with a clear statement: Clearly state your agreement with the argument or statement that the essay presents. Be firm, but avoid coming across as an overenthusiastic cheerleader.

Provide evidence and examples: Support your agreement with concrete evidence or examples. Show that you’ve done your research and you’re not just nodding along mindlessly. Remember, facts are your friends!

Explain your reasoning: This is where you get to shine! Explain the logic behind your agreement. Share your insights, thoughts, and reasoning in a clear and coherent manner . Let the reader understand why you’re on the same page.

Address potential counterarguments : Ah, the skeptics! Anticipate and address potential counterarguments against your agreement. Show that you’ve thought it through and considered different perspectives. It’s like playing chess and staying two steps ahead of your opponents.

Wrap it up with a bow: Conclude your agreement like a boss. Summarize your main points, leaving no room for doubt or confusion. You’ve successfully convinced the reader that you’re both head-nodding buddies.

Remember, agreeing in an essay is not about mindlessly accepting everything that comes your way. It’s about critically analyzing and providing support for your stance. So, agree with grace, confidence, and a touch of persuasive flair.

And there you have it! A mini FAQ to help you navigate the world of restricted response questions and agreeing like a boss in your essays. Happy writing, my fellow wordsmiths!

  • application
  • little room
  • major events
  • response question
  • specific guidelines

' src=

Sandra Vargas

Is high school a proper noun, what side is your femoral artery on understanding your body's inner mechanics, you may also like, ethical dimensions in information systems: exploring the five moral dimensions.

  • by Lindsey Smith
  • October 10, 2023

Is Ibuprofen an Antihistamine?

  • by Adam Davis
  • October 29, 2023

Is it Good to Listen to Audiobooks While Sleeping?

  • by James Short
  • October 7, 2023

Do NFL Season Ticket Holders Get Super Bowl Tickets?

  • by Matthew Morales

What to Do If You Swallow a Jolly Rancher: A Guide to Navigating Candy Mishaps

  • by Veronica Lopez

TNA Impact Wrestling on DirecTV: What Channel to Tune In for All the Action

  • by Jackie Hobbs
  • October 5, 2023

Teachers Institute

Types of Questions in Teacher Made Achievement Tests: A Comprehensive Guide

by Prince Kumar | Jan 5, 2024

When it comes to assessing students’ learning, teachers often turn to achievement tests they’ve created themselves. These tests are powerful tools that can provide both educators and learners with valuable insights into academic progress and understanding. But what types of questions make up these teacher-made tests? Understanding the various types of test items is crucial for designing assessments that are not only effective but also fair and comprehensive. Let’s dive into the world of objective and essay-type questions to see how they function and how best to construct them.

Table of Contents

Objective Type Test Items

Objective test items are those that require students to select or provide a very short response to a question, with one clear, correct answer. This section will explore the different types of objective test items , their uses, and tips for constructing them.

Supply Type Items

  • Short Answer Questions: These require students to recall and provide brief responses.
  • Fill-in-the-Blank: Here, students must supply a word or phrase to complete a statement.
  • Numerical Problems: Often used in math and science, these items require the calculation and provision of a numerical answer.

When constructing supply type items , clarity is key. Questions should be direct, and the required answer should be unambiguous. Avoid complex phrasing and ensure that the blank space provided is proportional to the expected answer’s length.

Selection Type Items

  • Multiple-Choice Questions (MCQs) : Students choose the correct answer from a list of options.
  • True/False Questions : These require students to determine the veracity of a statement.
  • Matching Items : Students must pair related items from two lists.

For selection type items , it’s important to construct distractors (wrong answers) that are plausible. This prevents guessing and encourages students to truly understand the material. In multiple-choice questions, for example, the incorrect options should be common misconceptions or errors related to the subject matter.

Essay Type Test Items

Essay test items call for longer, more detailed responses from students. These questions evaluate not just recall of information but also critical thinking, organization of thoughts, and the ability to communicate effectively through writing.

Extended Response Essay Questions

  • Exploratory Essays : These require a thorough investigation of a topic, often without a strict length constraint.
  • Argumentative Essays : Students must take a stance on an issue and provide supporting evidence.

In extended response essay questions, students should be given clear guidelines regarding the scope and depth of the response expected. Rubrics can be very helpful in setting these expectations and in guiding both the grading process and the students’ preparation.

Restricted Response Essay Questions

  • Reflective Essays : These typically involve a shorter response, reflecting on a specific question or scenario.
  • Analysis Essays : Students dissect a particular concept or event within a set framework.

Restricted response essay questions are valuable for assessing specific skills or knowledge within a limited domain. When constructing these items, ensure the question is focused and that students are aware of any word or time limits.

Examples and Guidelines for Constructing Effective Test Items

Now that we’ve understood the types of questions, let’s look at some examples and guidelines for creating effective test items.

Objective Item Construction

  • Multiple-Choice Example: “What is the capital of France? A) Madrid B) Paris C) Rome D) Berlin” – Ensure there’s only one correct answer.
  • True/False Example: “The Great Wall of China is visible from space.” – Provide a statement that is not ambiguously phrased.

When constructing objective items, make sure the question is based on important content, not trivial facts. The length of the test should be sufficient to cover the breadth of the material, and the items should vary in difficulty to gauge different levels of student understanding.

Essay Item Construction

  • Extended Response Example: “Discuss the impact of the Industrial Revolution on European society.” – This question allows for a broad exploration of the topic.
  • Restricted Response Example: “Describe two methods of conflict resolution and their effectiveness in workplace settings.” – This question limits the scope to two methods and a specific context.

Essay questions should be open-ended to encourage students to think critically and creatively. However, they should also be specific enough to prevent off-topic responses. Providing a clear rubric can help students understand what is expected in their answers and assist teachers in grading consistently.

Teacher-made achievement tests with a mix of objective and essay type questions can provide a comprehensive assessment of student learning. By understanding the different types of questions and following the guidelines for constructing them, educators can create fair, reliable, and valid assessments. This ensures that the results truly reflect students’ knowledge and skills, allowing for targeted feedback and further instructional planning.

What do you think? How can teachers balance the need for comprehensive assessment with the practical limitations of test administration time? Do you think one type of test item is more effective than the other in measuring student learning?

How useful was this post?

Click on a star to rate it!

Average rating 0 / 5. Vote count: 0

No votes so far! Be the first to rate this post.

We are sorry that this post was not useful for you!

Let us improve this post!

Tell us how we can improve this post?

Submit a Comment Cancel reply

Your email address will not be published. Required fields are marked *

Save my name, email, and website in this browser for the next time I comment.

Submit Comment

Assessment for Learning

1 Concept and Purpose of Evaluation

  • Basic Concepts
  • Relationships among Measurement, Assessment, and Evaluation
  • Teaching-Learning Process and Evaluation
  • Assessment for Enhancing Learning
  • Other Terms Related to Assessment and Evaluation

2 Perspectives of Assessment

  • Behaviourist Perspective of Assessment
  • Cognitive Perspective of Assessment
  • Constructivist Perspective of Assessment
  • Assessment of Learning and Assessment for Learning

3 Approaches to Evaluation

  • Approaches to Evaluation: Placement Formative Diagnostic and Summative
  • Distinction between Formative and Summative Evaluation
  • External and Internal Evaluation
  • Norm-referenced and Criterion-referenced Evaluation
  • Construction of Criterion-referenced Tests

4 Issues, Concerns and Trends in Assessment and Evaluation

  • What is to be Assessed?
  • Criteria to be used to Assess the Process and Product
  • Who will Apply the Assessment Criteria and Determine Marks or Grades?
  • How will the Scores or Grades be Interpreted?
  • Sources of Error in Examination
  • Learner-centered Assessment Strategies
  • Question Banks
  • Semester System
  • Continuous Internal Evaluation
  • Choice-Based Credit System (CBCS)
  • Marking versus Grading System
  • Open Book Examination
  • ICT Supported Assessment and Evaluation

5 Techniques of Assessment and Evaluation

  • Concept Tests
  • Self-report Techniques
  • Assignments
  • Observation Technique
  • Peer Assessment
  • Sociometric Technique
  • Project Work
  • School Club Activities

6 Criteria of a Good Tool

  • Evaluation Tools: Types and Differences
  • Essential Criteria of an Effective Tool of Evaluation
  • Reliability
  • Objectivity

7 Tools for Assessment and Evaluation

  • Paper Pencil Test
  • Aptitude Test
  • Achievement Test
  • Diagnostic–Remedial Test
  • Intelligence Test
  • Rating Scales
  • Questionnaire
  • Inventories
  • Interview Schedule
  • Observation Schedule
  • Anecdotal Records
  • Learners Portfolios and Rubrics

8 ICT Based Assessment and Evaluation

  • Importance of ICT in Assessment and Evaluation
  • Use of ICT in Various Types of Assessment and Evaluation
  • Role of Teacher in Technology Enabled Assessment and Evaluation
  • Online and E-examination
  • Learners’ E-portfolio and E-rubrics
  • Use of ICT Tools for Preparing Tests and Analyzing Results

9 Teacher Made Achievement Tests

  • Understanding Teacher Made Achievement Test (TMAT)
  • Types of Achievement Test Items/Questions
  • Construction of TMAT
  • Administration of TMAT
  • Scoring and Recording of Test Results
  • Reporting and Interpretation of Test Scores

10 Commonly Used Tests in Schools

  • Achievement Test Versus Aptitude Test
  • Performance Based Achievement Test
  • Diagnostic Testing and Remedial Activities
  • Question Bank
  • General Observation Techniques
  • Practical Test

11 Identification of Learning Gaps and Corrective Measures

  • Educational Diagnosis
  • Diagnostic Tests: Characteristics and Functions
  • Diagnostic Evaluation Vs. Formative and Summative Evaluation
  • Diagnostic Testing
  • Achievement Test Vs. Diagnostic Test
  • Diagnosing and Remedying Learning Difficulties: Steps Involved
  • Areas and Content of Diagnostic Testing
  • Remediation

12 Continuous and Comprehensive Evaluation

  • Continuous and Comprehensive Evaluation: Concepts and Functions
  • Forms of CCE
  • Recording and Reporting Students Performance
  • Students Profile
  • Cumulative Records

13 Tabulation and Graphical Representation of Data

  • Use of Educational Statistics in Assessment and Evaluation
  • Meaning and Nature of Data
  • Organization/Grouping of Data: Importance of Data Organization and Frequency Distribution Table
  • Graphical Representation of Data: Types of Graphs and its Use
  • Scales of Measurement

14 Measures of Central Tendency

  • Individual and Group Data
  • Measures of Central Tendency: Scales of Measurement and Measures of Central Tendency
  • The Mean: Use of Mean
  • The Median: Use of Median
  • The Mode: Use of Mode
  • Comparison of Mean, Median, and Mode

15 Measures of Dispersion

  • Measures of Dispersion
  • Standard Deviation

16 Correlation – Importance and Interpretation

  • The Concept of Correlation
  • Types of Correlation
  • Methods of Computing Co-efficient of Correlation (Ungrouped Data)
  • Interpretation of the Co-efficient of Correlation

17 Nature of Distribution and Its Interpretation

  • Normal Distribution/Normal Probability Curve
  • Divergence from Normality

Share on Mastodon

Home

  • CRLT Consultation Services
  • Consultation
  • Midterm Student Feedback
  • Classroom Observation
  • Teaching Philosophy
  • Upcoming Events and Seminars
  • CRLT Calendar
  • Orientations
  • Teaching Academies
  • Provost's Seminars
  • For Faculty
  • For Grad Students & Postdocs
  • For Chairs, Deans & Directors
  • Customized Workshops & Retreats
  • Assessment, Curriculum, & Learning Analytics Services
  • CRLT in Engineering
  • CRLT Players
  • Foundational Course Initiative
  • CRLT Grants
  • Other U-M Grants
  • Provost's Teaching Innovation Prize
  • U-M Teaching Awards
  • Retired Grants
  • Staff Directory
  • Faculty Advisory Board
  • Annual Report
  • Equity-Focused Teaching
  • Preparing to Teach
  • Teaching Strategies
  • Testing and Grading
  • Teaching with Technology
  • Teaching Philosophy & Statements
  • Training GSIs
  • Evaluation of Teaching
  • Occasional Papers

Home

Best Practices for Designing and Grading Exams

Adapted from crlt occasional paper #24: m.e. piontek (2008), center for research on learning and teaching.

The most obvious function of assessment methods (such as exams, quizzes, papers, and presentations) is to enable instructors to make judgments about the quality of student learning (i.e., assign grades). However, the method of assessment also can have a direct impact on the quality of student learning. Students assume that the focus of exams and assignments reflects the educational goals most valued by an instructor, and they direct their learning and studying accordingly  (McKeachie  & Svinicki, 2006). General grading systems can have an impact as well.  For example, a strict bell curve (i.e., norm-reference grading) has the potential to dampen motivation and cooperation in a classroom, while a system that strictly rewards proficiency (i.e., criterion-referenced grading ) could be perceived as contributing to grade inflation. Given the importance of assessment for both faculty and student interactions about learning, how can instructors develop exams that provide useful and relevant data about their students' learning and also direct students to spend their time on the important aspects of a course or course unit? How do grading practices further influence this process?

Guidelines for Designing Valid and Reliable Exams

Ideally, effective exams have four characteristics. They are:

  • Valid, (providing useful information about the concepts they were designed to test),
  • Reliable (allowing consistent measurement and discriminating between different levels of performance),
  • Recognizable   (instruction has prepared students for the assessment), and
  • Realistic (concerning time and effort required to complete the assignment)  (Svinicki, 1999). 

Most importantly, exams and assignments should f ocus on the most important content and behaviors emphasized during the course (or particular section of the course). What are the primary ideas, issues, and skills you hope students learn during a particular course/unit/module? These are the learning outcomes you wish to measure. For example, if your learning outcome involves memorization, then you should assess for memorization or classification; if you hope students will develop problem-solving capacities, your exams should focus on assessing students’ application and analysis skills.  As a general rule, assessments that focus too heavily on details (e.g., isolated facts, figures, etc.) “will probably lead to better student retention of the footnotes at the cost of the main points" (Halpern & Hakel, 2003, p. 40). As noted in Table 1, each type of exam item may be better suited to measuring some learning outcomes than others, and each has its advantages and disadvantages in terms of ease of design, implementation, and scoring.

Table 1: Advantages and Disadvantages of Commonly Used Types of Achievement Test Items

Many items can be administered in a relatively short time. Moderately easy to write; easily scored.

Limited primarily to testing knowledge of information.  Easy to guess correctly on many items, even if material has not been mastered.

Can be used to assess broad range of content in a brief period. Skillfully written items can measure higher order cognitive skills. Can be scored quickly.

Difficult and time consuming to write good items. Possible to assess higher order cognitive skills, but most items assess only knowledge.  Some correct answers can be guesses.

Items can be written quickly. A broad range of content can be assessed. Scoring can be done efficiently.

Higher order cognitive skills are difficult to assess.

Many can be administered in a brief amount of time. Relatively efficient to score. Moderately easy to write.

Difficult to identify defensible criteria for correct answers. Limited to questions that can be answered or completed in very few words.

Can be used to measure higher order cognitive skills. Relatively easy to write questions. Difficult for respondent to get correct answer by guessing.

Time consuming to administer and score. Difficult to identify reliable criteria for scoring. Only a limited range of content can be sampled during any one testing period.

Adapted from Table 10.1 of Worthen, et al., 1993, p. 261.

General Guidelines for Developing Multiple-Choice and Essay Questions

The following sections highlight general guidelines for developing multiple-choice and essay questions, which are often used in college-level assessment because they readily lend themselves to measuring higher order thinking skills  (e.g., application, justification, inference, analysis and evaluation).  Yet instructors often struggle to create, implement, and score these types of questions (McMillan, 2001; Worthen, et al., 1993).

Multiple-choice questions have a number of advantages. First, they can measure various kinds of knowledge, including students' understanding of terminology, facts, principles, methods, and procedures, as well as their ability to apply, interpret, and justify. When carefully designed, multiple-choice items also can assess higher-order thinking skills.

Multiple-choice questions are less ambiguous than short-answer items, thereby providing a more focused assessment of student knowledge. Multiple-choice items are superior to true-false items in several ways: on true-false items, students can receive credit for knowing that a statement is incorrect, without knowing what is correct. Multiple-choice items offer greater reliability than true-false items as the opportunity for guessing is reduced with the larger number of options. Finally, an instructor can diagnose misunderstanding by analyzing the incorrect options chosen by students.

A disadvantage of multiple-choice items is that they require developing incorrect, yet plausible, options that can be difficult to create. In addition, multiple- choice questions do not allow instructors to measure students’ ability to organize and present ideas.  Finally, because it is much easier to create multiple-choice items that test recall and recognition rather than higher order thinking, multiple-choice exams run the risk of not assessing the deep learning that many instructors consider important (Greenland & Linn, 1990; McMillan, 2001).

Guidelines for writing multiple-choice items include advice about stems, correct answers, and distractors (McMillan, 2001, p. 150; Piontek, 2008):

  • S tems pose the problem or question.
  • Is the stem stated as clearly, directly, and simply as possible?
  • Is the problem described fully in the stem?
  • Is the stem stated positively, to avoid the possibility that students will overlook terms like “no,” “not,” or “least”?
  • Does the stem provide only information relevant to the problem?

Possible responses include the correct answer and distractors , or the incorrect choices. Multiple-choice questions usually have at least three distractors.

  • Are the distractors plausible to students who do not know the correct answer?
  • Is there only one correct answer?
  • Are all the possible answers parallel with respect to grammatical structure, length, and complexity?
  • Are the options short?
  • Are complex options avoided? Are options placed in logical order?
  • Are correct answers spread equally among all the choices? (For example, is answer “A” correct about the same number of times as options “B” or “C” or “D”)?

An example of good multiple-choice questions that assess higher-order thinking skills is the following test question from pharmacy (Park, 2008):

Patient WC was admitted for third-degree burns over 75% of his body. The attending physician asks you to start this patient on antibiotic therapy.  Which one of the following is the best reason why WC would need antibiotic prophylaxis? a. His burn injuries have broken down the innate immunity that prevents microbial invasion. b. His injuries have inhibited his cellular immunity. c. His injuries have impaired antibody production. d. His injuries have induced the bone marrow, thus activated immune system

A second question builds on the first by describing the patient’s labs two days later, asking the students to develop an explanation for the subsequent lab results. (See Piontek, 2008 for the full question.)

Essay questions can tap complex thinking by requiring students to organize and integrate information, interpret information, construct arguments, give explanations, evaluate the merit of ideas, and carry out other types of reasoning  (Cashin, 1987; Gronlund & Linn, 1990; McMillan, 2001; Thorndike, 1997; Worthen, et al., 1993). Restricted response essay questions are good for assessing basic knowledge and understanding and generally require a brief written response (e.g., “State two hypotheses about why birds migrate.  Summarize the evidence supporting each hypothesis” [Worthen, et al., 1993, p. 277].) Extended response essay items allow students to construct a variety of strategies, processes, interpretations and explanations for a question, such as the following:

The framers of the Constitution strove to create an effective national government that balanced the tension between majority rule and the rights of minorities. What aspects of American politics favor majority rule? What aspects protect the rights of those not in the majority? Drawing upon material from your readings and the lectures, did the framers successfully balance this tension? Why or why not? (Shipan, 2008).

In addition to measuring complex thinking and reasoning, advantages of essays include the potential for motivating better study habits and providing the students flexibility in their responses.  Instructors can evaluate how well students are able to communicate their reasoning with essay items, and they are usually less time consuming to construct than multiple-choice items that measure reasoning.

The major disadvantages of essays include the amount of time instructors must devote to reading and scoring student responses, and  the importance of developing and using carefully constructed criteria/rubrics to insure reliability of scoring. Essays can assess only a limited amount of content in one testing period/exam due to the length of time required for students to respond to each essay item. As a result, essays do not provide a good sampling of content knowledge across a curriculum (Gronlund & Linn, 1990; McMillan, 2001).

Guidelines for writing essay questions include the following (Gronlund & Linn, 1990; McMillan, 2001; Worthen, et al., 1993):

  • Restrict the use of essay questions to educational outcomes that are difficult to measure using other formats. For example, to test recall knowledge, true-false, fill-in-the-blank, or multiple-choice questions are better measures.
  • Generalizations : State a set of principles that can explain the following events.
  • Synthesis : Write a well-organized report that shows…
  • Evaluation : Describe the strengths and weaknesses of…
  • Write the question clearly so that students do not feel that they are guessing at “what the instructor wants me to do.”
  • Indicate the amount of time and effort students should spend on each essay item.
  • Avoid giving students options for which essay questions they should answer. This choice decreases the validity and reliability of the test because each student is essentially taking a different exam.
  • Consider using several narrowly focused questions (rather than one broad question) that elicit different aspects of students’ skills and knowledge.
  • Make sure there is enough time to answer the questions.

Guidelines for scoring essay questions include the following (Gronlund & Linn, 1990; McMillan, 2001; Wiggins, 1998; Worthen, et al., 1993; Writing and grading essay questions , 1990):

  • Outline what constitutes an expected answer.
  • Select an appropriate scoring method based on the criteria. A rubric is a scoring key that indicates the criteria for scoring and the amount of points to be assigned for each criterion.  A sample rubric for a take-home history exam question might look like the following:

Number of references to class reading sources

0-2 references

3-5 references

6+ references

Historical accuracy

Lots of inaccuracies

Few inaccuracies

No apparent inaccuracies

Historical Argument

No argument made; little evidence for argument

Argument is vague and unevenly supported by evidence

Argument is clear and well-supported by evidence

Proof reading

Many grammar and spelling errors

Few (1-2) grammar or spelling errors

No grammar or spelling errors

For other examples of rubrics, see CRLT Occasional Paper #24  (Piontek, 2008).

  • Clarify the role of writing mechanics and other factors independent of the educational outcomes being measured. For example, how does grammar or use of scientific notation figure into your scoring criteria?
  • Create anonymity for students’ responses while scoring and create a random order in which tests are graded (e.g., shuffle the pile) to increase accuracy of the scoring.
  • Use a systematic process for scoring each essay item.  Assessment guidelines suggest scoring all answers for an individual essay question in one continuous process, rather than scoring all answers to all questions for an individual student. This system makes it easier to remember the criteria for scoring each answer.

You can also use these guidelines for scoring essay items to create grading processes and rubrics for students’ papers, oral presentations, course projects, and websites.  For other grading strategies, see Responding to Student Writing – Principles & Practices and Commenting Effectively on Student Writing .

Cashin, W. E. (1987). Improving essay tests . Idea Paper, No. 17. Manhattan, KS: Center for Faculty Evaluation and Development, Kansas State University.

Gronlund, N. E., & Linn, R. L. (1990). Measurement and evaluation in teaching   (6th  ed.). New  York:  Macmillan Publishing Company.

Halpern, D. H., & Hakel, M. D. (2003). Applying the science of learning to the university and beyond. Change, 35 (4), 37-41.

McKeachie, W. J., & Svinicki, M. D. (2006). Assessing, testing, and evaluating: Grading is not the most important function.   In   McKeachie's   Teaching tips: Strategies, research, and theory for college and university teachers (12th ed., pp. 74-86). Boston: Houghton Mifflin Company.

McMillan, J. H. (2001).  Classroom assessment: Principles and practice for effective instruction.  Boston: Allyn and Bacon.

Park, J. (2008, February 4). Personal communication. University of Michigan College of Pharmacy.

Piontek, M. (2008). Best practices for designing and grading exams. CRLT Occasional Paper No. 24 . Ann Arbor, MI. Center for Research on Learning and Teaching.>

Shipan, C. (2008, February 4). Personal communication. University of Michigan Department of Political Science.

Svinicki, M.   D.   (1999a). Evaluating and grading students.  In Teachers and students: A sourcebook for UT- Austin faculty (pp. 1-14). Austin, TX: Center for Teaching Effectiveness, University of Texas at Austin.

Thorndike, R. M. (1997). Measurement and evaluation in psychology and education.   Upper Saddle River, NJ: Prentice-Hall, Inc.

Wiggins, G. P. (1998). Educative assessment: Designing assessments to inform and improve student performance . San Francisco: Jossey-Bass Publishers.

Worthen, B.  R., Borg, W.  R.,  & White, K.  R.  (1993). Measurement and evaluation in the schools .  New York: Longman.

Writing and grading essay questions. (1990, September). For Your Consideration , No. 7. Chapel Hill, NC: Center for Teaching and Learning, University of North Carolina at Chapel Hill.

back to top

Center for Research on Learning and Teaching logo

Contact CRLT

location_on University of Michigan 1071 Palmer Commons 100 Washtenaw Ave. Ann Arbor, MI 48109-2218

phone Phone: (734) 764-0505

description Fax: (734) 647-3600

email Email: [email protected]

Connect with CRLT

tweets

directions Directions to CRLT

group Staff Directory

markunread_mailbox Subscribe to our Blog

COMMENTS

  1. Overheads for Unit 7--Chapter 10 (Essay Questions)

    Restricted-response essays: (i) require students to supply, not just identify, the answer and (ii) can target specific mental skills; Extended-response essays: emphasize integration and application of high-level skills; Can measure writing skills in addition to (or instead of) knowledge and understanding

  2. Tips for Creating and Scoring Essay Tests - ThoughtCo

    Restricted Response - These essay questions limit what the student will discuss in the essay based on the wording of the question. For example, "State the main differences between John Adams' and Thomas Jefferson's beliefs about federalism," is a restricted response.

  3. Classroom Assessment | Constructed Response

    Essay questions are used both as formative assessments (in classrooms) and summative assessments (on standardized tests). There are 2 major categories of essay questions -- short response (also referred to as restricted or brief ) and extended response.

  4. Restricted Response Questions: A Closer Look - AP PGECET

    In this blog post, we’ll explore what exactly a restricted response question is and how it differs from other question types. We’ll also delve into why they are used and offer some tips on how to answer them effectively.

  5. Essay type test | PPT - SlideShare

    Restricted Response Essay Questions Restricted response usually limits both the content and the response by restricting the scope of the topic to be discussed. Useful for measuring learning outcomes requiring interpretation and application of data in a specific area.

  6. Writing Better Essay Exams IDEA Paper #76 March 2019 - ed

    Restricted-response, or short-answer, questions likely have expected “correct” responses (e.g., “List the major components of Freytag’s triangle in dramatic structure.”). Extended-response questions (the focus of this paper) are those that typically come to mind when envisioning a traditional essay exam—questions

  7. Types of Questions in Teacher Made Achievement Tests: A ...

    Restricted response essay questions are valuable for assessing specific skills or knowledge within a limited domain. When constructing these items, ensure the question is focused and that students are aware of any word or time limits.

  8. Best Practices for Designing and Grading Exams | CRLT

    Restricted response essay questions are good for assessing basic knowledge and understanding and generally require a brief written response (e.g., “State two hypotheses about why birds migrate. Summarize the evidence supporting each hypothesis” [Worthen, et al., 1993, p. 277].)

  9. Essay Items - SAGE Publications Inc

    An open-ended (or unrestricted response) essay question is one where there are no restrictions on the response, including the amount of time allowed to finish, the number of pages written, or

  10. Assessment Fact Sheet: Constructed Response Assessments (Essays)

    Essay questions can be classified into two types:3 • Restricted response essays; limit both content and response as indicated within the question. The restricted response essay addresses a limited sample of the curriculum or learning outcomes.4 The restricted response essay may commonly be known as: