Defined Learning Educators Blog

Assessing Deeper Learning

Written by Jay McTighe and Chris Gareis | Apr 18, 2022 3:27:00 PM

The profound disruptions to traditional schooling over the past two years leave educators with unique opportunities to redefine school. Though it may be tempting to try to teach faster to "catch up" on missed content, we contend that the best course of action is to slow down and go into greater depth to ensure that students learn the most important ideas and skills deeply and are able to apply them in meaningful ways. 

Others have written about the instructional approaches needed to achieve deep learning (McTighe &  Silver, 2020; Fullan, Quinn, & McEachen, 2019; Mehta & Fine, 2019). But it is especially important to consider assessment when thinking about deeper learning. 

Deeper Learning, Unpacked 

The National Research Council defines deeper learning as a "process through which an individual becomes capable of taking what was learned in one situation and applying it to a new situation". We propose that modern education should prepare students to be able to apply their learning to new situations—in other words, to transfer. Grant Wiggins, co-author of Understanding by Design, highlights that transfer exists when a student makes use of their learning in a time, place, and circumstance different from that in which the learning first occurred. But how, exactly, will we know that students have learned deeply? 

This question reflects an if-then logic: If we believe that preparing students to transfer their learning is a fundamental aim of modern education, then we need to collect the necessary evidence to determine the extent to which students can demonstrate deeper learning through transfer. An unfortunate consequence of the standards-based testing movement is that many schools teach and test grade-level standards in isolation using multiple-choice or institute formalized "test prep"  assessments that mimic the format of standardized accountability tests. While such assessment methods may provide measures of students' acquisition of knowledge and basic skills, they don't tell us much about deep understanding and transfer. 

We contend that transfer of learning is shown best through performance assessments that require students to perform with their learning. Well-designed performance tasks can assess multiple standards and cut across subject areas, prompting complex thinking and a tangible product or performance that allows students to apply their learning. An athlete playing in a game needs to know the rules of the sport, apply the skills, and also display strategic understanding. Though a player could demonstrate knowledge of the rules on a paper and pencil test or through practice drills, the actual game requires the player to "put everything together" on the field. Similarly, effective performance assessments require knowledge, skills, and strategy, applied in context. 

Consider the three examples of performance assessment tasks in Figure 1. What must a student do to successfully complete each task? How do these assessments differ from selected-response tests?

Figure 1. Three Sample Performance Tasks 

Evaluate a Claim 

The Pooper Scooper Kitty Litter Company claims that their litter is 40 percent more absorbent than other brands. You are a consumer-advocates researcher who has been asked to evaluate their claim. Develop a  plan for conducting the investigation. Explain how the design of your investigation will enable you to evaluate the claim. 

Make Your Case 

You have an idea that you believe will make your school better, and you want to convince school leaders that they should act on your idea. Identify your audience (e.g., principal, PTSA board, the student  government association) and do the following: 

1. Describe your idea. 

2. Explain why and how it will improve the school. 

3. Develop a plan for acting on your idea. 

Your idea and plan can be communicated to your target audience in a letter, an e-mail, or a presentation.  Be sure to choose the means of communication that is most appropriate for your audience and purpose. What's the Pattern? 

Part 1 – Interpret the data on incidents of COVID-19 infections and associated mortality rates on each continent for the past three months. (Students are given data sets to analyze.) Prepare a chart, podcast, or newspaper article to help people understand any patterns you detect. 

Part 2 – Select four countries and compare their governmental policies enacted to mitigate the infection's spread. Provide an explanation of the link between their policies and the data on infections and mortality rates. Prepare a newspaper article, podcast, or vodcast to present your conclusions.

These examples reveal four characteristics of performance tasks for assessing deeper learning: 

  1. Performance assessment tasks call for students to (1) apply their learning in some context, and (2) explain what they have done. Whether a task calls for a written response (e.g., an academic essay or blog post), a spoken response (e.g., an audio recording or a live debate), or a visual or  physical communication (e.g., an infographic or an interpretive dance), students need to convey their reasoning, justify their decisions, and support their interpretations. 
  2. Any performance assessment of deeper learning needs to engage the student in transferring their learning to a novel situation, different from that in which it was initially learned. Benjamin  Bloom and his colleagues described this application category in their Taxonomy of the  Cognitive Domain in 1954: 
  3. An effective performance assessment task engages students in complex thinking. We recommend the use of Depth of Knowledge, a four-level framework developed by Norman Webb and his colleagues (2005) for analyzing the cognitive complexity of any task. When the goal is assessing deeper learning, the tasks need to operate at Levels 3 and 4, which require higher-order thinking such as analysis, interpretation, investigation, problem solving,  argumentation, and design. Use multiple-choice and short answer test and quiz items (at  Levels 1 and 2) to assess factual knowledge and discrete skills, and reserve performance tasks to assess deeper learning. 
  4. The best performance tasks establish a "real world" context for application when learners can effectively apply (i.e., transfer) their learning to realistic situations. 

 

Setting Goals and Tasks 

Authentic tasks reflect a worthy goal, a target audience, realistic constraints, a tangible product or performance, and success criteria. They can vary considerably in terms of their time frame, complexity,  nature of the products/performances, and whether the targeted content and skills are discipline-specific or interdisciplinary. Such performance tasks can range from conventional essays to open-ended mathematics problems to scientific experiments to a research project to tackle a community-based issue. Wiggins and McTighe offer a practical framework captured in the acronym GRASPS that can be used to develop an authentic context for an assessment task (see  Figure 2).

Figure 2. GRASPS Framework 

GRASPS 

- A real-world Goal, 

- A meaningful Role for the student, 

- An authentic (or simulated) Audience, 

- A contextualized Situation that involves real-world application, 

- Student-generated culminating Product and/or Performance, and 

- Success criteria by which student products and performances will be evaluated as evidence of learning.

Figure 3 presents an example of a performance assessment task used as part of the study of a state or  

province. Can you recognize the GRASPS elements contained within the task prompt? 

Figure 3: Sample GRASPS Performance Assessment 

State Tour 

A group of six exchange students is visiting your school for one month as part of an international exchange program. The principal has asked your class to plan and budget a four-day tour of [your state or province] to help the visitors understand the state's history, geography, economy, and cultural elements. You should prepare a written tour itinerary, including an explanation of why each site was selected. Include a map tracing the route for the four-day tour. Optional extension: Include a budget for the trip.

 

Attending to Validity and Reliability 

The authentic nature of performance tasks is often motivating to students. However, when these tasks are used for assessment purposes, they are obligated to meet measurement standards of validity and reliability. Performance assessment tasks must enable educators to answer key questions: 

  • Does a performance task provide the necessary and sufficient evidence to enable teachers  (and others) to determine the degree of deep learning of a student? 
  • Does the student's performance provide evidence that they can effectively apply their learning to new situations? 
  • Can we trust that a student's performance is not influenced (whether positively or negatively)  by raw chance, poor construction of the assessment itself, implicit biases, cheating, or inconsistency in the teacher's evaluation of students' work? 

A common concern about the validity and reliability of performance assessments is that their scoring is  "too subjective." While this can be a potential problem—especially with poorly designed tasks and scoring rubrics—the challenge is not insurmountable. After all, we use judgment-based evaluations routinely in state writing assessments, AP art portfolio reviews, judging in Olympic events, and when we rate a restaurant. Indeed, there are well-established practices that enable performance assessments and judgment-based evaluations to function fairly, consistently, and defensibly.

Deeper Learning for School and Life 

There's a cartoon that shows a recent graduate interviewing for a job. Across the desk is a besuited business executive who has just asked a question of the owl-eyed youngster. On the executive's face is an expression of dismay. The caption—spoken by the recent graduate—reads something like this: "Can  you give me four choices?" The comic is a satirical commentary on one of the unintended consequences of high-stakes standardized tests. If we feed students a steady diet of multiple-choice questions throughout their school careers, should we be surprised that they (and the graduates they become) will frame problems in terms of a set of selectable "answers"? The stone-faced expression on the prospective employer's face belies his experience in the "real world," where issues and problems are often complex and do not lend themselves to fixed-response solutions. Today's students need deeper learning to successfully navigate the opportunities and challenges they will face in the world beyond school. Achieving this goal will require shifts in curriculum, instruction, and especially,  assessment practices. 

As Grant Wiggins reminds us, "The point of school is not to get good at school but to effectively parlay  what we learned in school in other learning and in life." Deeper learning enables transfer. Performance assessments give us the evidence that students are indeed learning deeply—and are able to apply that learning in school and life. 

 *Article originally published in ASCD Express

About the authors:

Jay McTighe has a wealth of experience and a rich and varied career in education. He served as director of the Maryland Assessment Consortium, a collaboration of school districts working to develop and share formative performance assessments, and helped lead standards-based reforms at the Maryland State Department of Education.

Chris Gareis is a Professor of Educational Leadership in the School of Education at the College of William & Mary in Williamsburg, Virginia. He also conducts research and development on current issues, including classroom-based assessment, teacher preparation, mentoring, teacher evaluation, teacher compensation, strategic planning, and program evaluation.