Assessing Deeper Learning After a Year of Change

The profound disruptions to traditional schooling this past year leave educators with a unique  opportunity to redefine school when we resume in the fall. Though it may be tempting to try to teach  faster to "catch up" on missed content, we contend that the best course of action is to slow down and  go into greater depth to ensure that students learn the most important ideas and skills deeply and are  able to apply them in meaningful ways. 

Others have written about the instructional approaches needed to achieve deep learning (McTighe &  Silver, 2020; Fullan, Quinn, & McEachen, 2019; Mehta & Fine, 2019). But it is especially important to  consider assessment when thinking about deeper learning. 

Deeper Learning, Unpacked 

The National Research Council (2012) defines deeper learning as a "process through which an  individual becomes capable of taking what was learned in one situation and applying it to a new  situation" (p. 5). We propose that a modern education should prepare students to be able to apply  their learning to new situations—in other words, to transfer. Grant Wiggins (2012), co-author  of Understanding by Design, highlights that transfer exists when a student makes use of their learning in  a time, place, and circumstance different from that in which the learning first occurred. But how, exactly, will we know that students have learned deeply? 

This question reflects an if-then logic: If we believe that preparing students to transfer their learning is  a fundamental aim of a modern education, then we need to collect the necessary evidence to  determine the extent to which students can demonstrate deeper learning through transfer. An unfortunate consequence of the standards-based testing movement is that many schools teach  and test grade-level standards in isolation using multiple choice or institute formalized "test prep"  assessments that mimic the format of standardized accountability tests. While such assessment 

methods may provide measures of students' acquisition of knowledge and basic skills, they don't tell  us much about deep understanding and transfer. 

We contend that transfer of learning is shown best through performance assessments that require  students to perform with their learning. Well-designed performance tasks can assess multiple  standards and cut across subject areas, prompting complex thinking and a tangible product or  performance that allows students to apply their learning. An athlete playing in a game needs to know  the rules of the sport, apply the skills, and also display strategic understanding. Though a player could  demonstrate knowledge of the rules on a paper and pencil test or through practice drills, the actual  game requires the player to "put everything together" on the field. Similarly, effective performance  assessments require knowledge, skills, and strategy, applied in context. 

Consider the three examples of performance assessment tasks in Figure 1. What must a student do to  successfully complete each task? How do these assessments differ from selected-response tests?

Figure 1. Three Sample Performance Tasks 

Evaluate a Claim 

The Pooper Scooper Kitty Litter Company claims that their litter is 40 percent more absorbent than other  brands. You are a consumer-advocates researcher who has been asked to evaluate their claim. Develop a  plan for conducting the investigation. Explain how the design of your investigation will enable you to  evaluate the claim. 

Make Your Case 

You have an idea that you believe will make your school better, and you want to convince school leaders  that they should act on your idea. Identify your audience (e.g., principal, PTSA board, the student  government association) and do the following: 

1. Describe your idea. 

2. Explain why and how it will improve the school. 

3. Develop a plan for acting on your idea. 

Your idea and plan can be communicated to your target audience in a letter, an e-mail, or a presentation.  Be sure to choose the means of communication that is most appropriate for your audience and purpose. What's the Pattern? 

Part 1 – Interpret the data on incidents of COVID-19 infections and associated mortality rates on each  continent for the past three months. (Students are given data sets to analyze.) Prepare a chart, podcast, or  newspaper article to help people understand any patterns you detect. 

Part 2 – Select four countries and compare their governmental policies enacted to mitigate the infection's  spread. Provide an explanation of the link between their policies and the data on infections and mortality  rates. Prepare a newspaper article, podcast, or vodcast to present your conclusions.

These examples reveal four characteristics of performance tasks for assessing deeper learning (McTighe, Doubet, & Carbaugh, 2020): 

  1. Performance assessment tasks call for students to (1) apply their learning in some context, and (2) explain what they have done. Whether a task calls for a written response (e.g., an academic essay or blog post), a spoken response (e.g., an audio recording or a live debate), or a visual or  physical communication (e.g., an infographic or an interpretive dance), students need to  convey their reasoning, justify their decisions, and support their interpretations. 
  2. Any performance assessment of deeper learning needs to engage the student in transferring  their learning to a novel situation, different from that in which it was initially learned. Benjamin  Bloom and his colleagues described this application category of their Taxonomy of the  Cognitive Domain in 1954: 
  3. An effective performance assessment task engages students in complex thinking. We recommend the use of Depth of Knowledge, a four-level framework developed by Norman Webb and his colleagues (2005) for analyzing the cognitive complexity of any task. When the  goal is assessing deeper learning, the tasks need to operate at Levels 3 and 4, which require  higher-order thinking such as analysis, interpretation, investigation, problem solving,  argumentation, and design. Use multiple-choice and short answer test and quiz items (at  Levels 1 and 2) to assess factual knowledge and discrete skills, and reserve performance tasks  to assess deeper learning. 
  4. The best performance tasks establish a "real world" context for application, when learners can  effectively apply (i.e., transfer) their learning to realistic situations. 

Setting Goals and Tasks 

Authentic tasks reflect a worthy goal, a target audience, realistic constraints, a tangible product or  performance, and success criteria. They can vary considerably in terms of their time frame, complexity,  nature of the products/performances, and whether the targeted content and skills are discipline specific or interdisciplinary. Such performance tasks can range from conventional essays to open ended mathematics problems to scientific experiments to a research project to tackling a community based issue (Wren & Gareis, 2019). Wiggins and McTighe (2012) offer a practical framework captured in  the acronym GRASPS that can be used to develop an authentic context for an assessment task (see  Figure 2).

Figure 2. GRASPS Framework


- A real-world Goal, 

- A meaningful Role for the student, 

- An authentic (or simulated) Audience, 

- A contextualized Situation that involves real-world application, 

- Student-generated culminating Product and/or Performance, and 

- Success criteria by which student products and performances will be evaluated as evidence of learning.

Figure 3 presents an example of a performance assessment task used as part of the study of a state or  

province. Can you recognize the GRASPS elements contained within the task prompt? 

Figure 3: Sample GRASPS Performance Assessment 

State Tour 

A group of six exchange students is visiting your school for one month as part of an international  exchange program. The principal has asked your class to plan and budget a four-day tour of [your state or province] to help the visitors understand the state's history, geography, economy, and cultural  elements. You should prepare a written tour itinerary, including an explanation of why each site was  selected. Include a map tracing the route for the four-day tour. Optional extension: Include a budget for  the trip.

Attending to Validity and Reliability 

The authentic nature of performance tasks is often motivating to students. However, when these tasks  are used for assessment purposes, they are obligated to meet measurement standards of validity and  reliability. Performance assessment tasks must enable educators to answer key questions: 

  • Does a performance task provide the necessary and sufficient evidence to enable teachers  (and others) to determine the degree of deep learning of a student? 
  • Does the student's performance provide evidence that they can effectively apply their learning  to new situations? 
  • Can we trust that a student's performance is not influenced (whether positively or negatively)  by raw chance, poor construction of the assessment itself, implicit biases, cheating, or  inconsistency in the teacher's evaluation of students' work? 

A common concern about the validity and reliability of performance assessments is that their scoring is  "too subjective." While this can be a potential problem—especially with poorly designed tasks and  scoring rubrics—the challenge is not insurmountable. After all, we use judgment-based evaluations  routinely in state writing assessments, AP art portfolio reviews, judging in Olympic events, and when  we rate a restaurant. Indeed, there are well-established practices that enable performance  assessments and judgment-based evaluation to function fairly, consistently, and defensibly (Gareis &  Grant, 2015; McTighe 2013).

Deeper Learning for School and Life 

There's a cartoon that shows a recent graduate interviewing for a job. Across the desk is a besuited  business executive who has just asked a question of the owl-eyed youngster. On the executive's face is  an expression of dismay. The caption—spoken by the recent graduate—reads something like this: "Can  you give me four choices?" The comic is a satirical commentary on one of the unintended  consequences of high-stakes standardized tests. If we feed students a steady diet of multiple-choice  questions throughout their school careers, should we be surprised that they (and the graduates they  become) will frame problems in terms of a set of selectable "answers"? The stone-faced expression on  the prospective employer's face belies his experience in the "real world," where issues and problems  are often complex and do not lend themselves to fixed-response solutions. Today's students need  deeper learning to successfully navigate the opportunities and challenges they will face in the world  beyond school. Achieving this goal will require shifts in curriculum, instruction, and especially,  assessment practices. 

As Grant Wiggins reminds us, "The point of school is not to get good at school but to effectively parlay  what we learned in school in other learning and in life." Deeper learning enables transfer. Performance  assessments give us the evidence that students are indeed learning deeply—and are able to apply that  learning in school and life. 

 *Article originally published in ASCD Express

About the author:

Jay McTighe has a wealth of experience and a rich and varied career in education. He served as director of the Maryland Assessment Consortium, a collaboration of school districts working to develop and share formative performance assessments and helped lead standards-based reforms at the Maryland State Department of Education.

Chris Gareis is a Professor of Educational Leadership in the School of Education at the College of William & Mary in Williamsburg, Virginia. He also conducts research and development on current issues, including classroom-based assessment, teacher preparation, mentoring, teacher evaluation, teacher compensation, strategic planning, and program evaluation.


Subscribe to the #1 PBL Blog!

Receive new articles in the world of Project Based Learning, STEM/STEAM, and College & Career Readiness.