Amara had been volunteering at the campus tutoring center for three weeks when she realized something uncomfortable: she had no idea if her tutee was actually learning anything.
In This Chapter
- The Difference Between Teaching and Helping Someone Learn
- Cognitive Load Theory: The Designer's Framework
- The Worked Example Effect
- Scaffolding and Fading in Practice
- Retrieval Practice in the Classroom
- Spacing in Curriculum Design
- Interleaving Across Lessons
- Learning Objectives That Are Actually Meaningful
- Designing for Transfer
- Feedback Systems in Educational Design
- Amara Redesigns Her Tutoring Session
- Try This Right Now
- The Progressive Project
- A Final Note on the Gap Between Research and Practice
Chapter 34: Designing Learning Experiences
Amara had been volunteering at the campus tutoring center for three weeks when she realized something uncomfortable: she had no idea if her tutee was actually learning anything.
The sessions felt good. Amara would explain a concept — cell membrane transport, say, or the mechanics of a hormone feedback loop — and Jordan would nod and say "okay" and "that makes sense" and occasionally "oh, I get it now." Amara would feel the satisfying click of a clear explanation landing. She'd leave the session feeling like she'd helped.
Then Jordan would come back the following week and not remember most of what they'd covered.
The first time it happened, Amara chalked it up to Jordan not reviewing between sessions. The second time, she figured the material had been particularly hard. By the third time, she started to wonder if the problem was her.
She mentioned it to her supervisor at the tutoring center, a graduate student in education named Thomas. He asked a question that reframed everything: "When Jordan says 'that makes sense,' what does that actually mean?"
Amara started to say "it means she understands it," and then stopped. Because she didn't know. She had no way to tell. Jordan's nod of comprehension and Jordan's actual comprehension were two entirely separate things, and Amara had been measuring one and calling it the other.
Thomas smiled. "Welcome to the central problem of teaching."
The Difference Between Teaching and Helping Someone Learn
There's a version of teaching that is really just transmission: you have information, you convey it clearly, the other person receives it. On this model, the quality of teaching is measured by the clarity of the explanation. A brilliant explanation is great teaching.
This model is wrong, and the wrongness has consequences for every teacher, trainer, coach, and tutor in the world.
Information transmission is not learning. Sitting in a lecture, following a tutorial, watching an explainer video, reading a clear textbook — these are experiences of receiving information, but receiving information is not the same as learning it. The learner's subjective sense that something "makes sense" is a feeling of comprehension in the moment, not a record of durable memory. The feeling and the learning are, at best, weakly correlated.
This is not a flaw in your learners. It's a property of how human memory works. Understanding something in the moment requires that the information be active in working memory and coherent to the conscious mind. Remembering it later requires that the information have been encoded into long-term memory — a completely different process, one that requires effortful retrieval, not passive reception.
Amara's tutoring sessions were producing moments of comprehension. They were not producing durable memories, because nothing in the sessions was designed to produce durable memories.
The distinction between transmission and construction matters.
Transmission models treat the learner as a vessel to be filled. The instructor has knowledge; the learner lacks it; the instruction transfers it. Learning is measured by whether the transfer was received correctly.
Construction models treat learning as something the learner has to build. The instructor can create conditions under which construction is more or less likely — providing the materials, pointing at what needs to be built, offering scaffolding when the structure starts to lean — but no one can do the building for someone else. Learning is something the learner does, not something that happens to them.
Every principle in this book has been about the learner as active constructor: retrieval practice works because the act of retrieving builds the memory, not the act of reading. Spaced practice works because the effort of remembering after a gap is what consolidates the trace. Interleaving works because the problem of distinguishing between similar things is what forces deep processing.
When you design learning experiences for other people, the same principles apply. The question is not "how clearly can I explain this?" The question is "what will learners be doing that causes them to learn?"
[Evidence: Strong] The research comparing passive instruction (lecture, reading) to active instruction (retrieval practice, problem-solving, discussion) is among the most consistent in educational psychology. Freeman et al.'s (2014) meta-analysis of 225 studies found that active learning produced significantly higher exam scores and substantially lower failure rates than traditional lecture in undergraduate STEM education.
Cognitive Load Theory: The Designer's Framework
Before you can design effective learning experiences, you need a model of what's happening inside the learner's head when they're trying to learn. The most practically useful model available is John Sweller's cognitive load theory.
Here's the core insight in plain terms: your working memory — the part of your mind that actively holds and processes information — is severely limited. It can hold roughly four to seven distinct pieces of information at a time. When a learning experience demands more than that, cognitive overload happens: processing breaks down, understanding fails, and the learner often experiences frustration and disengagement without knowing why.
Most instruction routinely triggers cognitive overload without anyone realizing it. The instructor knows the material so well that what feels like a simple explanation is, for a novice, a cascade of unfamiliar terms, new relationships, and competing demands. The diagram that's "perfectly clear" to the designer requires five simultaneous mental operations from someone seeing it for the first time.
[Evidence: Strong] Cognitive load theory is one of the most extensively validated frameworks in instructional design, with over three decades of experimental research across dozens of domains, age groups, and instructional formats. Its core claims have been replicated far more robustly than most educational research.
Sweller identifies three types of cognitive load, and understanding the distinction between them is the key to better design.
Intrinsic Load
Intrinsic load is the inherent difficulty of the material itself — the number of interacting elements that must be held in working memory simultaneously to understand the concept. You cannot remove intrinsic load without removing the learning. Cell signaling pathways have high intrinsic load because understanding them requires simultaneously tracking multiple proteins, reactions, and cascades. Spelling the word "cat" has low intrinsic load.
What you can do with intrinsic load is manage it. You can sequence instruction so that complex concepts are introduced after their components have been learned individually. You can simplify early examples to the essential structure before introducing complications. You can chunk information so learners master one component at a time before integrating. None of this removes the difficulty — it sequences it so working memory isn't overwhelmed all at once.
Extraneous Load
Extraneous load is cognitive effort caused by poor design, not by the difficulty of the content. It's the mental work of dealing with a badly organized handout, deciphering jargon that could have been plain language, cross-referencing a diagram with a legend that should have been integrated, or processing text on a slide while simultaneously trying to hear a lecture. This load consumes cognitive resources without contributing to learning.
Extraneous load is waste. It uses up the limited working memory budget for navigation rather than learning.
Common sources of extraneous load: - Split-attention effects: When information that needs to be mentally integrated is presented in separate locations. The diagram of the cell cycle with a separate legend. The equation on the board that references variables defined in the textbook. The slide full of text that contradicts what the instructor is saying. Learners must mentally hold one piece while searching for the other. - Redundancy effects: When the same information is presented simultaneously in multiple formats (reading exactly what's on the slide aloud, for example) — the learner must process two channels saying the same thing, which wastes capacity. - Seductive details: Interesting but irrelevant additions that draw attention without contributing to the learning objective. A vivid anecdote that doesn't connect to the concept. A photograph that's thematically relevant but doesn't add analytical value. These feel like good teaching (engaging!) but consume cognitive resources. - Poorly organized structure: When the logical order of a presentation requires the learner to mentally reorganize it to make sense of it.
The design principle: ruthlessly eliminate extraneous load. Every element of an instructional experience should earn its place by contributing to the learning objective.
Germane Load
Germane load is the cognitive effort that actually produces learning. Trying to solve a problem. Explaining a concept in your own words. Connecting new information to prior knowledge. Noticing similarities and differences between concepts. Generating examples. This effort is demanding, but it's the good kind of demanding: it's the effort of building understanding.
The goal of learning design is a simple equation: minimize extraneous load, which frees cognitive resources for germane load.
When a presentation has slides full of text, visual clutter, poorly organized content, and jargon that hasn't been defined, learners are spending their cognitive budget on extraneous load. There's little left for germane. When the design is clean, the content is organized, and the visuals are integrated — learners can direct their full cognitive resources toward the hard work of actually understanding.
Practical Applications
Integrate text with visuals. Don't put labels in a separate legend; put them on the diagram. Don't describe a process in a caption below a figure; annotate the figure directly. Reduce the split-attention demand and cognitive resources are freed.
Minimal text on slides. Slides full of sentences create redundancy overload when the instructor is also speaking. Key words, diagrams, and structured outlines let the spoken word carry the explanatory load while the visual carries the conceptual structure. The learner's attention can integrate the two rather than having to process two competing streams of full language.
Eliminate the irrelevant. If an element of your design doesn't serve the learning objective, remove it. The interesting digression, the tangentially related video clip, the additional complexity that makes the example feel more realistic — all of these cost cognitive resources that could go toward the core learning.
Sequence complexity. Introduce the simplified version of a concept before introducing complications. Teach the basic mechanism before the exceptions. Let learners build a schema for the simple case before loading it with the full complexity.
The Worked Example Effect
Here is one of the most robust and practically important findings in all of instructional design, and most instructors have never heard of it.
For novice learners, studying worked examples of solved problems produces better learning outcomes than attempting to solve equivalent problems independently — even when total time on task is the same.
Let that sink in for a moment. Studying examples of solved problems is more effective than actually solving problems. This seems backwards. Isn't doing the work what builds the skill?
For novices, no. Here's why.
When a novice tries to solve an unfamiliar problem, almost all of their working memory is consumed by the management of the solution process: deciding what to try, keeping track of where they are, checking for errors, deciding whether to abandon a dead end and try something else. This navigation overhead leaves almost no cognitive capacity for understanding the underlying principles that make the problem solvable. They might eventually arrive at a correct answer through trial and error, but the cognitive work has gone into navigation, not learning.
When a novice studies a worked example, the navigation burden is gone. The solution is right there. The learner's cognitive resources can go entirely toward understanding why each step was taken, what principle it applied, and how the problem-solver knew what to do next. This is where the learning lives.
[Evidence: Strong] Across dozens of studies spanning mathematics, physics, chemistry, medicine, and computer science, worked examples consistently outperform equivalent problem-solving for novice learners in terms of both efficiency (time spent) and outcomes (performance on subsequent tests).
There is one crucial qualification: the learning benefit requires self-explanation. A learner who passively reads a worked example — who follows the steps without asking why — gets very little benefit. The learning mechanism is the active process of explaining to yourself why each step was taken.
Self-explanation prompts built into worked examples dramatically increase their effectiveness: - "Why was this step taken at this point?" - "What principle does this step apply?" - "What would have happened if the problem-solver had done X instead?" - "How would this approach change if the initial conditions were different?"
These questions prevent passive processing. They force the germane cognitive work that actually builds understanding.
The Expertise Reversal Effect
Here's the complication: the worked example advantage disappears — and eventually reverses — as learners develop expertise.
For an intermediate learner who has already built some competence, attempting to solve problems independently is more effective than studying worked examples. For an advanced learner, worked examples may actually impede learning by forcing them to process a provided solution when their existing knowledge would allow them to generate a better one themselves.
This is the expertise reversal effect: what helps novices hurts experts, and vice versa. The instructional design that's optimal for one learner may be counterproductive for another, depending on their level of expertise.
The practical implication: you need to know where your learners are. And as they develop, your design needs to change with them.
Fading: The Bridge from Worked to Independent
The research-supported solution is a graduated transition called fading: you begin with fully worked examples and systematically reduce the scaffolding as competence develops.
The fading sequence:
Stage 1: Fully worked example. Every step solved and explained. Appropriate for first exposure to a completely new type of problem. The learner's job is to understand, not solve.
Stage 2: Partially worked example. The first two-thirds of the solution are provided; the learner completes the remaining steps. Appropriate when the learner can recognize the problem type and apply the early steps confidently.
Stage 3: Problem with a hint or first step. The learner does nearly all the work, with minimal scaffolding to prevent the initial failure that consumes too much cognitive load. Appropriate when the learner succeeds most of the time independently but sometimes gets stuck at the start.
Stage 4: Fully independent problem. No scaffolding. Appropriate when the learner has demonstrated consistent accuracy.
This progression — worked to faded to independent — substantially outperforms both extremes: either keeping worked examples forever (which creates dependency) or immediately giving independent problems (which overwhelms novices). The middle path gives novices the support they need and expert learners the challenge they need.
[Evidence: Strong] Research by Renkl, Atkinson, and colleagues across multiple studies consistently demonstrates advantages for faded worked examples over both full worked examples and pure problem-solving, particularly for the transition from novice to intermediate competence.
Scaffolding and Fading in Practice
The scaffold metaphor is precise and worth taking literally: scaffolding on a building provides structural support while the building is under construction and cannot support itself. When the structure can stand on its own, the scaffolding is removed. Instructional scaffolding works the same way.
Lev Vygotsky's concept of the zone of proximal development describes the region between what a learner can do independently (too easy, not learning) and what they cannot do even with help (too hard, also not learning). The productive zone — where the learning happens — is the range of tasks that are too hard to do alone but manageable with appropriate support.
Good instructional design keeps learners in their zone of proximal development: challenging enough to require real effort, but supported enough that the effort is productive rather than paralytic.
Types of scaffolding available to instructional designers:
Structural scaffolding provides the architecture for thinking. Frameworks, templates, and cognitive tools that guide learners through a process without doing the thinking for them. A diagnostic algorithm in medical education — "when you see X, check for Y and Z" — is scaffolding. It doesn't diagnose the patient; it guides the reasoning. A writing framework ("claim, evidence, reasoning") is scaffolding. It doesn't write the essay; it structures the process.
Procedural scaffolding breaks complex tasks into explicit steps. Checklists, worked examples, step-by-step procedures, and decision trees all fall here. They reduce the cognitive load of figuring out what to do next, freeing capacity for understanding the content.
Metacognitive scaffolding prompts learners to monitor their own understanding. "What's the main idea here?" "What would you do if the next question looked like this?" "Where are you uncertain?" These prompts are particularly valuable for learners who haven't yet developed strong metacognitive habits.
Modeling shows learners what expert thinking looks like, explicitly. A surgeon who thinks aloud during a procedure, narrating their reasoning, is providing modeling scaffolding. An instructor who works through a problem on the board while making their decision process visible — including dead ends and corrections — is doing the same.
The dependency trap.
There is a failure mode in scaffolding called the dependency trap: scaffolding that is never removed produces learners who can only perform with support. The student who always has the formula sheet never internalizes the formulas. The trainee who always has the checklist never develops the diagnostic intuition the checklist was designed to build.
Scaffolding should be explicitly temporary. Learners should know that support will be gradually reduced. And the reduction should actually happen, based on demonstrated competence — not just assumed competence.
[Evidence: Moderate] Research on scaffolding and guided practice is extensive but methodologically varied. The clearest findings support the benefit of graduated support with explicit fading over both permanent scaffolding and immediate independence.
Signals that scaffolding can be faded: - Learners complete scaffolded tasks with consistently high accuracy (above 80-85%) - The time to complete with scaffold approaches the time to complete without - Learners can articulate the reasoning behind their steps, not just the steps themselves - Learners report the scaffold feeling unnecessary or obvious
Retrieval Practice in the Classroom
The single most evidence-backed learning intervention available to any instructor — and the one most consistently underused — is building retrieval practice into instruction.
Every class session that ends with learners having passively received information, without any retrieval activity, has underperformed. Every class that builds even a few minutes of active retrieval outperforms the equivalent passive session, measurably, on delayed assessments.
The evidence for this is overwhelming and consistent across subjects, ages, and formats. The only remaining question is implementation: how do you build retrieval practice into a course without it feeling punitive or taking over the curriculum?
[Evidence: Strong] Roediger, Karpicke, Agarwal, and colleagues have produced a decades-long body of classroom research — not just lab studies — demonstrating that frequent, low-stakes quizzing produces better retention than equivalent time spent on review or re-study. The effect holds in K-12 classrooms, university courses, and corporate training environments.
Opening Retrieval: The Highest-Return Five Minutes
The simplest and most powerful retrieval intervention: begin every class session with a brief retrieval of the previous session's material.
The procedure: before introducing new material, ask learners to take out paper (or open a document) and write down, from memory, three to five key concepts, mechanisms, or ideas from the previous session. Two to three minutes of quiet writing, then a brief review of the answers.
This takes five to seven minutes total. What it produces:
First, it provides spaced retrieval practice on the previous session's material. The gap between the previous class and today is a natural spacing interval — enough time for some forgetting to have occurred, which is exactly when retrieval is most beneficial for consolidating the memory.
Second, it activates prior knowledge before new material is introduced. Research on "knowledge activation" consistently shows that learners who enter a new instructional episode with relevant prior knowledge active in working memory integrate the new material more deeply. The retrieval exercise does this automatically.
Third, it creates accountability. Learners who know the first five minutes of every session will require retrieval from the previous session tend to review between sessions — not because they're told to, but because they know they'll need it.
The reframe for resistant learners and instructors: this is not a quiz. It's the most efficient form of studying available, compressed into five minutes at the start of class.
Embedded Questions During Instruction
Every fifteen to twenty minutes in a lecture or presentation, pause and ask a retrieval question about what was just covered.
Not a comprehension check ("did everyone follow that?") — an actual retrieval demand: "Without looking at your notes, what are the three mechanisms of antibiotic resistance we just discussed? Write them down. You have sixty seconds."
The brief pause and retrieval interrupts passive processing. During a lecture, information tends to flow through working memory without deep encoding — learners follow along but don't consolidate. The interruption for retrieval forces the active processing that consolidation requires.
Implementation options, from lowest to highest technology: - Paper retrieval (students write answers; instructor briefly reviews) - Think-pair-share (students write individually, then discuss with a partner before class review) - Whiteboards or mini-chalkboards (individual students write and hold up responses simultaneously, allowing instant instructor scan of the room) - Digital polling platforms (Poll Everywhere, Mentimeter, Kahoot — allow real-time visualization of response distribution across the class) - Cold calling on specific students (effective but requires a psychologically safe environment where getting it wrong is normatively okay)
The questions can be simple. "Name the two types we just covered." "What's the key difference between X and Y?" "What would happen in this scenario given what we just learned?" Complexity matters less than the act of retrieval.
Exit Tickets: The Last Five Minutes
The exit ticket is one of the most information-dense investments of class time available. Before learners leave, a two-to-three-minute written response that serves double duty: retrieval practice for learners, formative data for the instructor.
The simplest version: "Write down the three most important ideas from today's session, and one thing you're still uncertain about."
The retrieval component forces synthesis and recall. The confusion component gives the instructor honest, anonymous information about where the class is — information that arrives before the next session, when it can still be acted on.
The generative version applies the session's concept to a new situation: "Using what we covered today about X, predict what would happen in this scenario..." This demands transfer — the highest-value retrieval activity, because it tests genuine understanding rather than rote recall.
One caveat: exit tickets only work as formative assessment if the instructor actually uses them. Learners who watch their exit tickets get collected and never see any response — no acknowledgment, no "I noticed many of you were confused about Y, so let's start there today" — learn quickly that the exercise is performative. Respond to what the exit tickets tell you.
Low-Stakes Quizzing: The Evidence
[Evidence: Strong] Agarwal and colleagues' classroom studies with middle and high school students showed that frequent, low-stakes quizzing on previously covered material produced 10-15% better performance on final exams compared to equivalent time spent on review. The effect was consistent across subjects and persisted at delayed assessments administered a full year later.
"Low-stakes" is not a minor modifier. It's load-bearing. High-stakes frequent quizzing creates evaluation anxiety that impairs the learning benefit. The quiz needs to count for something — zero stakes creates zero engagement — but very little. A structure where many small quizzes together account for 10-15% of a course grade creates the right conditions: enough incentive to take them seriously, not enough to make each one stressful.
Learners who experience this structure consistently report, after initial resistance, that the quizzes feel like studying rather than testing. That reframe is accurate.
Spacing in Curriculum Design
Most curricula have a fatal structural flaw: they teach material to mastery, then move on and never formally return.
The logic seems reasonable: once students understand something, they've learned it. Time to cover the next thing. Reviewing old material is going backward, not forward. There's too much content to keep revisiting what's already been done.
This logic is wrong in a way that quietly destroys learning across entire courses. Material from early in a course is largely forgotten by the end — even by students who scored well on the early assessments — because the curriculum has no mechanism that forces students to maintain it. The midterm material is gone by finals. Week one is gone by week ten.
The solution is structural: build review of prior material into the curriculum itself. Don't rely on students to space their own review. They generally don't, and even when they try, they don't do it systematically.
The spiral curriculum. Jerome Bruner proposed the spiral curriculum in 1960: revisit concepts multiple times across a course (or across a curriculum), each time at greater depth and complexity. The concept is never "done" — it reappears, richer, connected to more of the knowledge that has developed since.
Applied to an introductory biology course:
Week 3: Introduce cell membrane transport as a basic concept — what it is, why it matters, the basic mechanisms.
Week 8: Revisit membrane transport in the context of specific organ systems — how kidney tubules use selective transport for filtration and reabsorption.
Week 14: Revisit membrane transport in the context of disease — how disrupted transport mechanisms underlie specific conditions, and how drug interventions target those mechanisms.
Each return encounter is spaced, more complex, and integrated into a larger knowledge structure. The concept grows rather than being filed away.
[Evidence: Strong] Cepeda et al.'s (2006) meta-analysis of over 200 studies on spacing effects found that distributed practice consistently outperforms massed practice for long-term retention, with effect sizes that are among the largest in educational psychology.
Cumulative Assessment
The simplest structural intervention that forces spaced retrieval without changing instruction: make assessments cumulative.
Exams that include material from previous units — not just the most recent one — change learner behavior automatically. If 30% of every exam includes prior material, students cannot afford to study only the current unit. They have to maintain earlier learning. This is the spaced retrieval practice built into the course structure, with no additional instructional effort required.
Most courses fail to do this. Every exam covers only the recent material; everything before it can be forgotten without consequence. Cumulative assessment closes this gap.
Interleaved Problem Sets
For courses with practice problems or homework exercises, the organization of those problems matters more than most instructors realize.
Blocked practice: All problems from Chapter 5 are grouped together; all Chapter 6 problems follow. This is how most textbooks and courses organize practice. It feels efficient because each problem is immediately preceded by similar problems, so learners can apply the same approach repeatedly.
Interleaved practice: Problems from different chapters and types are mixed together, in no particular order. This feels harder — and it is harder. The learner must first identify what type of problem they're facing before they can solve it. This extra step is cognitively demanding.
The interleaved version produces significantly better long-term retention and transfer. The difficulty is the learning. The discrimination between problem types that feels like an obstacle is actually the cognitive work that builds flexible understanding.
[Evidence: Strong] Rohrer and Taylor's (2007) research on interleaved practice in mathematics showed substantially better exam performance for interleaved vs. blocked practice, even when total practice time was equivalent. The effect held at both immediate and delayed testing.
Interleaving Across Lessons
Interleaving applies not just within a problem set, but across lessons and units. Mixing the order in which topics are covered — rather than completing one topic fully before moving to the next — produces better long-term retention, even though it produces more confusion during learning.
Most curricula are organized for thematic coherence: all the material on Topic A, then all the material on Topic B, then all the material on Topic C. This organizational logic serves the instructor's presentation needs. It does not serve the learner's memory needs.
Why interleaving works across lessons:
When topics are blocked — presented one at a time, to completion — learners can apply the same mental approach to every item in the block. There's no need to discriminate between approaches. When topics are interleaved, learners must first identify what kind of thing they're looking at before they can respond to it. This identification step is exactly what transfer requires: the ability to recognize which concept applies in a new situation.
The discrimination difficulty is the learning. Every time a learner pauses to figure out which approach applies, they're practicing the most important skill: reading a situation and identifying the relevant principle.
Implementing interleaving in curriculum design:
For instructors: mix review of earlier topics into current instructional sessions. Don't complete Unit 3 entirely before beginning Unit 4 — weave in Unit 1 and Unit 2 material as the course progresses.
For assessment: interleaved assessments — tests that mix question types from multiple units without labeling which unit each question comes from — are more authentic tests of real-world performance than blocked assessments. In the real world, problems don't come pre-labeled with which approach to use.
For learners doing problem sets: deliberately mix problem types. Don't exhaust all questions of one type before moving to the next. Shuffle, scramble, and mix. The resulting confusion is the point.
The resistance to interleaving is real and predictable. Students and trainees consistently prefer blocked practice because it feels more efficient. They complete more questions per hour, feel more productive, and report less confusion. They are also retaining less and will perform worse on delayed tests.
This is one of the clearest places where what feels good during learning and what produces good learning are directly opposed. The instructor who knows this can design against the preference, explain why, and help learners reframe the confusion as a sign that learning is happening rather than a sign that the instruction is failing.
Learning Objectives That Are Actually Meaningful
Learning objectives are the most frequently written and least frequently useful component of instructional design. Most learning objectives are vague ("students will understand X"), unmeasurable ("learners will appreciate Y"), or describe activities rather than outcomes ("participants will complete a module on Z").
The problem with vague objectives is not cosmetic. Objectives drive assessment design, and assessment drives what learners study and how instructors teach. A vague objective leads to an assessment that measures the wrong thing, which leads to learners who have optimized for the wrong performance.
Bloom's Taxonomy provides a practical vocabulary for writing objectives that specify the cognitive level required, not just the content domain:
Level 1 — Remember: Recall facts and basic concepts. Verbs: name, list, recall, define. Example: "Name the six levels of Bloom's taxonomy."
Level 2 — Understand: Explain ideas in your own words. Verbs: explain, summarize, classify, describe. Example: "Explain the difference between intrinsic and extraneous cognitive load."
Level 3 — Apply: Use information in a new situation. Verbs: use, solve, demonstrate, calculate. Example: "Apply cognitive load principles to identify sources of extraneous load in a sample instructional design."
Level 4 — Analyze: Draw connections, examine structure. Verbs: differentiate, compare, contrast. Example: "Compare blocked and interleaved practice and analyze when each is appropriate."
Level 5 — Evaluate: Make judgments based on criteria. Verbs: assess, judge, defend, critique. Example: "Evaluate a given course design and defend specific changes that would improve learning outcomes."
Level 6 — Create: Produce new work from multiple elements. Verbs: design, build, construct, develop. Example: "Design a three-session tutoring curriculum that incorporates retrieval practice, scaffolding, and formative feedback."
The misalignment problem. Most courses have objectives at Level 1-2 but test at Level 3-4. Students who prepare for Level 1-2 performance are blindsided by assessments requiring Level 3-4 performance. Writing objectives at the actual level required by your assessment — and telling learners what that level is — aligns the whole system.
The useful formula: "By the end of this unit, learners will be able to [Bloom's verb] [specific content] [in this context or to this standard]."
"By the end of this unit, learners will be able to apply the diagnostic criteria for major depressive disorder to novel clinical vignettes and identify cases that do and do not meet criteria." This objective specifies content, cognitive level, and context. It's assessable. It tells the learner exactly what performance is expected. It drives both study behavior and assessment design.
Designing for Transfer
The purpose of education is not performance on the original task. It's the ability to use what you've learned in new situations — transfer. A medical student who can answer anatomy exam questions but can't apply anatomy to clinical diagnosis has learned the wrong thing. A business school graduate who can solve textbook cases but freezes in real organizational situations has been trained for the test, not the job.
Transfer is the hardest outcome to produce, and most instruction doesn't systematically design for it. But there are principles that increase the likelihood of transfer.
Variety of examples. Transfer requires the ability to recognize when a principle applies, not just the ability to apply it in familiar contexts. Learners who encounter a concept through many varied examples — different domains, different surface features, different contexts — develop a more flexible representation of the underlying principle. They learn to see through the surface to the structure.
A learner who sees Newton's second law applied only in physics problems with standard formats learns to recognize physics-problem formatting, not Newton's second law. A learner who encounters the same principle applied in biomechanics, vehicle dynamics, sports science, and everyday situations develops a representation of the principle itself.
Explicit abstraction. After learners work through specific examples, explicitly identify the underlying principle. "In all of these cases, what's the same? What's the general rule that connects them?" This kind of abstraction work — connecting specifics to a general principle — is what cognitive scientists call "schema formation," and it's strongly predictive of transfer.
Far transfer requires deliberate design. Near transfer — applying a skill in a slightly different but structurally similar situation — is relatively easy to produce. Far transfer — applying a principle across substantially different domains or contexts — is rare and requires intentional design. The way to produce far transfer is to expose learners to the principle across genuinely different contexts and then explicitly discuss the connections.
The surface features problem. A well-documented finding in transfer research is that learners often fail to recognize that a principle applies in a new context when the surface features differ — even when the underlying structure is identical. Students who have learned to solve a particular type of physics problem in a laboratory context may fail to recognize that the same principles apply to a real-world scenario with different objects, language, and framing.
The solution is dual: deliberate variety of examples (so learners don't associate the principle with one surface format) and explicit attention-direction ("notice that these two seemingly different problems have the same underlying structure"). Calling attention to structure, explicitly and repeatedly, builds the ability to see through surface to principle — and that ability is the foundation of genuine transfer.
Transfer and assessment alignment. One of the most common failures of instructional design is assessing transfer when instruction produced only performance on the original task. If you teach learners to solve problems in one context and then test them in a different context without any transition instruction for that context, poor transfer performance tells you nothing useful. If you want to design for transfer, your assessment must test in varied contexts — and your instruction must explicitly practice applying the principle in multiple contexts.
[Evidence: Moderate] The research on transfer is less consistent than research on retention. Near transfer is reliably produced by good instruction. Far transfer is more variable and more dependent on explicit abstraction and variety of examples. Instruction that explicitly labels underlying principles and practices applying them in varied contexts consistently outperforms instruction that covers the same material without this explicit structure work.
Feedback Systems in Educational Design
Feedback is the most powerful tool in instructional design and the most frequently implemented poorly.
Most feedback in education is evaluative: correct or incorrect, a grade, a score. This feedback tells learners how they did. It doesn't tell them how to do better.
Consider the difference:
"Your answer was wrong." (Evaluative feedback)
"Your answer treated these two mechanisms as equivalent when they operate through different pathways. You correctly identified the starting point but made an error at the step where the processes diverge. The key distinction is Y." (Corrective feedback)
The second response requires more time to write, but it produces fundamentally different learning. The learner now understands not just that they were wrong but why, and they have information they can use to reconstruct their understanding.
The characteristics of feedback that actually works:
Specificity. Feedback that precisely identifies what went wrong and where in the reasoning the error occurred is far more useful than feedback that identifies that something went wrong. "You need to work on this" is not feedback; it's a verdict.
Timing. Feedback that arrives while the material is still relevant produces learning. Feedback on a midterm exam, returned three weeks later when the course has moved on to different material, has minimal impact on understanding of the tested material. For formative purposes, same-day or next-day feedback is the target.
Forward direction. Effective feedback is future-oriented: what specifically should the learner do differently? Not just "this was wrong" but "next time you encounter a problem like this, the distinguishing feature to check for is X."
Process focus over outcome focus. Telling learners what they got wrong (outcome feedback) is less effective for improvement than telling them specifically where in their reasoning process the error occurred (process feedback). An instructor who says "your reasoning about the mechanism was sound, but you applied an incorrect assumption when you reached this specific step" gives the learner a map of where their thinking broke down.
Feedback that replaces thinking is not feedback. An instructor who over-explains, who provides such complete corrective feedback that the learner never has to figure anything out, is doing the cognitive work for the learner. Effective feedback is calibrated to provide the minimum information needed for the learner to reach understanding — not to reach understanding for them.
The key question: after receiving this feedback, does the learner have to do any thinking? If the answer is no — if the feedback is so complete that comprehension is delivered rather than constructed — it's over-feedback, and it will not produce durable learning.
[Evidence: Moderate-Strong] Hattie and Timperley's (2007) meta-analysis of feedback research identified four levels of feedback with increasing effectiveness: feedback about the task, about the process, about self-regulation, and about the person. The first three levels consistently outperform outcome feedback; feedback about the person ("you're smart" or "you're failing") is often counterproductive.
Feedback Timing and the Spacing Effect
An important design decision in any learning experience: when should feedback arrive?
For formative feedback intended to improve subsequent learning, earlier is usually better. Feedback that arrives the same day as the performance allows learners to connect the correction directly to the reasoning that produced the error. Feedback that arrives two weeks later — when the learner may not clearly remember the thinking that led to the wrong answer — has reduced learning value.
But there is a nuance for retrieval practice. Research on "feedback delay" shows that for retrieval-based activities, slightly delayed feedback — a few minutes to a day — can be more beneficial than immediate feedback, because the time gap allows learners to attempt to retrieve the correct answer before it's provided. The retrieval attempt itself, even an incorrect one, strengthens subsequent memory. Immediate feedback that arrives before the learner has made a genuine attempt to retrieve deprives them of this benefit.
The design principle: for practice that follows instruction immediately, delay feedback slightly to create retrieval opportunity. For practice done independently over longer periods, get feedback back as quickly as possible so the learner can connect the correction to still-active memory.
Self-Generated Feedback
The highest-quality feedback in many learning contexts is feedback that learners generate themselves — through self-testing, self-explanation, peer discussion, and application to new problems.
Peer review is a particularly powerful form of feedback generation: explaining to someone else what's wrong with their work requires understanding the correct version well enough to articulate the gap. The reviewer often learns as much as the reviewee. This is the protégé effect applied to feedback: teaching or evaluating someone else's work activates all the generative processes that produce durable understanding.
Designing learning experiences to include peer feedback, self-assessment with explicit criteria, and self-explanation exercises produces multiple benefits simultaneously: it provides feedback to the learner being assessed, activates deep processing in the assessor, and builds metacognitive skills in both. This is not a replacement for expert feedback — expert feedback remains more accurate and more targeted — but it is a valuable complement that scales in ways that expert feedback often cannot.
Amara Redesigns Her Tutoring Session
The week after her conversation with Thomas, Amara went back to her notes on everything she'd been learning about learning science and applied it to the structure of her tutoring sessions.
The old structure: Jordan arrives. Amara asks how the week went. Jordan identifies what she's confused about. Amara explains it clearly. Jordan says it makes sense. Session ends.
The new structure:
Amara now starts every session by asking Jordan to close her notes and write down — from memory — the three most important things they covered in the previous session. Jordan's first attempt was painful to watch. She remembered fragments. She got things partially right. She had significant gaps.
Amara didn't step in immediately. She let Jordan struggle with the retrieval for a full two minutes. Then she asked what Jordan thought she'd missed. Then she filled in the gaps — not by re-explaining the full session, but by asking questions that guided Jordan to reconstruct the missing pieces.
This took twenty minutes. The old version — Amara explaining and Jordan nodding — had taken the same amount of time.
Then Amara moved to the new material. But instead of explaining it comprehensively, she used a different structure: she worked through one example completely, narrating her reasoning as she went, then gave Jordan a partially worked example and asked her to complete it. When Jordan got stuck, Amara didn't tell her what to do — she asked what Jordan thought the next step should be, and why.
At the end of the session, instead of asking "does that make sense?", Amara asked Jordan to take two minutes and write down the main point of what they'd just covered, and to identify one thing she was still uncertain about.
Jordan struggled more in the new sessions. She found them harder than the old ones. She occasionally expressed frustration.
Two weeks later, Jordan came to a session and said something she'd never said before: "I actually remembered what we covered last week. Like, when I was in class and the professor mentioned it, I knew what she was talking about."
Amara had made one major shift. She'd stopped trying to put knowledge into Jordan and started designing conditions under which Jordan would construct it herself. The confusion and effort in the sessions weren't signs that the teaching was failing. They were the learning.
Try This Right Now
If you currently teach, tutor, coach, or train anyone, try this in your next session:
Start with two minutes of retrieval from the previous session. No notes, no hints. Just "what do you remember from last time?" Sit with the silence. Don't fill it with prompting until at least sixty seconds have passed.
Notice how different this is from asking "do you remember what we covered last time?" That's a yes/no question that almost always gets an optimistic yes. "Write down what you remember" surfaces what's actually there.
If you don't currently teach anyone, design a three-session mini-curriculum for teaching someone something you know well — a skill, a concept, a process. Before you design it, write down what your learner should be able to do at the end, not just what they should have been exposed to. This single shift in how you frame the objective will change every design decision that follows.
The Progressive Project
If you teach, train, coach, or present, this chapter is a design audit for one session, module, or curriculum you currently run. Choose something real — a class you're actually teaching, a training program you've actually designed, a coaching session you actually conduct.
Work through these questions:
Learning objective audit. What should learners be able to do at the end of this experience? Write it as a verb-plus-context statement: "Learners will be able to [apply/analyze/evaluate] [this concept] [in this type of situation]." If your current objective is "understand X," rewrite it at the level of cognitive demand your assessment actually requires.
Retrieval inventory. Where in your current design does active retrieval happen? If the answer is "only on assessments," identify two places where you can add low-stakes retrieval: one at the beginning of a session, one at the end.
Spacing audit. When does prior material reappear? Map out when concepts introduced in week one are revisited. If the answer is "never," identify one concept that should spiral back in a later session and plan the return.
Cognitive load check. Where is extraneous load highest in your design? What is the one change — to a slide, a handout, a presentation structure, a set of instructions — that would most reduce it?
Scaffolding assessment. Where are your learners novices who would benefit from worked examples and explicit scaffolding? Where have they developed enough competence that scaffolding should be faded? What's your plan for the transition?
Feedback quality review. Look at the last piece of substantial feedback you gave a learner. Was it specific? Was it process-focused? Did it tell the learner what to do differently, or just that something was wrong? What would better feedback have looked like?
Pick the single highest-impact change from this audit and implement it in your next session. Then assess whether learning outcomes change — not by asking learners if the session was clearer or more enjoyable, but by testing what they actually retained.
Amara's instinct to explain things clearly was not wrong — clear explanations have value. What was wrong was treating the clear explanation as the endpoint. The explanation is the beginning: it creates conditions for understanding to be constructed. Retrieval, practice, feedback, and scaffolding are what convert the moment of comprehension into durable knowledge. Designing for learning means designing for what learners do, not just for what instructors say.
A Final Note on the Gap Between Research and Practice
The principles in this chapter — retrieval practice, spacing, cognitive load reduction, worked examples with fading, scaffolding, transfer-oriented design — are not new. The research underlying most of them has been available for twenty to forty years. Some of the foundational work is even older.
And yet most instruction, in schools, universities, and workplaces, still operates as though none of this research existed. Most courses are still designed around content coverage and delivery. Most assessments are still designed to measure final performance rather than to promote learning. Most feedback still focuses on outcomes rather than processes.
The gap between what the science says and what practice does is one of the largest evidence-to-practice gaps in any applied domain.
This chapter is not primarily for professional instructional designers, though it's relevant to them. It's for anyone who teaches anyone anything — formally or informally, in a classroom or a kitchen, in a corporate training room or a one-on-one conversation. The principles apply everywhere that one person is trying to help another person learn.
The science is clear. The techniques are practical. The only thing standing between "instruction that feels good" and "instruction that actually works" is the decision to design for learning rather than for delivery.