Why Do You Forget 77% of What You Read? The Science Explained
Ebbinghaus forgetting curve, testing effect, spaced repetition, dual coding: what memory science says about reading retention and why passive reading fails.
Elliott Tong
March 14, 2026
15 min read
Why Do You Forget 77% of What You Read? The Science Explained
You forget roughly 77% of what you read within a week. Not an estimate. Ebbinghaus documented this in 1885 and it's been replicated many times since. Passive reading without review produces this outcome reliably. The science also shows exactly what changes it, and why most common strategies (highlighting, re-reading, summarising) don't make a dent.
Three years into university, I sat down and tried to recall something specific from the 40-odd articles I'd read that semester.
Not a vague recollection. Something concrete. An argument, a number, a name I could actually use.
Most of it was gone. A few scraps, maybe. I'd spent hundreds of hours reading and had almost nothing to show for it. I thought the problem was me. My attention, my discipline, my memory.
It wasn't. The problem was something much more boring: physics. Memory follows predictable laws, and passive reading violates almost all of them.
This piece is the science behind what goes wrong. If you want the practical strategies, those are in a companion guide to improving reading retention. What follows is the mechanism: why your brain discards what you read, what the research on memory actually shows, and why most intuitive study habits don't work.
What Does the Ebbinghaus Forgetting Curve Actually Show?
Ebbinghaus's forgetting curve shows that memory decays exponentially after learning stops. Without any review, you'll lose about 56% of new material within an hour, 66% within a day, and roughly 77% within a week.
Hermann Ebbinghaus spent years in the late 1800s memorising nonsense syllables (sequences with no prior meaning, like "XOQ") and then testing his own recall at various intervals. What he found was that forgetting isn't random. It follows a reliable exponential pattern, dropping sharply in the first hours and then leveling off.
The curve looks brutal in a table:
| Time after learning | Retention (passive reading) |
|---|---|
| 20 minutes | ~58% |
| 1 hour | ~44% |
| 1 day | ~34% |
| 1 week | ~23% |
| 1 month | ~21% |
That 23% after a week is what researchers call the "long-term storage" component: material that encoded deeply enough to survive initial decay. For passive reading of ordinary articles, most of what you read never reaches that depth.
The shape of the curve matters as much as the numbers. Forgetting is front-loaded. Most of what you lose, you lose in the first hour. This is why reading something right before you need to use it feels more effective than reading it a week ago. You're catching the material before the sharpest part of the drop.
Two things flatten the curve: repetition and active processing. Both work because they force re-encoding, which is different from re-reading. Each time you successfully retrieve a memory, you extend its half-life. The material that has survived three retrievals is far more stable than material you read for the first time this morning, even if both feel familiar in the moment.
The curve also explains why massed reading (reading the same article twice in a row, "massed practice") produces so little retention gain. You're returning before the decay has happened, so there's no meaningful retrieval happening. Just recognition of something that's still active in short-term memory.
Why Does Testing Yourself Outperform Re-Reading?
Retrieval practice produces better long-term retention than restudying because the act of retrieval itself strengthens memory. Roediger and Karpicke (2006) showed that students who practiced retrieval retained 50% more after one week than students who restudied the same material.
This is called the testing effect, and it's one of the most replicated findings in cognitive psychology.
The intuition most people have is that testing measures learning and studying causes it. That's backwards. Testing causes learning. The act of pulling information out of memory is what strengthens the memory trace. Re-reading just adds another encoding episode to a memory that was never properly retrieved.
The Roediger and Karpicke study was clean. Three conditions:
- Read the passage, study it repeatedly (SSSS)
- Read it once, test yourself once (STTT)
- Read it once and do four retrieval tests (STTT)
After five minutes, the restudying group scored highest. After a week, the retrieval practice group had retained roughly 50% more. The effect flipped because retrieval practice built durable memory. Re-reading built temporary familiarity.
This matters because familiarity feels like knowledge. When you re-read something you've already seen, it feels fluent and accessible. Your brain interprets that fluency as retention. But fluency in the moment says almost nothing about recall a week later. This is what researchers call the recognition trap: you can recognise information without being able to retrieve it independently.
The meta-evidence reinforces the individual study. Dunlosky et al.'s 2013 review of ten common study techniques rated practice testing the highest utility, "strongly recommended as a general learning strategy." Highlighting, underlining, re-reading, summarization: all rated low or moderate utility. Only practice testing and distributed practice reached the top tier.
Why Most Study Habits Aren't Working
The techniques students and professionals rely on are systematically misaligned with how memory works:
| Technique | Dunlosky et al. Rating | Why it fails |
|---|---|---|
| Highlighting | Low utility | Passive; creates familiarity, not recall |
| Re-reading | Low utility | Recognition, not retrieval |
| Summarization | Low utility | Requires skill to do well; rarely tested |
| Keyword mnemonics | Low utility | Narrows learning to surface features |
| Imagery for text | Low utility | Inconsistent encoding benefit |
| Elaborative interrogation | Moderate utility | Works, but skill-dependent |
| Self-explanation | Moderate utility | Works, but time-intensive |
| Interleaved practice | Moderate utility | Strong evidence, complicated to implement |
| Practice testing | High utility | Forces retrieval; strengthens memory |
| Distributed practice | High utility | Spaces encoding; fights forgetting curve |
The two that work are the two that create genuine retrieval demands on memory. Everything else feels productive but isn't.
What Is Spaced Repetition and Why Does It Work Better Than Cramming?
Spaced repetition schedules material for review at increasing intervals, timed to catch each memory just before it would decay. A 2006 meta-analysis by Cepeda et al. covering 317 experiments confirmed that distributed practice consistently outperforms massed practice, with the gap widening over longer delays.
The logic is counterintuitive. Reviewing material right after you learned it produces almost no memory benefit. The memory trace is still active, so retrieval requires no effort. But reviewing after a gap, when the memory has partially decayed, forces genuine reconstruction. That reconstruction is the mechanism that builds durable storage.
The ideal interval isn't "review as often as possible." It's "review as infrequently as possible while still being able to retrieve the material successfully." This produces the maximum memory gain per unit of time spent.
The History of Spaced Repetition Systems
The practical history of spaced repetition as a tool runs like this:
1885: Ebbinghaus documents the forgetting curve and proposes that spaced review can flatten it.
1932: C.A. Mace suggests "distributed practice" as a principle in psychology.
1972: Sebastian Leitner builds the first practical spaced repetition system using physical flashcard boxes. Cards you know get moved to boxes reviewed less frequently. Cards you don't know stay in the box reviewed daily. Simple, manual, effective.
1987: Piotr Wozniak designs the SM-2 algorithm (now the basis of Anki's default scheduling). Cards are assigned an "ease factor" starting at 2.5. Each review multiplies the interval by the ease factor. Hard ratings decrease ease permanently. This works well for most cards but creates "ease hell," where difficult cards get stuck at minimum ease and reviewed every 3-5 days forever.
2023: The FSRS algorithm (Free Spaced Repetition Scheduler, designed by Jarrett Ye and trained on 700 million reviews from 20,000+ users) replaces ease factors with a full memory model: stability (how long before the memory decays), retrievability (current probability of recall), and difficulty (inherent complexity). FSRS achieves 20-30% better efficiency than SM-2 and eliminates ease hell. It became Anki's default scheduler in November 2023.
The progression isn't just algorithmic refinement. Each generation fixed a specific failure mode in the previous one. SM-2 beat Leitner boxes because it personalized intervals. FSRS beats SM-2 because it models memory as a continuous state rather than a card-level ease factor.
The 85% Rule
One finding from this research has a surprisingly clean empirical basis. A 2019 study in Nature Communications by Wilson and Shenhav derived mathematically that the optimal error rate for learning is approximately 15%. You should succeed roughly 85% of the time in any retrieval session.
Below 70% accuracy, you're in frustration territory. The failure rate is too high for productive encoding, and the emotional cost starts outweighing the memory benefit.
Above 95% accuracy, the material is too easy. You're not creating genuine retrieval demand, so memory isn't meaningfully strengthened.
This 85% rule aligns with Vygotsky's "zone of proximal development," the productive difficulty band where challenge exists but success is achievable. The specific mechanism differs (Vygotsky was describing instructional scaffolding, not spaced repetition), but the target zone is similar.
For practical reading, this means you want to review material when recall feels effortful but possible. Not so early that retrieval is trivial, not so late that you've forgotten everything.
Why Do Audio and Visual Together Improve Retention?
Dual coding theory (originally Clark and Paivio) holds that combining verbal and visual information creates two distinct memory traces: a verbal one and an imagery-based one. Both can serve as retrieval cues independently. More retrieval pathways means more chances to access the memory later.
Richard Mayer extended this into the modality principle: audio combined with synchronized visual representation improves learning more than visual alone, even when the total information presented is identical. Across 17 experiments, the effect size was d = 1.02. Large by any measure in psychology.
The mechanism matters here. It's not that audio is inherently better than visual, or that redundancy helps. What matters is that each channel activates a different cognitive processing system. The visual cortex encodes the written word as a visual symbol. The auditory cortex encodes its phonological form. When both are active simultaneously, the memory is encoded with two distinct sets of retrieval cues rather than one.
This is why word-by-word synchronized text highlighting plus audio is different from just listening. Pure listening encodes via the auditory channel. Reading encodes via the visual-verbal channel. Both together, in sync, engages both systems at once.
The practical implication: if you want to retain something, passive audio (podcasts, audiobooks without visual support) is better than nothing but leaves encoding incomplete. Audio synchronized with visual text gets closer to optimal dual-channel encoding.
Human speech is also the evolutionarily older channel. Written language is approximately 5,000 years old. Spoken language is at least 100,000 years old, possibly much older. Your brain spent 95% of its evolutionary history processing information through sound. The auditory system is not a secondary channel for information. For most of human history, it was the primary one.
What Is Desirable Difficulty and Why Does Easy Reading Mean Poor Retention?
Desirable difficulty is a term coined by Robert Bjork at UCLA to describe learning conditions that slow apparent progress but improve actual long-term retention. The paradox at the core: the conditions that make learning feel harder in the moment are often the conditions that make it stick.
Bjork's explanation is that memory has two components: storage strength (how deeply encoded a memory is) and retrieval strength (how easily accessible it is right now). These are independent. High retrieval strength in the moment (you just read it, it feels very accessible) can coexist with low storage strength (without reinforcement, it will decay quickly).
Desirable difficulties work by deliberately reducing retrieval strength during study: spacing out reviews until partial forgetting has occurred, presenting material in varied contexts, requiring the learner to generate information rather than recognise it. Each of these reductions in retrieval strength, when successfully overcome, produces a gain in storage strength that outlasts the temporary difficulty.
The critical qualifier: not all difficulties are desirable. A difficulty is only desirable if the learner can successfully overcome it. If an item is so hard that retrieval consistently fails, there's no memory benefit. Just frustration. The 85% rule applies here too: difficulty should produce struggle, not failure.
For passive reading, the problem runs the other direction. Reading feels fluent. Words go in smoothly. There's no friction. And that smoothness, that ease of processing, signals to your brain that the material doesn't need to be encoded deeply. Easy processing leads to shallow encoding, which leads to rapid forgetting.
This is why re-reading something right after first reading it produces so little benefit. The material is maximally accessible (retrieval strength is high), so the brain treats it as already known and doesn't allocate more encoding resources.
What Is the Generation Effect and Why Does Producing Information Beat Consuming It?
The generation effect is the finding that producing information from memory (writing an explanation, completing a sentence, articulating a concept in your own words) produces better retention than passively reading the same information.
A meta-analysis of 86 studies found a d = 0.40 effect for the generation effect, with hit rates of 87% for generated items versus 65% for read items. The neural evidence adds precision: generation activates a broader network than reading, including prefrontal cortex and inferior temporal gyrus. More encoding resources get recruited at the moment of production.
The mechanism is related to desirable difficulty. Generating information requires reconstructing it, which is the same cognitive operation as retrieval. Each time you produce a memory rather than receive it, you strengthen the retrieval pathways. Each time you consume information passively, you strengthen the recognition trace but not the recall trace.
This distinction between recognition and recall is the crux of why passive reading fails at building usable knowledge. You can recognise information you've seen before; it feels familiar. But recognition and recall are different cognitive operations. Familiarity doesn't predict recall. You can recognise an actor's face without being able to recall their name. You can recognise an article you read without being able to retrieve its argument.
For knowledge that actually needs to be retrieved (cited in a conversation, applied to a decision, connected to new information you're reading) recall is what matters, not recognition.
Why Passive Reading Is Designed for Consumption, Not Learning
This is the uncomfortable truth that the research points toward: reading, as most people practice it, is optimised for consumption rather than retention. You move through text. You register the content. You feel like you've absorbed it. And then it leaves.
The tools built around reading have mostly reinforced this. Read-it-later apps optimise for saving. Note-taking apps optimise for capture. Highlighting tools optimise for marking. None of these ask whether the person came away knowing more: whether they could retrieve the material a week later, apply it in a different context, connect it to something they'd read before.
The bottleneck was never access to content. More articles, faster reading, easier saving. None of that touches the actual problem. The bottleneck is encoding. Getting information from your working memory into long-term storage in a form you can actually retrieve.
What Does All This Mean in Practice?
The science converges on a small number of mechanisms that actually move the needle. Worth naming them explicitly:
Retrieval practice outperforms re-study. Testing yourself on material, even imperfectly, produces more durable memory than reading it again. The specific format matters less than the retrieval demand: free recall, cued recall, filling in blanks, explaining from memory. All of these work better than re-reading.
Spacing beats massing. Reviewing material once a day for three days produces more retention than reviewing it three times in a single day. The forgetting that happens between sessions is not a problem. It's the mechanism. Retrieving from partial forgetting is what builds storage strength.
Generation beats consumption. Producing information in your own words, in a new context, encodes it more deeply than receiving it. This is why writing notes in your own words works better than copying, and why explaining a concept to someone else works better than re-reading it.
Dual coding strengthens both retrieval paths. Combining audio and synchronized visual text engages two separate memory systems. More encoding pathways mean more retrieval pathways. The memory becomes accessible through both channels instead of one.
Difficulty signals depth. If reading feels effortless, you're probably not encoding deeply. The friction of retrieval, the work of reconstructing a memory rather than recognising it, is the active ingredient in durable learning.
None of these are complicated. What makes them hard is that they contradict what feels productive. Re-reading feels like learning. Highlighting feels like learning. Moving through text quickly feels like progress. The science is clear that most of these feelings are wrong.
How Alexandria Applies This Research
The reading tools most people use were designed for consumption. They optimise for saving articles, reading faster, or clearing the backlog. None of them were built around the encoding mechanisms that memory science says actually matter.
Alexandria is built on the research above. FlowRead's word-by-word sync highlighting with audio is dual coding in practice: the visual and auditory channels active simultaneously, in sync. The knowledge extraction system is built around the generation effect: knowledge blocks are structured summaries, not passive highlights, and they're the inputs to retrieval practice.
Spaced repetition is built into the system. The same knowledge blocks Alexandria extracts are the ones it schedules for review at increasing intervals, using FSRS, not the older SM-2 algorithm. No separate flashcard system to maintain. No manual entry. The review happens at the right time, at the level of the material you actually read.
The goal isn't to read more. The goal is to retain what you read. Those are different problems, and they need different tools.
If you want the practical guide to improving retention, start with how to actually remember what you read. If you want to try the system that applies this science, Alexandria is free to start.
See also:
- Why You Forget Every Article You Read (and What to Do About It) explores the Google Effect and why your brain treats the internet as an external hard drive.
- Why Your Brain Gives Up on Articles (Before You Finish Them) covers cognitive load and working memory, the bottleneck that makes long reads feel impossible.
- Is AI Making You Dumber? What the Research Actually Shows examines what happens when tools do your thinking for you, and why passive consumption is accelerating.
Frequently Asked Questions
What percentage of what you read do you actually remember?
Research based on the Ebbinghaus forgetting curve shows passive reading loses roughly 56% of new material within one hour, 66% within a day, and approximately 77% within a week. Without any review or active processing, almost nothing remains accessible in long-term memory after a month.
What is the Ebbinghaus forgetting curve?
The Ebbinghaus forgetting curve describes how memory decays over time without rehearsal. Hermann Ebbinghaus discovered in 1885 that forgetting follows a predictable exponential pattern: roughly 56% of new material is lost within an hour, 66% within a day, and 77% within a week. The curve flattens after repeated spaced review.
What is the testing effect in learning?
The testing effect (also called the retrieval practice effect) is the finding that testing yourself on material produces much better long-term retention than re-reading it. Roediger and Karpicke (2006) showed students who practiced retrieval retained 50% more after a week than students who restudied. Testing doesn't just measure learning. It causes it.
Does highlighting help you remember what you read?
No. Dunlosky et al.'s 2013 meta-analysis of 10 common study techniques rated highlighting as "low utility," performing no better than plain re-reading. Highlighting feels productive because it requires attention, but it's passive. Effective retention requires actively retrieving information from memory, not marking it while you read.
What is spaced repetition and does it actually work?
Spaced repetition is a review system that schedules material at increasing intervals, reviewing just before you'd forget it. It works exceptionally well. Dunlosky et al. (2013) rated it "high utility," one of only two techniques to receive that rating across 10 studied. Modern algorithms like FSRS achieve 20-30% more efficient retention than older systems.
What is dual coding theory and how does it improve reading retention?
Dual coding theory (Clark and Paivio) holds that combining verbal and visual information creates stronger memory traces than either alone. Mayer's modality principle extends this: audio plus synchronized visual (like word-by-word highlighted text) improves learning with an effect size of d = 1.02 across 17 experiments. Two channels mean two retrieval pathways.
What is desirable difficulty in learning?
Desirable difficulty, coined by Robert Bjork at UCLA, describes learning conditions that slow initial performance but improve long-term retention. Things that feel harder to learn often stick better. Spacing reviews until you've partially forgotten material, generating information rather than re-reading it: these are desirable difficulties that build durable storage strength.
What is the generation effect in memory?
The generation effect is the finding that producing information from memory (filling in a blank, writing an explanation, completing a sentence) produces stronger recall than passively reading the same information. A meta-analysis of 86 studies found a d = 0.40 effect, with hit rates of 87% for generated items versus 65% for read items.
What is the difference between storage strength and retrieval strength in memory?
Robert Bjork distinguishes storage strength (how deeply a memory is encoded in long-term memory) from retrieval strength (how easily accessible it is right now). These are independent. Something can feel very accessible right after reading but be poorly stored. Retrieval practice builds storage strength. Re-reading mostly reinforces retrieval strength temporarily.
What is the FSRS algorithm and how does it improve on SM-2?
FSRS (Free Spaced Repetition Scheduler) was trained on 700 million reviews from 20,000+ users and models memory as three variables: stability, retrievability, and difficulty. Unlike SM-2, which uses a fixed ease factor, FSRS recalculates the full memory model after each review. It achieves 20-30% better efficiency than SM-2 and became Anki's default scheduler in November 2023.