#developer#ebbinghaus#forgetting-curve#spaced-repetition#learning-science

The Ebbinghaus Forgetting Curve Explained for Developers (With Code Analogies)

The Ebbinghaus Forgetting Curve is the reason every vocabulary study method fails β€” unless it schedules reviews at the right time. Here's the science explained with software engineering analogies.

✍ Wordrop TeamπŸ“… April 4, 2026⏱ 8 min read

The Ebbinghaus Forgetting Curve Explained for Developers (With Code Analogies)

Every vocabulary study method you've tried has a hidden flaw. You learn a word. You feel like you know it. Then you don't see it for a week, and it's gone.

This isn't a memory limitation β€” it's a scheduling problem. And like most scheduling problems in software, the solution is counterintuitive until you understand the underlying mechanism.

Hermann Ebbinghaus discovered that mechanism in 1885. It's still the most important finding in all of learning science.


The Forgetting Curve: What It Actually Is

In 1885, German psychologist Hermann Ebbinghaus conducted the first rigorous experiments on human memory, testing himself over years by memorizing and testing thousands of nonsense syllables. His finding β€” later called the Ebbinghaus Forgetting Curve β€” has been replicated consistently ever since:

> Memory decays exponentially over time β€” unless the information is rehearsed.

Specifically:

  • Within 20 minutes of learning, you forget ~42% of new material

  • Within 24 hours, forgetting reaches ~67%

  • Within one week, ~77% is lost

  • Within one month, only ~21% remains without reinforcement
  • This isn't a personality flaw. Your brain is doing exactly what it's designed to do: aggressively pruning information it has no reason to believe is important.

    The second finding is the one that changes everything: each time you successfully recall something, the forgetting curve resets β€” but at a shallower slope. The memory becomes progressively more resistant to decay with each successful retrieval.


    The Software Analogy: Memory as a Cache

    Think of your long-term memory as a distributed cache with a garbage collector.

    ``
    // Memory: Simplified mental model
    interface Memory {
    item: string;
    ttl: number; // Time To Live β€” how long before it's pruned
    strength: number; // Resistance to GC
    }

    function onSuccessfulRecall(item: Memory): void {
    // Each recall resets TTL and increases strength
    item.ttl = item.ttl * item.strength;
    item.strength = item.strength * 1.3; // SM-2 ease factor equivalent
    }

    function onFailedRecall(item: Memory): void {
    // Failed recall resets TTL to minimum and reduces strength
    item.ttl = 1; // Back to 1-day interval
    item.strength = Math.max(item.strength * 0.75, 1.3); // Floor at 1.3
    }
    `

    Without reinforcement, your brain's GC runs on schedule and removes the item. Every time you successfully recall before the GC runs, you increase the TTL and reduce the strength β†’ the next review interval gets longer.

    This is the core mechanism of spaced repetition: schedule the review at the last possible moment before the GC would run.


    Why Most Study Methods Fail: The Scheduling Bug

    Here's the bug in most vocabulary study methods, framed as a software problem:

    Problem: Fixed-interval review regardless of memory state

    `javascript
    // What Duolingo and most apps do (simplified)
    function scheduleReview(word, daysSinceLastStudy) {
    return "tomorrow"; // Fixed interval, ignoring memory state
    }
    `

    What should happen: Adaptive review based on recall performance

    `javascript
    // What spaced repetition does
    function scheduleNextReview(word, recallPerformance, currentEaseFactor) {
    if (recallPerformance < 0.6) {
    return 1; // Failed recall: review tomorrow
    }

    const interval = word.lastInterval * currentEaseFactor;
    return Math.round(interval); // Successful recall: exponential growth
    }
    `

    The bug in most study methods is that they treat all vocabulary as equivalent regardless of individual recall performance. They review words you've already mastered too often and words you keep forgetting not often enough.

    Spaced repetition is the fix: personalized scheduling based on per-word recall data.


    The SM-2 Algorithm: The Scheduler That Implements the Fix

    Modern spaced repetition uses SM-2 (SuperMemo 2), developed by Piotr WoΕΊniak in 1987. The algorithm is simple:

    Every item has:

  • Interval (I): Days until next review. Starts at 1.

  • Ease Factor (EF): How aggressively the interval grows. Starts at 2.5.
  • After each review, you rate recall quality (0–5):

    RatingMeaningEffect on IntervalEffect on EF ------------------------------------------------- Again (0)Complete blankReset to 1 dayEF decreases by 0.20 Hard (1)Recalled with difficultyΓ—1.2EF decreases by 0.15 Good (3)Recalled correctlyΓ—EFUnchanged Easy (5)Instant recallΓ—EFEF increases by 0.10

    For a word consistently rated "Good" with default EF = 2.5:
    `
    Day 1 β†’ Day 2 β†’ Day 7 β†’ Day 18 β†’ Day 45 β†’ Day 112 β†’ Day 278...
    `

    For a word consistently failed (EF drops to floor of 1.3):
    `
    Day 1 β†’ Day 1 β†’ Day 1 β†’ Day 1... (forever daily until you recall it)
    `

    This is the key insight: SM-2 doesn't let you "graduate" a word you haven't actually learned. It's honest about recall performance in a way no fixed schedule can be.


    Why Micro-Sessions Beat Blocked Study: The Desirable Difficulty Effect

    Here's a counterintuitive finding from cognitive science:

    Harder retrieval conditions produce stronger memories β€” even when accuracy is lower.

    This is called the desirable difficulty effect (Bjork, 1994). When retrieval is easy (you just reviewed this yesterday), the recall act itself is cheap, and the memory barely strengthens. When retrieval is effortful (you haven't seen this word in 3 weeks and have to really search), the recall act is expensive, and the memory strengthens significantly.

    For developers, this explains why:

  • Reviewing words right after you learn them feels efficient but isn't. The recall is too easy to build durable memory.
  • Forgetting partially before reviewing is actually beneficial. It makes the retrieval effortful, which strengthens the memory more.
  • Short sessions distributed across the day outperform one long session. More distinct retrieval events = more memory strengthening = faster learning.
  • `python
    # Simplified desirable difficulty model
    def memory_strengthening(recall_effort: float) -> float:
    """
    recall_effort: 0.0 (effortless) to 1.0 (maximum effort)
    returns: memory_strength_gained
    """
    # Non-linear: harder recalls produce disproportionately stronger memories
    return 0.3 + (0.7 recall_effort * 0.5)

    # Easy recall (yesterday's word)
    strengthening = memory_strengthening(0.1) # Returns ~0.52

    # Effortful recall (3-week-old word)
    strengthening = memory_strengthening(0.8) # Returns ~0.93
    ``

    The practical implication: don't try to make vocabulary review easy. The struggle is the mechanism.


    What This Means for Your Study Schedule

    If you've been studying vocabulary by:

  • Reading word lists before sleep

  • Reviewing flashcards until you feel confident

  • Doing Duolingo lessons daily on a fixed schedule
  • You've been scheduling reviews at the wrong time β€” too early (while the recall is still easy) rather than just before the forgetting curve takes the memory below threshold.

    The optimal schedule:

    DayExpected RetentionAction --------------------------------- Day 0100% (just learned)Learn the word Day 1~58%First review β€” this is the critical window Day 3–4~40%Second review β€” effort required, strengthening maximized Day 10–14~35%Third review Day 30–35~30%Fourth review Day 90+~25%Now in long-term stable memory

    Notice: each review interval gets longer as the memory strengthens. This is the compounding effect of spaced repetition β€” after 4–5 reviews, the word requires only occasional reinforcement to stay permanently accessible.


    The Engineering Principle: Run the Review Before the GC Fires

    The fundamental insight from the Ebbinghaus research, translated for developers:

    > Review just before the forgetting threshold β€” not immediately after learning, not long after forgetting, but precisely at the point of maximum desirable difficulty.

    Spaced repetition systems automate this scheduling. They track each word's individual decay curve and schedule the review at exactly the right time. This is why a well-designed SRS can maintain 85–95% retention with only 10–20 minutes of daily review, while ad-hoc studying at far higher time investments produces 20–40% retention.

    It's not about effort. It's about scheduling.


    Frequently Asked Questions

    Is the Ebbinghaus Forgetting Curve still considered valid in modern cognitive science?
    Yes. The exponential decay pattern has been replicated in hundreds of studies across different materials, populations, and time scales. The specific percentages vary based on material difficulty and individual differences, but the exponential decay shape and the stabilizing effect of spaced retrieval are among the most robust findings in psychology.

    Why doesn't the brain just keep everything?
    Energy conservation. The brain consumes approximately 20% of the body's energy despite being only 2% of body weight. Retaining all encoded information would be metabolically unsustainable. The GC β€” forgetting β€” is a feature, not a bug.

    If I already know what the forgetting curve is, why am I still forgetting words?
    Because knowing about the forgetting curve doesn't automatically generate a review schedule. The curve tells you what happens. Spaced repetition tells you when to review. Without automated scheduling, you're still relying on willpower to review at the right intervals β€” which consistently fails.

    How many reviews does a word need before it's in long-term memory?
    With SM-2, approximately 5–7 successful reviews over 3–6 months. After reaching a review interval of 60+ days, words reliably persist in memory for years with minimal reinforcement.


    Your brain has a garbage collector. The forgetting curve describes its schedule. Spaced repetition is the caching strategy that runs reviews before the GC fires β€” so your words stay in memory instead of being pruned.

    The mechanism is as reliable as the algorithm it's based on.

    See spaced repetition in practice with Wordrop β†’

    W

    Wordrop Team

    Building tools to make language learning effortless and evidence-based.

    [ APPLY WHAT YOU JUST LEARNED ]

    START BUILDING VOCABULARY TODAY

    DOWNLOAD WORDROP FREE β†’