← All articles
Learning Science

What developmental science says about attention in early learners

Attention spans grow on a known curve. Most kids' apps are designed against it.

Tim de Vallée8 minTBD

A 4-year-old can pay close attention to a worm in the driveway for eleven minutes. The same child, asked to sit through a tap-tap-tap reading exercise on a tablet, taps out in ninety seconds and starts pulling at the carpet.

Both numbers are real. Neither tells you the child has "a short attention span." They tell you that attention is not a single thing, and that the design of what's in front of the child matters as much as the child's age.

Most kids' apps are built as if attention were a battery — fill it with stimulus, drain it for engagement, recharge with a streak notification. Developmental science describes something very different: a layered, slowly maturing set of cognitive skills that respond to specific kinds of input and break under others. If you understand the layers, you stop blaming the kid for bailing on the app, and you start asking better questions about the app.

Attention is not one skill

Researchers generally describe three functional types of attention that matter for early learning. The terminology varies by field — Michael Posner's work on attention networks (orienting, alerting, executive) is one common framework — but for parents the three useful distinctions are:

Sustained attention is the ability to keep focus on one task over time without external prompting. This is the worm-in-the-driveway skill. It's also the skill required to follow a story, work through a puzzle, or stick with a hard word.

Selective attention is the ability to focus on the relevant thing and ignore competing input — the buzzing dishwasher, the sibling, the bright animated character bouncing in the corner of the screen.

Divided attention is the ability to do two cognitive tasks at once — listen and type, read and remember, watch and respond. In young children this is essentially unavailable; what looks like divided attention is usually rapid switching, which carries a measurable cost every time the child switches back.

These three develop on different timelines and with different ceilings. Conflating them is where most kids' app design goes wrong.

What the curve actually looks like

The often-cited rule of thumb — "two to five minutes of attention per year of age" — comes from clinical observation, not a single landmark study, and it understates what kids can do under the right conditions. Holly Ruff and Mary Klevjord Rothbart's developmental work (summarized in their 1996 book Attention in Early Development, Oxford University Press) describes sustained attention emerging between ages 3 and 6, with significant individual variation and a strong dependence on what the child is attending to.

A few research-backed reference points:

  • Ages 3-4: Sustained attention to a chosen activity can stretch past ten minutes when the child has agency. Attention to an adult-imposed task drops sharply — often under three minutes. Selective attention is weak; background stimuli pull focus easily.
  • Ages 5-6: Sustained attention to structured tasks reliably reaches 10-15 minutes. Selective attention is still developing — the child can ignore one distractor but not several.
  • Ages 7-8: Sustained attention can hold 20-30 minutes on the right material. Selective attention matures rapidly during this window. This is part of why formal schooling starts here in most countries.
  • Ages 9-10: Approaches adult-like patterns for sustained and selective attention. Divided attention is still poor and stays poor through adolescence — Russell Barkley's extensive work on executive function (paraphrased — see Executive Functions: What They Are, How They Work, and Why They Evolved, Guilford Press, 2012) places full executive maturation well past the early-learner range. [VERIFY: page reference for the specific developmental window]

The U.S. Centers for Disease Control and Prevention publishes age-by-age developmental milestones that align with this picture and are worth bookmarking: see the CDC's "Learn the Signs. Act Early." milestones.

How digital interfaces help or hurt each type

Digital design choices interact with each attention type differently. The same feature can help one and hurt another.

Sustained attention

Helped by: a single clear goal at a time, response latency under one second, content that adapts to the child's pace rather than the clock's, and the ability for the child to drive depth ("tell me more about volcanoes").

Hurt by: long response delays (the child fills the gap with something else and doesn't come back), constant difficulty jumps that break flow, and timers visible to the child.

If you want the full latency argument, see why a ten-second delay kills your child's learning — the working-memory math is sharper than most parents expect.

Selective attention

Helped by: visually quiet interfaces, audio that comes from one source at a time, and a clear figure-ground relationship between the task and everything else on screen.

Hurt by: animated mascots in the periphery, autoplay sidebars, achievement popups during a task, and "engagement features" that compete with the learning task for the child's focus. The U.S. National Institute of Child Health and Human Development (NICHD), part of the National Institutes of Health, summarizes media-and-attention research that points in this direction at nichd.nih.gov.

Divided attention

Mostly: leave it alone. Young children can't divide attention well, so interfaces that demand it (tap this while listening to that while watching the third thing) are just training rapid switching. Switching has a cost that researchers in cognitive control have measured repeatedly in adults — and the effect is larger in children. The American Academy of Pediatrics' guidance on media use (aap.org/en/patient-care/media-and-children) leans heavily on this point: quality of attention matters more than minutes on a clock.

What evidence-based design looks like

If you took the science seriously and designed an app for a 5-year-old from scratch, you would build something that looks almost nothing like the brightly animated, streak-driven, badge-collecting product most kids' apps still ship. You'd build something that looks closer to a calm conversation with a patient adult.

Evidence-based design, distilled:

  1. One task on screen at a time. No competing animations, no mascots waving from the corner.
  2. Response latency under one second for any interaction the child initiates. Anything longer and you've lost their working memory.
  3. Difficulty driven by the child's actual responses, not a fixed difficulty curve. A child who is bored needs harder material; a child who is struggling needs to back up, not "try again."
  4. Sessions that end when the child is done, not when a streak demands one more round. Engagement maximization and learning are different goals and frequently in conflict.
  5. Speech and listening before tapping. Kids' receptive and expressive language outpaces their fine motor and reading skills by years. Voice interfaces meet the child where they are.
  6. Quiet between turns. Silence is not dead air — it's the space the child uses to think.
  7. No notifications. A learning app that pings the child (or the parent on the child's behalf) is optimizing for return visits, not learning.

This is the same design philosophy underneath the screen-quality argument — covered in more depth in screen time is the wrong question, screen quality is the right one.

A parent's checklist for any kids' app

Use this to evaluate any learning app on the market — Lumikids included, and any of its competitors. Score generously: an app needs to clear most of these, not all.

  • Latency check. From the moment your child finishes speaking or tapping, how many seconds pass before the app responds? Time it on your phone. Under 1 second is excellent. 1-3 seconds is workable. Over 5 seconds is breaking your child's working memory.
  • Visual quiet. Open the app, look at the screen during a task. Count the moving things. More than two animated elements means the design is competing with itself for the child's attention.
  • One-source audio. Is the app ever playing music underneath a voice prompt? If yes, it's making the selective-attention problem harder, not easier.
  • Adaptive on what. Ask (or read the docs): does the app change difficulty based on the child's actual answers and pauses, or on a fixed level system? Fixed systems are not adaptive in any meaningful sense.
  • The streak test. Does the app try to bring your child back with notifications, streaks, or "don't break the chain" messaging? If yes, it's optimizing for retention metrics, not for your child.
  • Exit behavior. Can your child stop mid-session without losing progress or being guilted? Healthy learning apps let kids leave.
  • Observability. Can you see what your child actually did — the specific questions, the actual responses, where they got stuck? Or only a summary score? Summaries are for the company's dashboard; specifics are for you.
  • Data minimum. What is collected on your child, where does it live, who has access? If the answer is "we use industry-standard analytics," that's not an answer.
  • Off-screen handoff. Does the app encourage your child to do something away from the screen — read a real book, look up at the sky, ask you a question? Or does it try to keep them in-app?
  • The boredom signal. Can your child get bored and the app notices, or does it double down with animations and rewards? An app that can't tell the difference between bored and engaged isn't paying attention to your child.

If you want a deeper version of this framework specifically for AI tutors — questions about what model is running, where data lives, how adaptation actually works — see a parent's framework for evaluating any AI tutor.

The honest takeaway

Attention in early learners is not short. It's specialized. A 5-year-old can sustain attention for a long stretch on the right material delivered in the right way; the same child will check out of the wrong material in ninety seconds, and the app will blame the kid.

The science is consistent enough that parents can use it as a filter. If an app's design fights against how attention actually develops — long delays, busy screens, engagement traps, fixed difficulty — no amount of marketing language about "learning science" or "personalization" makes up for it.

Lumikids is the app I built for my son Remi after watching him lose attention to the response delays in a major reading platform — sub-second responses, quiet screen, voice-first, no streaks, sessions that end when he's done. If you want to see the design choices in action, the beta is open.

Image brief

  • Hero image: A 5-year-old at a kitchen table looking up from a tablet mid-sentence, natural morning light, parent's hand visible at the edge of frame, no app branding on the screen.
  • Inline image 1: A simple line chart showing sustained attention duration by age (3-10), with three curves layered for "free play / chosen task," "structured adult-led task," and "passive media." Place after "What the curve actually looks like."
  • Inline image 2: A side-by-side wireframe diagram — "busy app interface" with mascots, badges, timer, and progress bar vs. "quiet app interface" with a single task and a microphone icon. Place after "What evidence-based design looks like."

Internal link suggestions

  • "Why a ten-second delay kills your child's learning" — anchor text: why a ten-second delay kills your child's learning
  • "Screen time is the wrong question. Screen quality is the right one." — anchor text: screen time is the wrong question, screen quality is the right one
  • "A parent's framework for evaluating any AI tutor" — anchor text: a parent's framework for evaluating any AI tutor

Editor's note

Tim — three things to verify before publish. (1) The Ruff & Rothbart reference: I cited their 1996 Oxford book Attention in Early Development, which is real and widely cited, but please confirm the specific age ranges I attributed to them match what's in the book vs. what's general consensus across the field. (2) The Barkley reference is paraphrased and marked [VERIFY] — if you want a tighter citation we should pull the actual page on early-learner executive function rather than the general 2012 book. (3) The "ninety seconds" and "eleven minutes" in the opening are illustrative composites, not from a named study — if you've got an actual Remi anecdote that lands the same point, swap it in. Also worth confirming: is the beta page actually at /beta or somewhere else?

One more thing —

Lumi is in open beta and free for the first 100 families. If reading time at your house ever feels harder than it should, we built this for you.