The Why and How of Research and the Science of Reading

Blog Banner
14 March, 2026

science of reading

reading research

Blast from the Past: This blog first posted on November 3, 2018, and was reposted in revised form on March 14, 2026. The original title was, “The Whys and Hows of Research and the Teaching of Reading.” As you can see, I’ve retitled it. The term “science of reading” is both widely used and widely misunderstood today, so I have provided considerable revision here – though the overall points are the same. If you want to see the original, it is linked at the bottom (along with the more than 50 comments it elicited).

I talk a lot about research in this space.

I argue for research-based instruction and policy.

I point out a dearth of empirical evidence behind some instructional schemes, and champion others that have been validated or verified to my satisfaction.

Some readers are happy to find out what is “known,” and others see me as a killjoy because the research findings don’t match well with what they claim to “know.”

Members of this latter group are often horrified by my conclusions. They often are certain that I’m wrong because they read a book for teachers that had lots of impressive citations contradicting my claims. Sometimes these days, it doesn’t even take that. A few nasty assertions on X can do the trick.

I’m often challenged with statements like, “I thought that was part of the science of reading.”  

What is clear from these exchanges is that many educators, legislators, journalists, and parents don’t know what research is, why we should rely on it, or how to interpret research findings. That, perhaps, is not surprising but it does make these very important people susceptible to groundless claims and false promises.

Research is used to try to answer a question, solve a problem, or figure something out. It requires the systematic and formal collection and analysis of empirical data – using methods able to provide the kind of answer being sought. That means not all research can be used to provide a valid answer to all questions. Even with that, research can never prove something with 100 percent certainty. It is useful for reducing uncertainty, for increasing the likelihood that we will get something right – it cannot guarantee it.

“Systematic and formal” means that there are rules or conventions for how data in a research study should be handled; the rigor of these methods is what makes the data trustworthy and allows research to reduce uncertainty. If a researchers want to compare the effectiveness of two instructional approaches, they must make sure the groups to be taught are equivalent at the beginning. Without that we can’t be sure which method did best. Likewise, we are more likely to trust a survey that defines its terms, or an anthropological study that immerses the observer in the environment for a long period of time.

Research reports do much more than just report the results or outcomes of a study, but they explain—preferably in detail—how those results were obtained. Most people may find such detailed description of methods to be mind-numbingly boring, but it is that detail that allows us to determine what a study really means and how much weight to place on it.

Here are some simple guidelines to remember.

1. Just because something is written, doesn’t make it research.

Many practitioners think that if an idea is in a book or magazine (or even just comments on social media) that it is research. Some even think this blog is research. It is not, and neither is the typical Reading Teacher article or Heinemann book.

That’s not a comment on their quality or value, but a recognition of what such writing provides. In some cases, as with my blog, there is a serious effort to summarize research findings accurately, and to distinguish opinions (what someone believes) from research findings (what we know).  

Many publications for teachers are no more than compendia of opinions or personal experiences, which is fine. However, these suffer all the limits of that kind of thing. Just because someone likes what they’re doing (e.g., teaching, investing, cooking) and then writes about how well they’ve done it… doesn’t necessarily mean it’s great. That’s why 82% of people believe they’re in the top 30% of drivers; something that obviously can’t be true. (Often teachers will tell me they are getting great results doing something that contradicts my research summaries. I always request supporting test data that demonstrate the claim. Not surprisingly, they almost never write back and when they do it is not to offer any proof.)

As human beings we all fall prey to overconfidence, selective memory, and just a plain lack of systematicity in how we evaluate what we are doing and its impact.

Often teachers tell me that kids now love reading because of the teaching they have provided them. I ask how they know. What evidence do they have? Usually the answer is something like, “A parent told me that their child now likes to read.” Of course, that doesn’t reveal how the other 25 kids are doing, or whether that parent is a good observer of such things, or even what the motivation may have been for the praise.

Even when you’re correct about things improving, it’s impossible—from personal experience alone—to know the source of the success. It could be the teaching method, or maybe just the force of your personality. If another teacher adopted your methods, things might not be so magical.

And, then there is opportunity cost. We all struggle with this one. No matter how good an outcome, I can’t possibly know how well things might have gone had I done it differently. The roads not traveled may have gotten me someplace less positive—but not necessarily. You simply can’t know.

That’s where research comes in… it allows us to avoid overconfidence, selective memory, lack of systematicity, lack of reliable evidence, incorrect causal attribution, and the narrowness of individual experience.

When I first published this in 2018, my concern was that many teachers were placing too much confidence in first-hand accounts of other teachers and were too likely to buy into junk from Teachers Pay Teachers. It was not that I didn’t believe teachers could gain worthwhile ideas from such sources. My concern was that teachers were often placing these sources above sound research results.

If an article recommends a technique that the authors claim worked for them, I might want to give it a spin in my own classes to see what I think. But I’d have much more confidence in this if someone had objectively evaluated it.  

How have things changed over the past several years?

I’m noticing that teachers are getting a lot more advice about how to teach from journalists, social media influencers, and scientists who may study reading but not reading instruction. These sources usually have little or no experience teaching reading and their knowledge of the relevant research and their ability to evaluate it are questionable at best. I don’t mind when these folks make recommendations as to what they want kids to learn (everybody has a vested interest in such curricular issues), but the best ways to teach something? Nah!

My 2026 take on this is: For the best advice on how to teach reading, turn to studies that explore the effects of reading instruction and what works best. Teachers’ advice based on their own experiences in delivering such instruction can’t replace the research studies but can provide valuable practical information that would never make its way into a research study. The value of advice on how best to teach reading from folks who don’t teach reading and who aren’t relying on instructional research? Not so much.

 2.     Research should not be used selectively.

Many educators use research the same way advertisers and politicians do—selectively, to support their beliefs or claims—rather than trying to figure out how things work or how they could be made to work better.

I wish I had a doughnut for every time a school official has asked me to identify research that could be used to support their new policy! They know what they want to do and seek research that will allow them to say their idea is a good one. What they should be doing is turning to the research first and then using that to determine the best way forward.

Cherry-picking an aberrant study that matches one’s claims or ignoring a rigorously designed study in favor of one with a preferred outcome may be acceptable debater’s tricks but bad science. That can only lead to bad instructional practice.

When it comes to determining what research means, you must pay attention not just to results that you like. Research is at its best when it challenges us to see things differently.

I vividly remember early in my career when Scott Paris challenged our colleagues to wonder why DISTAR, a scripted teaching approach was so effective, even though most of us despised it. Clearly, we were missing something. Our theories were so strong that they were blinding us to the fact that the approach to teaching that we didn’t like was providing something beneficial for kids—at least for some kids or under some conditions (the kinds of things that personal experience can’t reveal).

Anything to add to this in 2026?

These days my concern is not just with cherry-picked evidence but with evidence inappropriate to the question. Recently, I was speaking with legislators who were telling me about the importance of phonics instruction because of the studies of how the brain works.

Those studies are fascinating and have value, but their value is not that they reveal how we should teach phonics, how much phonics to teach, which programs to use, which curricular sequence to focus on, what activities to use and so on.

These studies provide us with a possible explanation for why phonics instruction has been successful in so many instructional studies. Who knows, someday this brain work might even suggest new ideas for designing or redesigning some of that instruction, but we won’t know the benefits of such invention until someone tries it out in classrooms in the context of a rigorous instructional study. All those things mentioned in the previous paragraph about how to teach decoding is something that those neurological studies can’t tell us. Instructional studies need to be the source of the determinative evidence as to how instruction should proceed – if we are really trying to reduce our uncertainty about how best to increase reading achievement. Legislating instructional practice based on brain scans is dopey.

 3.     Research, and the interpretation of research, require consistency.

Admittedly, interpreting research studies is as much an art as a science. During the nearly 60 years of my professional career, the interpretation of pedagogical research has changed dramatically.

It used to be entirely up to the discretion of each individual researcher as to which studies they’d include in a review and what criteria they’d use to weigh these studies.

That led to some pretty funky science: research syntheses that identified only studies that supported a particular teaching method or inconsistent criteria for impeaching studies. What that would mean is that a study would be ignored because it suffered from a serious design flaw unless its findings matched with what the researcher wanted to find – then that design flaw was no longer a problem.

Since I wrote this, research has found long-term negative consequences for Reading Recovery (May, Blakeney, Shrestha, Mazal, & Kennedy, 2024). Studies have long reported short-term gains (e.g, May, Gray, Sirinides, Goldsworthy, Armijo, Sam, Gillespie, & Tognatta, 2015; Shanahan & Barr, 1995), but the long-term harm of the approach is greatly concerning.

When this blog first posted, those long-term consequences were unknown. At that time, it was common to hear the short-term positive results of Reading Recovery dismissed because it was taught one-to-one (“anything works if it is taught one-on-one,” critics would say).

That dismissal disturbed me, not because I was a big Reading Recovery fan (I was always skeptical), but because the folks who were pooh-poohing Reading Recovery in that way, were simultaneously embracing the phonics results of the National Reading Panel (NICHD, 2000). The problem with this? Many of the studies that reported strong positive results had evaluated one-on-one phonics instruction, which didn’t seem to bother these critics at all.  

I’ve retained this paragraph from the 2018 version: “I’m persuaded that both phonics and Reading Recovery work (because they both have multiple studies of sufficient quality showing their effectiveness). That doesn’t mean I think they work equally well, or that they are equally efficient, or that they even accomplish the same things for students.”

That paragraph still makes the important original point that we need to evaluate all research and research results using the same standards and criteria. But it now also reveals the fact that what we think we know from research may change over time. New studies may overturn previous research-based practices. That is not a problem, that is a strength of science.

In the arguments over the “science of reading,” phonics opponents often have argued that it is inappropriate to require phonics instruction since there is no such thing as “settled science“ (e.g., Tierney & Pearson, 2024). They’re not wrong about the nature of science – new better research can always alter our perspectives and practices and in that sense science can never be settled.

Nevertheless, that does not mean we cannot proceed to apply science to our lives. No physician would withhold today’s most effective cancer treatments because there is no such thing as “settled science.” We apply the best knowledge that we have and continue through science – not rhetoric – to keep adding to what we know, even if it means that in the future we will need to alter our course (which is a good reason to resist overly prescriptive legislation even if it appears to be consistent with today’s scientific findings, since up the road science may make such legislation problematic).   

If we want to do better in reading education, we’re going to need to rely on science. That means valuing science over loud opinionizing, evaluating scientific information consistently, and rejecting arguments that there is never enough research to dictate our actions.

Let’s do better.

References

May, H., Blakeney, A., Shrestha, P., Mazal, M., & Kennedy, N. (2024). Long-term impacts of Reading Recovery through 3rd and 4th Grade: A regression discontinuity study. Journal of Research on Educational Effectiveness, 17(3), 433–458. https://doi.org/10.1080/19345747.2023.2209092

May, H., Gray, A., Sirinides, P., Goldsworthy, H., Armijo, M., Sam, C., Gillespie, J. N., & Tognatta, N. (2015). Year one results from the multisite randomized evaluation of the i3 scale-up of Reading Recovery. American Educational Research Journal, 53(3), 547-581.

National Institute of Child Health and Human Development. (2000). Teaching Children to Read: An Evidence-Based Assessment of the Scientific Research Literature on Reading and Its Implications for Reading Instruction: Reports of the Subgroups. Bethesda, MD: National Institute of Child Health and Human Development, National Institutes of Health.

Shanahan, T., & Barr, R. (1995). Reading recovery: An independent evaluation of the effects of an early instructional intervention for at-risk learners. Reading Research Quarterly, 30(4), 958–996. https://doi.org/10.2307/748206

Tierney, R. J., & Pearson, P. D. (2024). Fact-checking the science of reading: Opening up the conversation. Literacy Research Commons. https://literacyresearchcommons.org

November 3, 2018 Version

Shanahan On Literacy Podcast

Comments

See what others have to say about this topic.

What Are your thoughts?

Leave me a comment and I would like to have a discussion with you!

Comment *
Name*
Email*
Website
William Feerick Mar 14, 2026 11:34 AM

Thank you Dr. Shanahan for re-posting this piece. I love learning from your knowledge, years of experience, wisdom, and commitment to evidence and research. I’ve been teaching first grade for 27 years so when new teachers come to me and ask for recommendations about how to teach reading effectively, I often refer to your blog.

If a new first grade teacher were to come to you and ask, “Where should I start if I want to learn more about how to teach reading effectively since my graduate classes featured texts (many by Heineman) advocating approaches that weren’t always research-based?” Where would you have them start?

Dr. Bill Conrad Mar 14, 2026 11:40 AM

Congratulations Tim on 60 years of amazing contribution and inspiration for the quality teaching of reading!

While the use of quality research is important in supporting reading instructional practices, I think it is also important to go beyond addressing individual teaching practices and pay attention to the big picture.

The theory of action for learning to read is important to keep in mind. In order to successfully read. Students need to be able to hear and recognize the 44 small sounds that make up oral speaking - phonemic awareness. Next students need to be able to link these sounds with written letters - phonics. Next students need to be able to read with automaticity - fluency. Once these elements are in pace students can focus on comprehension - deriving meaning from the text. Vocabulary, writing, and spelling also play a key role in the process of reading effectively.

Of course, you should be given the credit for discovering this research-based theory of action for reading!

Too often teachers make the claim that they know their kiddos! They may claim that all of them are reading at grade level. I encountered this phenomenon within the Redwood City School District in California. To test their claim, we independently assessed the key elements of reading using the valid and reliable DIBELS assessment. Based upon test results we found that only about 30% of students in grades K-3 were reading on grade level.

Teacher impressions of student reading ability must be backed up with data derived from valid and reliable reading assessments aligned with the essential elements of reading. Accept no substitutes!

Lauren Mar 14, 2026 03:38 PM

Thank you for this thoughtful post. I am interested in the study that found long term negative consequences with Reading Recovery. When I was using a very systematic phonics/decodable only program, I reached the conclusion that it created an artificial reading experience for students...sort of a "bubble". Within the bubble students appeared to be making progress, but when they tried to read texts that were not completely decodable, they floundered.

By the way, Dr. Conrad, I would trust a teacher with many years of experience teaching children to read over those online testing systems. The charts and color coding are seductive, but they are fraught with problems and inconsistencies. Trusting AI over the judgement of professional, experienced human beings is the real danger of our times.

Dr. Bill Conrad Mar 14, 2026 04:54 PM

Hello Lauren,

Thank you for your insights.

Research indicates that teacher judgments of student academic success are generally moderately correlated with standardized test scores, typically explaining 25% to 35% of the variance.

Teachers are reasonably accurate, but often overrate weak readers and factor in non-test elements like classroom effort, behavior, and motivation, which can cause discrepancies.

Key Findings in Research Support:
Moderate Accuracy: Teacher judgments show moderate accuracy in predicting student achievement as measured by standardized tests, according to studies such as Südkamp et al., 2012 and others.

Different Measures: Teacher-assigned grades and standardized tests measure different but complementary aspects of achievement; teacher judgment often captures student engagement, This complementary effectiveness of student motivation, and effort better than standardized tests, which measure specific domain knowledge. This complementarity aligns with your claim about teacher professional diasgnosis of strudent performance overall.

Contextual Factors: Teachers tend to be more accurate in assessing students' academic self-concept when they have deep knowledge of the student, but may struggle with objective evaluation due to student behavior or classroom dynamics.

Discrepancy in Weak Readers: Several studies suggest that teachers tend to overestimate the skills of weak readers compared to their actual performance on standardized reading tests.

Predictive Validity: While standardized tests can be more consistent in measuring attainment, the combination of teacher judgments and standardized tests often provides the most robust view of a student's true academic success. This in alignment with your thinking Lauren.

Long-Term Impact: Research by Chetty et al. (2014) indicates that teacher effectiveness, as predicted by test score gains, is directly related to long-term student outcomes like college attendance and higher income.

Factors Influencing Accuracy:
Bias & Expectations: Teachers' expectations can create self-fulfilling prophecies, where lower-performing students may receive less focus, causing them to underperform, according to research from Rubie-Davies (2010).

Professional Development: Training on how to use assessment data can improve the accuracy of teacher ratings in relation to objective test results.

Grade Inflation/Differentiation: Grading practices often deviate from standardized measurement techniques, leading to differences between class grades and standardized test scores, notes studies from Cross & Frary (1996).

In summary, teacher judgment and standardized tests are not entirely aligned, but rather complementary. Standardized tests offer a more standardized, objective measure of achievement which of course is critical in assessing students’ ability to read while teacher judgments provide context, including behavioral and motivational factors that are important for overall student success, say many experts.

So let’s not rage against the thermometers and instead use them in complementarity ways with teacher professional judgment.

Timothy Shanahan Mar 14, 2026 05:41 PM

William--
Any of these sites could be useful in such a situation:

Timothy Shanahan www.shanahanonliteracy.com
Florida Center for Reading Research https://fcrr.org/educators
Johns Hopkins University https://bestevidence.org/category/reading/
What Works Clearinghouse https://ies.ed.gov/ncee/wwc/
WestEd https://www.wested.org/focus-area/literacy/

tim

Mary Baker-Hendy Mar 14, 2026 06:24 PM

Hi Dr. Shanahan and William,
Seeing the list of reliable resources for educational research reassures me that I have using and recommending the right sources over the years.
I’d like to add a plug...William, make sure you check out the IES Practice Guides on the What Works Clearinghouse site! These guides are written by the expert researchers of specific educational topics and issues. They describe a set of recommendations based on the compiled and reliable research reviewed by the authors. They describe the practices, include examples and resources, and are concise and readable.
Check out Dr. Shanahan’s, Guide on Improving Reading Comprehension in Kindergarten Through 3rd Grade, 2010. Reading the names of the panel members reads like a ‘who’s who’ list for this old timer! They’ll steer you right. Mary BH

Lauren Mar 14, 2026 08:28 PM

Thank you for your detailed response, Dr. Conrad. I agree with you that it does not have to be either or... I can only share with you my practical concerns as a person who administers these tests to students.

Loss of instructional time: With the test you mentioned, we lose fifteen days of Reading Intervention instruction per year to administer the test to every student in the school at the beginning, middle, and end of the year. If you do the amount of progress monitoring recommended,
you have to take at least one period of intervention twice a month to complete the testing for a small group. This is an additional twenty days of lost instruction. State tests take two weeks of instruction, and there are other online testing platforms aimed at vocabulary, fluency, and comprehension that take at least a week to administer.

Mismatched levels/moving target: If a reading intervention student reads at a 1.8 level coming into third grade, they are tested at a 3.0 ORF level on these tests. They do poorly (not a surprise). If they make great progress and are reading at 2.5 by the middle of the year, they are tested at 3.5. They do poorly (not a surprise). If they are at 3.3 by the end of the year, they are tested at 4.0. They do poorly (no surprise).

Timed component: Tests are timed. This is why they are a good screener to identify students who process reading/ language more slowly. This is why they are not a particularly good progress indicator for struggling readers.

Nonsense words: Nonsense words can be very confusing for struggling readers and all young readers. An example is that students will decode short vowel words such as "tap" as "tape" once they start studying long vowel sounds in class. There are better ways to measure decoding skills. The whole nonsense word thing really should be eliminated.

Only test to find out what you don't know: These tests don't tell me anything that I don't know from working with my students daily, so for me, they are a waste of time.

I hope these insights are informative.



Dr. Bill Conrad Mar 14, 2026 10:44 PM

Hi Lauren,

You make some good points. However, we should never let perfect be the enemy of the good. We need a method for gauging student phonics knowledge as opposed to sight word knowledge.

Some additional points can be found below arguing both sides of the argument.

Teachers' frustration with Nonsense Word Fluency (NWF) tests often stems from a "teach-to-the-test" culture, but research suggests these assessments are uniquely effective at identifying whether a child is actually decoding or simply memorizing.

The Teacher's Perspective (The "Point")
Teachers often argue these tests are a waste of time or inaccurate for several practical reasons:

Artificial Task: Real reading involves meaning. Decoding "wug" or "zot" feels disconnected from the ultimate goal of comprehension.

Instructional Distraction: When schools prioritize NWF scores, teachers may spend valuable class time practicing nonsense words rather than reading real stories. Experts like Timothy Shanahan call this practice a "waste of time" that invalidates the test.

Penalty for "Good" Readers: Some advanced students instinctively "correct" nonsense words to real ones (e.g., reading mump as jump), which counts as an error despite being a sign of high-level linguistic processing.

Time Consumption: Administering these tests 1-on-1 every few weeks can eat up hours of instructional time across a classroom.

The Counter-Argument (Why They Are Used)
The primary defense for NWF is that it provides a "pure" measure of decoding that real words cannot.

Prevents "Faking It": Students can memorize thousands of real words by sight or use picture clues to guess them. Nonsense words force a student to use their phonics knowledge because the word is guaranteed to be unfamiliar.
Identifies "The Wall": Many struggling readers (including those with dyslexia) appear to read fine in 1st grade by memorizing simple words. They hit a "wall" in 3rd or 4th grade when words become too complex to memorize. NWF catches these gaps early.

Predictive Power: Studies show that NWF scores in early grades are strong predictors of later reading comprehension and fluency

George LILLEY Mar 15, 2026 12:53 AM

Tierney and Pearson’s argument that the science remains unsettled is defensible. This post itself illustrates why: May et?al. (2024) is presented as a robust study, despite the fact that the What Works Clearinghouse—an authority you also endorse—found that it failed to meet WWC standards for multiple reasons, most notably attrition rates above 75%.

Angela Healy Mar 15, 2026 09:51 AM

I appreciated the blog’s reminder that research should be interpreted carefully and that evidence should be judged using consistent standards. As someone involved in teacher education and the implementation of Reading Recovery in Ireland, I take that responsibility seriously.
For that reason, I was struck by the use of the word harm in relation to Reading Recovery.
When the word harm is used in relation to an intervention, it sets a very high bar for evidence. The May study reports a negative long-term association in later test scores, but moving from that finding to a conclusion that the intervention itself causes harm seems like a stronger claim than the current evidence may support. In most areas of education research, conclusions of harm would normally require consistent findings across multiple studies. As teacher educators we try to help teachers interpret research carefully, and that usually means distinguishing between findings that raise important questions and conclusions and those that suggest an intervention is actually harmful.
I also wonder how differences in later instructional support are accounted for. Those of us who work both in teacher education and directly with the lowest-achieving readers are very aware of how complex children’s learning trajectories can be over several years.
In any case, I hope you have a very happy St Patrick’s weekend. I suspect many of my Reading Recovery colleagues here may be too busy celebrating to have read the blog just yet. We welcome all research into our intervention here as it helps us challenge our assumptions, refine practice, and improve outcomes for all children.

Lauren Mar 15, 2026 05:36 PM

Thank you for presenting both sides of the nonsense word issue, Dr. Conrad. I don't agree with your assertion that we need a method for gauging student phonics knowledge vs. sight word knowledge. Do we? Children have been learning to read in public schools for many decades without these online testing platforms. Who is making money off of these platforms? Why do we accept that they are a useful, necessary thing? Classroom teachers teach and informally assess these skills daily as they have for many years. I can tell you that the struggling readers who process language more slowly do poorly on these measures all through their primary grades, because the tests are timed, and the passages and sight words are not set to their ability. The nonsense word reading is confusing. Why make them take these assessments over and over again? It's kind of like the definition of insanity isn't it? Are NAEP scores going up?

Dr. Bill Conrad Mar 15, 2026 06:45 PM

Lauren,

Let’s consider the data.

The 2022 NAEP Reading test report that fewer than 1/3 of 4th grade students read at proficient or above reading levels.

The teacher is the most important variable in determining whether students learn to read! There is a wide discrepancy between how many students are learning to read versus how many students actually learn to read!

Houston? We have a problem!

The psyshometricians who design. Validate and test the reliability of tests like DIBELS and NAEP are far more knowledgeable about actual student performance than the average teacher! No?

Surgeons don’t question the ability of heart-Lung machines Phone users don’t question the competence of the engineers who design the phones.

Better to stop raging against the thermometers. Use them well and stay in your lane.

What am I missing?

Jennifer S. Mar 15, 2026 08:48 PM

Thanks for another great article! You say the best research to follow for teaching kids to read is research on reading instruction. What about cognitive psychology and things like cognitive load? Is there reading instruction research that speaks for or against cognitive load theory and/or the instructional hierarchy when designing reading lessons and instruction?

David Bermingham Mar 15, 2026 10:08 PM

Hi Tim, I totally agree with many of the points you raise in this post. We have to be careful consumers of research. We must be aware of the strengths and weaknesses of different forms of research. One of the most common pitfalls in evaluating educational research is the desire to elicit simple and actionable lists of ‘do’s and don’ts’, that can be enacted undigested in the classroom. As we know, classrooms are noisy and multifaceted environments with groups of children who present with varieties of strengths and challenges, motivations, interests and backgrounds. The young learner’s brain is a rich and complex assembly of billions of neurons creating different pathways that results in each different learner perceiving and learning from their environment in a unique fashion. This is why the enactment of research to practice requires agentic and well informed professionals who are tentative, observant and contingent. This is also the key tenet of the scientific approach, no answer is definitive, we are always on an iterative journey, evaluating evidence, proposing a refined thesis based on this, evaluating again and refining. As a teacher with over thirty years experience at all levels in primary schools and researcher who has conducted study in the area of literacy development I have always been fascinated by the how our brains engage with literacy. I find it’s important to remember that written language, particularly, is a human artefact, a technology and a relatively new one, evolutionary speaking. This is why brain scans show that the act of reading and writing engages many different parts of the brain simultaneously, and that there is no one specific ‘reading and writing’ part. Knowing this, we must be especially resourceful and flexible in how we approach the teaching of it. Neuroscientific research does show that our brains are ultimately pattern seekers, and making meaning of our environment is fundamental. We also know that the human brain is hugely energy consuming and cognitive drain occurs quickly and easily, particularly with literacy learners experiencing challenge. Any approach that doesn’t keeping meaning at its core will be vulnerable to learners experiencing reduced motivation to persist, particularly those for whom isolated skills and drills lack connection to what their minds perceive to be stimulating, engaging and applicable to the curiosities they have about the world around them. This is why an approach that tries to divide the skills and strategies needed to become a successful reader and writer, to to a neatly sequenced (adult created) set of steps, risk undervaluing the little human minds you need to reach. This is particularly true for those who, for any number of reasons, struggle in the whole class environment.
I have been RR teacher and teacher-leader for over a decade as well as a researcher in the area of literacy acquisition. The approaches taken in this intervention have proven to me that when we work with a struggling literacy learner on an individual basis we have the privilege and responsibility to ensure that the pupil becomes an accurate reader who is motivated to continue to speak, read and write, and engage as a confident and competent individual. They must be able to engage with the mechanics of language with accuracy and automaticity and must be happy to do this because they find it meaningful and therefore motivational. Most of all they must become an independent and agentic learner themselves with the capacity to employ the skills and strategies they have learned as they engage with language in all its forms with their peers back in the whole class environment. The approach is to harness the pupils strengths to conquer their challenges. Because of the one to one nature, we have the ability to make the programme completely bespoke to the child themselves, using what sense they have made of the world around them to build new connections and understanding; as Hebb’s rule states ‘neurons that fire together wire together’, so we build forward to ensure that unhelpful habits or misunderstandings are extinguished and replaced with generative skills and strategies. RR approach is not necessary for every pupil, it does not exist to replace good systematic classroom pedagogy but to support those few learners who are experiencing confusions at this level. If left unsupported such confusions can become entrenched and more difficult to untangle as the child progresses through school. The danger for such children is the old maxim that ‘practice makes permanent, only perfect practice makes perfect’.
Having said all that, there are a certain cohort of children who present with more profound literacy processing challenges and will need to continue to receive classroom or specialised supports for a longer period of time, in this situation RR serves as a longitudinal observational diagnostic support, which, along with other diagnostic tools, helps the school to create the best and most appropriate supports for such learners. Finally, when a child experiences accelerated learning during the RR intervention and is experiencing success and independence that allows them to benefit from the classroom teaching environment, we have delivered them to the starting line, not the the finish. As early learners (kindergarten/first grade) they still have a long journey ahead but at least now they are at the same starting point as their peers, allowing the classroom teacher to create a more inclusive learning environment

Timothy Shanahan Mar 15, 2026 10:24 PM

Jennifer-
Such research is quite valuable. It provides wonderful hypotheses as to how we can improve education. However, it is always insufficient. No matter how convincing such studies may be, their results have to be translated into pedagogical practices and those practices, ultimately, must be evaluated in terms of their actual effectiveness. Quite often the insights drawn from such research turn out not to be translatable into successful teaching.

tim

Timothy Shanahan Mar 15, 2026 10:28 PM

Angela--
I was referring to research that identified students who were qualified for RR. Those students were then assigned to the RR treatment or not. By the end of grade 1, the RR students had clearly made gains. However, at a later time, the comparison showed not that they lost their gains, but that they had fallen behind the students who had not received the treatment. That is a harm.

tim

Angela Healy Mar 16, 2026 12:28 PM

Tim, thank you for clarifying that. I understand the point you are making about the later comparison.
In truth the question I still wrestle with is whether that later difference can confidently attributed to the Reading Recovery intervention itself, or whether it reflects the many different instructional experiences children encounter in the years afterwards. In practice, children’s literacy pathways often diverge significantly after first grade, and long-term studies inevitably struggle to account fully for those differences.
I also notice that the study reports quite high attrition (over 75%) in the follow-up sample. In my experience working with the lowest-achieving readers, many come from highly mobile or vulnerable backgrounds where school attendance, family circumstances and wider socio-economic factors can significantly shape learning trajectories. Tracking outcomes for the same children over several years in those contexts is understandably challenging.
My understanding of regression discontinuity designs is that they are extremely powerful for examining effects close to the intervention point, which may explain why the study reproduces the strong short-term gains of the intervention reported elsewhere. Interpreting outcomes several years later, however, becomes much more complex. For that reason, it seems premature to interpret the later difference as evidence that the intervention itself causes harm.

Gaynor Chapman Mar 18, 2026 08:44 AM

The main ingredient in scientific research is truth and unfortunately not every researcher is committed to the truth but instead their pet ideology. Literacy history shows ideology trumped truth for decades.
I still struggle to understand how academia dominated literacy for so many decades promoting destructive anti phonic methods .
I have lived through the reading wars since my mother resolved to try and change the system when she saw so many normal and bright children failing to learn to read in a high income suburb. She had trained in the 1930s when you lost grading as an infant teacher in NZ unless you succeeded in having every child in the class reading by 7 years old . This was the era of universal literacy and it was an absolute slog to achieve this but quite impossible without traditional intensive phonics .
Then along came the ideology in the form of Progressive Education and the aim was no longer literacy but all sorts of social aims. Consequently there was reading failure such as had not been seen before . Marie Clay 's, wonderful new idea of 'guessing' and Whole Language was supposed to be the answer to the reading failure. She did some crumby , dishonest research to prove her theory and despite mountains of good honest research showing otherwise, she dominated the scene . Academics who tried to promote phonics were denied research money , censored and prevented from having their papers published. Many teachers and parents saw Whole Language and Reading Recovery weren't working but they weren't 'clever ' academics ' so were ignored. It is astonishing to me how something that was clearly failing continued on for so long. But that is the nature of ideology -it captures and snares people's thinking even though they may be very clever.
How do we prevent such a thing happening again ? Whole Language has destroyed the futures of so many millions of children . Note that it has also been private bodies who have conducted significant endeavours to get us back on track, like the Mississippi Miracle sponsored by a generous millionaire and the research sponsored by the shire council of Clackmannashire Scottish research as well as the efforts of academics like Henry May et al. . Emily Stanford with her journalism also helped . Where academia failed being hindered in producing truth others stepped in . Academia need to also humbly acknowledge they went badly astray and also much of what is being done now is reinventing the wheel from what we had in the past . Recent research confirms this. Honesty is paramount .

Timothy Shanahan Mar 18, 2026 06:00 PM

Angela--

A study this well done, with nearly 10,000 kids, and no bias evident in the attrition, deserves attention (especially given the unlikelihood of it ever attracting this level of research again). Even if you set aside the negative results and just conclude that there was no longterm benefit to RR, that should be enough to discontinue the program (given its expense). That expense is especially problematic given another finding in the May study: these students required as much or more follow up remediation as their similar classmates who were not enrolled in RR. That means the cost savings (and life improvements) promised by RR advocates were a pipe dream.

Tim

Angela Healy Mar 18, 2026 10:28 PM

Tim, thank you for taking the time to respond.

I do think the discussion has shifted slightly. We began with the question of harm, and are now moving toward a question of long-term benefit and cost. Those are valid policy considerations, but they raise a different set of issues. For those of us working in disadvantaged school contexts, interventions like Reading Recovery are judged not only on long-term averages, but on whether they provide early access to literacy, accelerate learning at a critical point, and change a child’s trajectory at the start of schooling. The evidence — including in this study — suggests that this early acceleration does occur.

For that reason, I want to return briefly to the use of the term harm, which is a significant claim, particularly for those of us working closely with the most vulnerable readers. The same study shows that children receiving Reading Recovery make substantial gains in first grade, and for many of these children that represents their first experience of success in literacy. In that context, it is difficult to reconcile those early gains with a characterisation of the intervention itself as harmful.

In terms of the study itself, I would be cautious about how far we can generalise from the long-term findings. The comparison group consisted of children just above the cut-off for intervention, rather than the lowest-achieving readers, and the Reading Recovery group included a wide range of students, including those who did not complete a full series of lessons. Over time, those groups are likely to diverge further in their instructional experiences. It is also notable that the study is not included in What Works Clearinghouse intervention evidence, which reflects the challenge of meeting the highest standards for causal interpretation in long-term follow-up studies of this kind.

In relation to the high levels of attrition in the study, I agree that while statistical checks may not indicate bias, in practice they do not eliminate real-world factors such as mobility and vulnerability. The most vulnerable students are often the hardest to track over time, which raises questions about how representative the remaining sample is of the original population. More broadly, the analysis cannot fully capture the range of factors that shape children’s trajectories over several years — including attendance, additional supports, wider socio-economic context, and the instruction they receive after the intervention.
Inevitably, the study simplifies a highly complex educational reality.

Timothy Shanahan Mar 18, 2026 11:48 PM

Angela--
As former director of reading for the Chicago Public Schools where I was responsible for 437,000 children 85% of whom were from low income families, I think I know something about the nuance of this:

"For those of us working in disadvantaged school contexts, interventions like Reading Recovery are judged not only on long-term averages, but on whether they provide early access to literacy, accelerate learning at a critical point, and change a child’s trajectory at the start of schooling."

You are saying here that the only thing that matters is the amount of immediate growth you are likely to see in a very small sample of grade 1 children. As someone who has taken much more responsibility than that, I must say that we need to worry not about first graders but preschoolers to 12th graders, and, indeed, we need to worry about more than one years growth for this small group of children but also about long term gains for all boys and girls. For most schools, Reading Recovery costs their entire expenditure on reading improvement for all kids. That it is likely to only help 4 or 5 children in grade 1 and that those long term gains are at best transient and at worst a burden for the children who attain them, that is a very big issue for those of us who take responsibility for disadvantaged children.

You're missing the nuance that has been apparent with the research on RR for a very long time (go back and read my 1995 analysis of those studies and you'll see what I mean).

tim

tim

Angela Healy Mar 19, 2026 12:23 PM

Tim, thank you for your response. I agree that long-term outcomes, whole-school priorities and value for money are essential considerations. My point was not that immediate gains are the only thing that matters, but that early acceleration for the lowest-achieving readers is one important part of the picture and should be considered alongside longer-term evidence.
Where I still hesitate is in treating several different questions as one. Whether an intervention produces short-term gains, whether those gains are sustained, and whether it represents the best use of limited resources are related but distinct issues. The May study may well raise important questions about sustainability and cost-effectiveness. I am less persuaded that it justifies the stronger claim that the intervention is harmful.
From my perspective, that still calls for a more cautious interpretation of the evidence and a higher evidential bar than a single long-term study can provide. That also seems consistent with the principle in your blog about applying careful and consistent standards when interpreting research evidence.

Angela

Susan Bodman Mar 19, 2026 03:24 PM

Tim, as you said in 2022, “There are no magic beans when it comes to early literacy. The trick is to catch kids up early and then to continue to strive to keep them caught up.” It’s that second part — keeping them caught up — that’s under scrutiny here.
Acceleration of rate of progress at one point (demonstrated by May and colleagues) followed by normal developmental variation cannot reliably support an interpretation that the intervention is doing the ‘harm’. What evidence would be needed to claim in a safe and objective way that an intervention does ‘harm’? Surely evidence would need to include:
• long term negative outcomes directly attributable to the intervention
• consistent patterns across multiple high-quality independent studies in different geographical regions and contexts.
In fact, studies like those on the Every Child a Reader project in England do show long term benefits, contradicting the interpretation that the intervention itself is harmful. Simply, there is not the research base to conclude whether any long term dip is actually due to the quality of subsequent instruction, or the realities of children’s lives as they move through school.
What this debate vividly illustrates is that the available evidence around the long term gains of any early intervention is mixed, highly contested, and highly dependent on study design, attrition, and the quality of instruction students receive after the intervention.
Thank you for reading,
Sue

Lauren Mar 19, 2026 03:42 PM

I think we will have to agree to disagree on this one, Dr. Conrad. The color coded data gives everybody something to look at at staff meetings and board meetings, but if you actually administer these assessments to young children, you quickly see that there are issues with validity and reliability. They are fairly useful as screening tools, but everyone already knows which students need intervention help anyway. I again go back to the issue of wasted instructional time. We are overtesting. It also goes back to the fact that a culture is created where people teach to the test instead of teaching children to read: Teaching nonsense words, and teaching children to read as fast as they can with no regard for
comprehension or prosody. There is no common sense here. What algorithms are used to come up with the numbers and colors? If a teacher has students practice nonsense words, the scores will go up. (Wow, these students are really making progress!). It's ridiculous.

Gaynor Chapman Mar 20, 2026 12:22 AM

I have strong emotions on the damaging effects of WL and RR. My own child was taught a good phonic home programme as a preschooler , last century and could read simple controlled vocabulary readers . But after six weeks at school with WL and the leveled predictable texts she could not read at all . Her brain was somehow scrambled . It took months to remediate her. Other teachers have noted this handicap, as well.
I was an assistant to my mother who remediated 1500 students with intensive phonics in the WL and RR era late last century. . The confusion and compulsive guessing of many students required some to be homeschooled for a while until they had stopped making wild guesses which the schools were actually encouraging . It was such hard work remediating RR failures that my mother told parents she was reluctant to take them as students.
Besides the crazy guessing there were the behaviour problems these failing children developed. Australian Kerry Hempenstall has written about this in detail . Failing children stop trying anymore .
Because of these experiences I would not give any struggling reader those infernal leveled readers which promote 3 cueing strategies. even if these WL methods are not employed explicitly. This probably means not those from disadvantaged homes or the lowest 25% in any class. Certainly not dyslexics,
What Marie Clay promoted , I believe, selectively damaged less able students. She was a fraud because those are the very students she proclaimed she was helping.

Marcia Long Mar 20, 2026 10:58 PM

Hi, Tim,

I love the way you use "schemes" when talking about instructional programs or plans. I see a lot of schemes in my line of work as I support different schools. I also appreciate the analogy with relying on science to save cancer patients; however, I still get the response that "you can find research that says anything" from administrators and district leaders. One big issue for me right now is trying to persuade teachers and principals that we must continue teaching students to read and not abandoning teaching for the last quarter of the year. They still have the mindset that testing students is the same as teaching students, no matter how much research I present. One big issue I face is that they do not support teaching syntax, morphology, or language as the state test approaches because our state's interim reading assessments highlight comprehension as the issue. I thought about using an analogy of teaching any other foreign language: We would never just continually hand students studying a second language texts with questions and say "do this" without teaching how the language works. Even our best and brightest struggle with Shakespeare, Hawthorne, and difficult texts of all genres, and two-thirds of our students are more than one grade level or more behind in reading. Many are not fluent. Our state (NC) also hurts the cause because they give back interim assessment data in standards, so it reinforces the misconception that we can isolate standards, such as theme/central idea, tone, author's purpose, text structure, and we should focus on one area at a time through multiple-choice tests. So, the focus is wrong and the way they want to remediate is not research or evidence-aligned. Do you think the foreign language analogy is good? Any suggestions for persuading leaders who will not read the research or listen? I thought about trying a campaign of reading myth busting, similar to one of your blogs, but in an incremental fashion.

Timothy Shanahan Mar 21, 2026 04:16 PM

Marcia--
I am never sure how to get administrators on board who are sure they know how to raise reading achievement (usually by some kind of ineffective or irrelevant test practice), and consequently resist any thing that works. This was a huge problem for me in the Chicago Public Schools. The Mayor wanted a big emphasis on test prep and I resisted. They eventually over-ruled me, putting in some test prep at the last minute (perhaps a week or so before the tests) -- not enough time to do much damage. Historically, the district had started test prep immediately after winter break, and this would continue until April when the tests were given. By not imposing such a scheme, we managed to accomplish the biggest test gains in history -- all that extra teaching paid off. Nevertheless, I was unable to persuade the powers that be, that teaching can beat test prep. Good luck.

tim

Comments

The Why and How of Research and the Science of Reading

27 comments

One of the world’s premier literacy educators.

He studies reading and writing across all ages and abilities. Feel free to contact him.

Timothy Shanahan is one of the world’s premier literacy educators. He studies the teaching of reading and writing across all ages and abilities. He was inducted to the Reading Hall of Fame in 2007, and is a former first-grade teacher.  Read more

60 E Monroe St #6001
CHICAGO, Illinois 60603-2760
Subscribe