Opinion
School & District Management Opinion

What Does It Mean to Call a Program ‘Evidence-Based’ Anyway?

There’s a good reason not to stop after finding a single positive study
By Fiona Hollands, Yuan Chang & Venita Holmes — June 09, 2021 5 min read
A researcher points to charts and data
  • Save to favorites
  • Print

Millions of federal dollars are currently streaming into schools to help address learning shortfalls related to the pandemic. While it’s not clear what strings might be attached to these new funds, recent versions of the Elementary and Secondary Education Act have encouraged and, for some uses, required education decisionmakers to invest federal funds in evidence-based activities. Many school leaders themselves want evidence to inform their choices. But how does that work in practice?

The current federal approach may seem either needlessly restrictive or absolutely sound, but the fact is the bar for what counts as evidence isn’t awfully high: The Every Student Succeeds Act, the latest ESEA iteration, asks for only one study showing that a program improves student outcomes. And if there isn’t one, a sound rationale for the activity and a plan to evaluate it will often suffice.

To figure out how well schools and districts are able to meet this directive, our research team worked with a large, Southern school system to determine what proportion of its Title I funds are spent on evidence-based activities. Our findings—which relate to prepandemic federal funds—reveal weaknesses both in the law and in the evidence base. As a result, we have some advice to share with educators who are doing their best to adhere to ESSA, serve students well, and now wisely spend large amounts of COVID-19 relief funds in record time.

First, let’s address the question of what should count as adequate evidence when choosing where to invest education funds. Does it really make sense to declare a program “evidence-based” if one study shows positive outcomes for students while a bunch of others don’t?

In our analysis, we first determined what percentage of Title I funds at the district office and four sampled schools were spent on programs and practices backed by just one study: 70 percent in a three-year period. It’s likely that this number would have been substantially higher if we had descended below the top three tiers of evidence laid out in ESSA to that nebulous world of activities for which there is a research-based rationale and an ongoing evaluation plan. To evade this squishiness, we limited our definition of positive evidence to studies using one of three rigorous approaches for evaluating effectiveness: experimental, quasi-experimental, or correlational methods. Additionally, each study needed to show gains for students of approximately the same age as those being served at the district.

Next, we considered all the relevant studies we could find in four research repositories: What Works Clearinghouse, Evidence for ESSA, Education Endowment Foundation, and Education Resources Information Center (ERIC). We agreed on a summary rating for each program or practice to reflect the overall body of evidence. Using these ratings, we found that only 50 percent of the funds were spent on evidence-based activities. The punch line here is that, for 20 percent of the Title I investments, the conclusions drawn from a single study run counter to conclusions drawn from multiple studies of the same intervention.

v40 36 Hollands Chart C

The takeaway for those of you holding the purse strings is don’t stop at the first positive study you find.

Ironically, the more often a program is studied, the less likely it is always found to be effective. This means that ESSA can inadvertently tip the scales toward less studied—but not necessarily more promising—interventions. So a second lesson is when the evidence is mixed, look carefully at what kinds of students were served in each study, which outcomes were improved, and what skills and resources were needed to implement the program. If these don’t match your own context in at least some studies that found student improvements, be cautious about the likelihood that you can replicate positive results.

There’s also the problem that, for some applications of Title I funds suggested by ESSA, there just isn’t any evidence at all. This means that expecting 100 percent of Title I funds to be invested in evidence-based activities is an unreasonable goal. There are no rigorous studies, for instance, of many widely implemented and seemingly essential programs like student-data tracking and school nurses.

Most important, it’s worth questioning the apparent assumption that educators have the time and skills to search for and evaluate all this evidence. It took us a year to complete our study even though we’ve been trained extensively in how to do this kind of work. Who has that kind of time in a school or district office? It’s unrealistic and inefficient to expect educators and administrators in every school and district to thoroughly investigate each new practice they consider. It would be more helpful for state education departments and district central offices to provide clearer guidance to the decisionmakers they support on suitable investments of federal funds.

See also

Every Student Succeeds Act NCLB Rewrite Sets New Path on School Research
Sarah D. Sparks, January 5, 2016
7 min read

Some education agencies, for example, the New York state education department, are already doing this. Right now, suggestions on how to address learning loss are desperately needed—and these need to be more specific than “high-quality tutoring.” Detailed exemplars are called for to show what that looks like in practice to address different student needs.

Our research leads to one more major piece of advice. Even a well-curated list of evidence-based activities cannot assure effective implementation of promising practices. Schools may be more successful at scaling up effective programs they are already using or at tweaking implementation of their underperforming programs than at introducing brand-new strategies, even if the evidence for them is favorable. For example, one district was surprised to find that Reading Recovery was not producing positive results overall for its students despite ample research evidence to support its effectiveness elsewhere. Instead of ditching the program, district officials carefully reviewed whether schools with the greatest number of struggling readers were getting the appropriate share of trained teachers.

Our own research team’s evidence repository lists items currently in use in districts with which we work and aims to help decisionmakers know where to steer their efforts and their funding. It may behoove states and districts to exert their energy in supporting educators to better implement potentially effective programs than expecting them to become researchers overnight.

A version of this article appeared in the June 16, 2021 edition of Education Week as What Does It Mean To Call a Program ‘Evidence-Based’ Anyway?

Events

School Climate & Safety K-12 Essentials Forum Strengthen Students’ Connections to School
Join this free event to learn how schools are creating the space for students to form strong bonds with each other and trusted adults.
This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of Education Week's editorial staff.
Sponsor
Assessment Webinar
Standards-Based Grading Roundtable: What We've Achieved and Where We're Headed
Content provided by Otus
This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of Education Week's editorial staff.
Sponsor
Reading & Literacy Webinar
Creating Confident Readers: Why Differentiated Instruction is Equitable Instruction
Join us as we break down how differentiated instruction can advance your school’s literacy and equity goals.
Content provided by Lexia Learning

EdWeek Top School Jobs

Teacher Jobs
Search over ten thousand teaching jobs nationwide — elementary, middle, high school and more.
View Jobs
Principal Jobs
Find hundreds of jobs for principals, assistant principals, and other school leadership roles.
View Jobs
Administrator Jobs
Over a thousand district-level jobs: superintendents, directors, more.
View Jobs
Support Staff Jobs
Search thousands of jobs, from paraprofessionals to counselors and more.
View Jobs

Read Next

School & District Management Israel-Hamas War Poses Tough Questions for K-12 Leaders, Too
High school students have joined walkouts, while charges of antisemitism in three districts will be the focus of a House hearing this week.
9 min read
Officers with the New York Police Department raid the encampment by pro-Palestinian protesters at Columbia University on April 30, 2024, in New York. The protesters had seized the administration building, known as Hamilton Hall, more than 20 hours earlier in a major escalation as demonstrations against the Israel-Hamas war spread on college campuses nationwide.
New York City police officers raid the encampment of pro-Palestinian protesters at Columbia University on April 30, 2024. Although not as turbulent as what is happening on many college campuses, K-12 schools in some pockets of the country are also contending with conflict stemming from the Israel-Hamas war.
Marco Postigo Storel via AP
School & District Management What the Research Says A New Way for Educators to Think About School Segregation
Seventy years after the Supreme Court's ruling in Brown v. Board, Stanford researchers find racial, economic isolation spiking in schools.
4 min read
First-graders listen to teacher Dwane Davis at Milwaukee Math and Science Academy, a charter school in Milwaukee on Oct. 20, 2017. Charter schools are among the nation's most segregated, an Associated Press analysis finds — an outcome at odds, critics say, with their goal of offering a better alternative to failing traditional public schools.
First-graders listen to teacher Dwane Davis at Milwaukee Math and Science Academy, a charter school in Milwaukee on Oct. 20, 2017. Charter schools are among the nation's most segregated, an Associated Press analysis finds—an outcome at odds, critics say, with their goal of offering a better alternative to failing traditional public schools.
Carrie Antlfinger/AP
School & District Management Opinion How We Can Fix Chronic Absenteeism
Experts on school attendance lay out five steps to ramping up family and student engagement.
Hedy N. Chang & Catherine M. Cooney
6 min read
A young student is sitting at the desk in the classroom and looking worried at the test. The students around him are absent.
Vanessa Solis/Education Week + E+/Getty
School & District Management Letter to the Editor Women Still Face Barriers to Leadership
A letter to the editor discusses the challenges women face in education leadership positions.
1 min read
Education Week opinion letters submissions
Gwen Keraval for Education Week