ESEA Bill Boosts Research Role in Ed. Programs
Measure would solidify authority of IES and funding for evaluation
A slew of quiet changes in the Senate bill to reauthorize federal education law would substantially increase the role of research in federal education programs.
The latest version of the bill reauthorizing the Elementary and Secondary Education Act, put forth by U.S. Sens. Tom Harkin, D-Iowa, the chairman of the Senate education committee, and Sen. Michael B. Enzi of Wyoming, the committee’s lead Republican, was approved 15-7 by the panel on Oct. 20 after its introduction last week.
The bill has been controversial from the get-go, with civil rights groups criticizing its overhaul of the accountability system set up under the current law, the No Child Left Behind Act, and many states and administrator groups voicing concern that its provisions are still too restrictive.
In comparison, the bill’s research-related provisions seemed to be flying under the radar. They would:
• Greatly increase the percentage of federal program funds devoted to evaluation and technical assistance; from 0.5 percent to 1 percent in the case of federal Title I anti-poverty programs, and to 3 percent in most other programs.
• Set the Institute of Education Sciences, the U.S. Department of Education’s primary research arm, as the lead agency to evaluate federal education programs and require the IES to help federal programs establish criteria for program effectiveness.
• Establish a permanent grant for the Investing in Innovation, or i3, program, originally created under the fiscal stimulus law, which provides three- to five-year competitive grants to conduct research to develop and scale up promising education programs and interventions.
• Require more research evidence backing up the use of school improvement programs in some cases.
“I think they’ve got the fundamentals for a great [research] infrastructure in there,” said James W. Kohlmoos, the president of the Knowledge Alliance, a Washington-based group representing regional educational laboratories and other research organizations that get federal research funding. “There’s a recognition in total in this draft that we might be moving further down the road from ESEA being primarily an accountability bill to a solutions-oriented one that provides support for improvement.”
The bill wouldn’t allow the IES to pool all evaluation money into a single pot—a change long requested by IES officials and the National Board for Education Sciences that advises it—but the increase in evaluation set-asides could mean millions of additional dollars for the IES.
According to Sue Betka, the deputy director of administration and policy at the IES, the institute spent $66.8 million for evaluations in fiscal 2011, including: $11.1 million in general ESEA set-asides; $7.8 million for Title I evaluation specifically; $20.2 million for national program evaluations; and $11.5 million for special education evaluations. Only $16.2 million came from IES’ general research, development and dissemination funds.
The bill also would solidify the institute’s authority to evaluate Education Department programs, rather than splitting responsibility between the IES and the department’s various program offices. The department had already announced that starting this fiscal year, the IES’ National Center of Education Evaluation and Regional Assistance would conduct all program implementation and effectiveness studies of 18 months or longer, but the bill would increase that authority.
It calls for the IES to coordinate all federal program evaluations for the department and help establish the evaluation criteria for those programs on the front end.
“Generally I’m pleased the bipartisan bill shows confidence in IES and our ability to do this work,” said IES Director John Q. Easton. “We are very eager to work with program people from the very beginning to design programs that can result in strong evaluations with little burden on participants.”
In the past, evaluations have been planned after large grants such as Reading First went out, making it more difficult to set criteria and collect data to measure the program’s effectiveness.
“It’s awkward to have it divided
that way,” said Grover J. “Russ” Whitehurst, the director of the Brown Center on Education Policy at the Washington-based Brookings Institution and a former director of the IES. “It’s much better to have one captain of the ship rather than confused coordination of the bridge.”
He said “having a real secure pot of money, [and] knowing it’s going to be there over a number of years would change the nature of the enterprise. Predictability of the fund would allow much more forward thinking about an evaluation plan. It would allow the evaluation of lots of programs that don’t get evaluated now, simply because there’s not enough in smaller programs to fund a real evaluation.”
One example of the bill’s tougher requirements for research evidence comes in the area of school improvement. Under the bill, the bottom 5 percent of schools identified as “persistently low-achieving,” either for overall performance or for achievement gaps, would have to adopt one of a number of school improvement strategies. Among these, the “whole school reform” strategy allows districts to create their own turnaround plans—but only using programs and interventions that have demonstrated statistically significant improvements in student outcomes on “more than one well-designed or well-implemented experimental or quasi-experimental study.”
That provision raised red flags for Mr. Whitehurst. “I think it repeats the error in NCLB of using ‘scientifically based research’ as a requirement for using federal funds when there wasn’t a sufficient base of scientifically rigorous research to do what was required,” Mr. Whitehurst said.
While the Harkin-Enzi bill includes fewer references to “scientifically based research” than the NCLB law does, it sets a more stringent definition for programs backed by scientifically based research. They must include randomized, controlled experiments or quasi-experimental studies—considered the “gold standard” of education research—which have results that can be repeated and have been accepted by a peer-reviewed journal or panel of independent experts.
Since the NCLB law set randomized, controlled experiments as the bar for “scientifically based research,” the number of such trials has exploded. In the past five years, for example, regional educational laboratories conducted 25 studies using experimental designs. Yet critics have argued the time, expense, and ethical dilemmas posed by full experiments make them difficult to use for education.
The Baltimore-based Success for All Foundation offers one of the few whole-school interventions with evidence backed by experimental studies yet its chairman, Robert E. Slavin, admitted: “If we [as a field] had to [implement the school improvement model] this afternoon, it would be trouble; there wouldn’t be more than a half-dozen programs for whole-school reform that have that kind of evidence.” (Mr. Slavin also writes an opinion blog for edweek.org.)However, Fred Doolittle, the vice president and K-12 policy director of the New York City-based research firm MDRC, argued most schools could weave individual evidence-based interventions into their larger school improvement plan, and the requirement would drive demand for large-scale research on effective programs.
“We’re not starting from scratch; over the last decade, there has been a push for this kind of evidence in education, and as a result, the capacity to do this kind of work has expanded,” he said.
Jon Baron, the president of the Washington-based Coalition for Evidence-Based Policy, agreed.
“You don’t need a million programs; you need a few that are backed by strong evidence, and you’ll build more over time,” Mr. Baron said. “One of the reasons building additional valid evidence of things that work is so important is a lot of the interventions currently in use that people think are effective might not be achieving their intended goals.”
Among the experts interviewed, there also seemed to be a consensus that a permanent Investing in Innovation grant program could, more quickly, help expand the research-backed programs available. The program provides tiers of grants, with various funding levels, based in part on the level of evidence a program has already established.
“You encourage not just scaling up interventions backed by strong evidence, but these lower tiers are an opportunity for innovation, to see whether these programs are things that can work or not,” Mr. Baron said. “It’s attractive to much larger groups and constituencies.”
During the mark-up of the Harkin-Enzi bill last week, U.S. Sen. Michael F. Bennet, D-Colo., introduced—but later withdrew—an amendment to create a set-aside within the i3 program to develop the Advanced Research Projects Agency-Education, or ARPA-Ed, modeled on a cutting-edge research group within the U.S. Department of Defense.
ARPA-Ed which was originally proposed as a $90 million project in President Barack Obama’s 2012 budget plan, would focus on special projects “to aggressively pursue technological breakthroughs that transform educational technology and empower teaching and learning,” a statement from Sen. Bennet’s office says.
Mr. Slavin said he thought an ARPA-Ed addition could fill in gaps in i3. “When you want Disney or Pixar or National Geographic to get involved and solve some long-standing problems in education, I think [ARPA-Ed] could get them involved in a way that frankly a $5 million [i3] grant over five years isn’t going to do,” Mr. Slavin said.
Vol. 31, Issue 09, Pages 16,18
Get 10 free stories, e-newsletters, and more!
- Director of Technology
- St. Paul's School for Girls, Brooklandville, MD
- Director of Information Technology
- Charles E. Smith Jewish Day School, Rockville, MD
- Head of School
- Augusta Preparatory Day School, Martinez, GA
- Chief Academic Officer
- Cristo Rey Network, Chicago, IL
- Director of Auxiliary Programming
- Lovett School, Atlanta, GA