NCLB Technical Panel Could Be Influential

As reauthorization stalls, Ed. Dept. committee to advise on accountability.

Article Tools
  • PrintPrinter-Friendly
  • EmailEmail Article
  • ReprintReprints
  • CommentsComments

With Congress having effectively postponed the reauthorization of the No Child Left Behind Act, the Department of Education is turning to a panel of experts for advice about the technical issues that linger nearly seven years after the law’s passage.

Lawmakers will wait until at least next year to decide how to amend the federal education law, and that leaves in place the current requirements for standards, testing, and accountability.

Until Congress acts, though, the new advisory committee will help outline regulatory adjustments that the Education Department can make so the law’s accountability system is fairer and easier for school districts to implement.

“We think you’re going to be a good resource for the department,” Raymond J. Simon, the deputy secretary of education, said on Sept. 16 at the first meeting of the National Technical Advisory Panel. “You’re going to be a good resource for Congress.”

And Mr. Simon said the 16-member body will help answer important questions for the next presidential administration about what powers it has to effectively alter how the law works, and what changes must be adopted by Congress in a reauthorization.

Policy Implications

No Child Left Behind Act Technical Panel

Secretary of Education Margaret Spellings has appointed experts in testing and other policies to be members of the National Technical Advisory Committee:

CHAIRMAN: Tom Fisher, a consultant based in McMinnville, Tenn., and a former director of Florida’s testing program


David Abrams, assistant commissioner for standards, assessment, and reporting, New York state department of education

Anthony Alpert, director of assessment, Oregon department of education

Diane Browder, professor of special education, University of North Carolina at Charlotte

Wesley Bruce, assistant superintendent for the Center for Accreditation, Assessment, and Licensing, Indiana department of education

Wayne Camara, vice president for research and development, the College Board, New York City

Kevin Carey, research and policy manager, Education Sector, Washington

Gregory Cizek, professor of educational measurement and evaluation, University of North Carolina at Chapel Hill

Carl Cohn, former superintendent of San Diego and Long Beach, Calif., school districts

Denise Collier, chief academic officer, Dallas school district

Robert Costrell, professor of education reform and economics, University of Arkansas at Fayetteville

Harold Doran, principal research analyst, American Institutes for Research, Washington

Margo Gottlieb, lead developer, World-Class Instructional Design and Assessment Consortium, University of Wisconsin-Madison

Suzanne Lane, professor of research methodology, University of Pittsburgh

Scott Marion, vice president, National Center for the Improvement of Educational Assessment Inc., Dover, N.H.

John Poggio, professor, department of educational psychology and research, University of Kansas, Lawrence

Secretary of Education Margaret Spellings appointed the panel in August to advise the department on standards, assessments, and accountability. The group includes state assessment directors, test designers, and other policy experts.

At its first meeting here, department officials filled the agenda with such questions as how to design growth models and create indexes that provide separate cumulative scores to report a school’s achievement in reading and mathematics.

The discussion of issues surrounding such accountability topics quickly became technical among the experts. But answering such questions can lead to important changes in policy in determining whether schools make adequate yearly progress—or AYP—under the law. Schools that fail to make AYP may have to pay for tutoring for their students and take other actions to improve achievement.

For example, the background paper for the meeting asked whether the Education Department should change the questions peer reviewers ask when evaluating states’ latest applications to use growth models.

The alternative way of measuring AYP can be a more accurate measure of schools’ success because it examines students’ growth over time, rather than comparing students at one grade level to the previous year’s cohort.

But the criteria used to evaluate states’ applications for growth models could make the accountability system “so flexible as to make AYP less meaningful,” the background paper said.

“In short, is it possible for an accountability system to provide too many ways to make AYP?” the document asked the panel members to address.

Secretary Spellings has approved 11 states to use growth models in making AYP decisions and has invited remaining states to submit growth models for potential approval for implementation as soon as the current school year.

Indexing Performance

At the panel’s Sept. 16 meeting, it was clear that the work would be technical.

The first item discussed was what criteria the department should use to evaluate a state’s performance index­—a way of combining a variety of results to determine the success of a school. Most indexes give states partial credit for students who achieve at levels below the NCLB law’s goal of proficiency. So far, the department hasn’t approved a plan that gives a school extra credit for the number of students who achieve at a level above proficiency.

The department has approved 12 states’ performance indexes and would consider proposals from other states to make their AYP determinations based on such indexes, said Kerri Briggs, the assistant secretary of elementary and secondary education.

But, Ms. Briggs added, the department would not approve plans that don’t maintain key elements of the law, such as measuring whether students are on track to meet the goal of universal proficiency in reading and mathematics by the end of the 2013-14 school year.

Testing experts turned the discussion to issues of validity and the reliability of test scores.

They also asked whether the department allows states to design indexes to report the results on the alternate assessments given to students with disabilities. So far, Ms. Briggs said, the department hasn’t done so.

“If you’re going to allow index systems for all students, I can’t see a reason why you couldn’t apply the same index system to the alternate assessment,” said Scott Marion, a panel member and a vice president of the National Center for the Improvement of Educational Assessment, a nonprofit consultant to states based in Dover, N.H.

Vol. 28, Issue 05, Pages 23-24

Published in Print: September 24, 2008, as NCLB Technical Panel Could Be Influential
Notice: We recently upgraded our comments. (Learn more here.) If you are logged in as a subscriber or registered user and already have a Display Name on, you can post comments. If you do not already have a Display Name, please create one here.
Ground Rules for Posting
We encourage lively debate, but please be respectful of others. Profanity and personal attacks are prohibited. By commenting, you are agreeing to abide by our user agreement.
All comments are public.

Back to Top Back to Top

Most Popular Stories