Test Firm, N.Y.C. Officials Say Scores Were ‘Overstated’

By Catherine Gewertz — June 20, 2001 3 min read
  • Save to favorites
  • Print

The New York City board of education and CTB/McGraw-Hill, one of the nation’s largest test-makers, said that more than 60,000 6th graders in the city received higher scores on a reading test last year than they should have. But each side offered differing theories on how it happened.

In a joint statement issued June 8, the city’s schools chancellor, Harold O. Levy, and the president of CTB/McGraw-Hill, David M. Taggart, said the results of the April 2000 reading tests that the city and the test-maker designed together were “overstated.”

The problems stemmed from “a combination of factors related to test design” and had no effect on the accuracy of students’ answers, they said. In an interview later, Mr. Taggart speculated about a variety of factors that could have worked together to create an overly high aggregate result.

Students simply might have delivered an exceptionally strong performance, he said. Some students might have been familiar with some of the questions because 12 of the 50 multiple-choice questions had been used in a 1999 version of the test, he said. In addition, he suggested, the results could have been affected if the administration of the test was inconsistent from one school to another.

“Were those scores wrong? No, they were not wrong,” Mr. Taggart said. “What may have happened is that those scores may overstate to some extent the level of skills those students had. But it’s important to remember those students were improving from 1999 to 2000. There’s no question about that. The question is how much.”

The executive director of assessment and accountability for the 1.1 million-student New York school system had a different view. Robert Tobias theorized that the group’s scores might have been inflated because the test was deemed to be more difficult than it actually was, allowing students to receive more points for answering fewer of the questions correctly.

On the 2000 test, he said, students were judged to meet grade-level standards if they answered 70 percent of the questions correctly. On the 2001 test, students needed to answer 80 percent to achieve the same benchmark.

Mr. Tobias disagreed that familiarity with questions used in a previous year contributed to the problem, noting that students scored about the same on the 12 repeated questions as they did on the new ones.

Warning Signs

Mr. Tobias and CTB/McGraw-Hill officials became suspicious right away, when the 2000 test results showed that 15 percent more students had achieved grade-level standards than had done so in 1999—a gain of more than double what is normally expected, Mr. Tobias said. But the test-maker stood by the results, and the district agreed to wait for the next year’s scores to draw conclusions.

But when the 2001 results showed a 13 percent drop in the number of students meeting grade-level standards, that finding bolstered officials’ suspicions that the 2000 scores were too high. In 1999, 28.6 percent of 6th graders had met the standard; the next year, the figure jumped to 43.9 percent; and this year, it dropped to 30.9 percent—a spiky performance line that “really got our attention,” Mr. Tobias said.

The problem brought back memories of a foul-up with another CTB/McGraw-Hill test in 1999, when 9,000 students were sent to summer school because of an incorrectly scored reading test.

Mr. Tobias said he was confident that no students were promoted improperly as a result of the 2000 glitch because test scores are considered in combination with attendance and classroom performance in making such decisions.

Mr. Taggart and Chancellor Levy said in their statement that the 2001 results were accurate because the questions raised by the 2000 test were addressed in this year’s version. Mr. Tobias and Mr. Taggart explained that the 2001 test contained all new items and greatly increased the portion of questions that have been tested on a national sample of students.

Mr. Tobias said that CTB/McGraw-Hill had done a good job on many of the city’s tests, but that the city plans to seek out other makers of reading assessments. Those plans were made before the current problem surfaced. But in issuing a new contract, Mr. Tobias said, “part of the consideration is the track record of the company.”

“Obviously,” he said, “any anomaly that occurs, particularly those that aren’t explained, shakes your credibility in the products and services you are receiving.”

Related Tags:

A version of this article appeared in the June 20, 2001 edition of Education Week as Test Firm, N.Y.C. Officials Say Scores Were ‘Overstated’


Special Education Webinar Reading, Dyslexia, and Equity: Best Practices for Addressing a Threefold Challenge
Learn about proven strategies for instruction and intervention that support students with dyslexia.
This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of Education Week's editorial staff.
Personalized Learning Webinar
No Time to Waste: Individualized Instruction Will Drive Change
Targeted support and intervention can boost student achievement. Join us to explore tutoring’s role in accelerating the turnaround. 
Content provided by Varsity Tutors for Schools
Student Well-Being K-12 Essentials Forum Social-Emotional Learning: Making It Meaningful
Join us for this event with educators and experts on the damage the pandemic did to academic and social and emotional well-being.

EdWeek Top School Jobs

Teacher Jobs
Search over ten thousand teaching jobs nationwide — elementary, middle, high school and more.
View Jobs
Principal Jobs
Find hundreds of jobs for principals, assistant principals, and other school leadership roles.
View Jobs
Administrator Jobs
Over a thousand district-level jobs: superintendents, directors, more.
View Jobs
Support Staff Jobs
Search thousands of jobs, from paraprofessionals to counselors and more.
View Jobs

Read Next

Assessment Opinion Rebooting Assessment and Accountability Post-Pandemic: What Now?
The disruptions of the pandemic have made this an ideal time to rethink accountability and assessment.
3 min read
Image shows a multi-tailed arrow hitting the bullseye of a target.
DigitalVision Vectors/Getty
Assessment Don’t Use State Tests ‘Punitively,’ Ed. Secretary Cardona Warns
As federal accountability restarts after two years, guidance from the department underscores how complicated that could be.
5 min read
Image of data, target goals, and gaining ground.
Assessment Latest Round of Federal Grants Aims to Make States' Assessments More Equitable, Precise
The U.S. Department of Education awarded over $29 million in competitive grants to 10 state education agencies.
2 min read
Assessment review data 599911460
Assessment Opinion Are There Better Ways Than Standardized Tests to Assess Students? Educators Think So
Student portfolios and school community surveys are but two of the many alternatives to standardized tests.
3 min read
Illustration of students in virus environment facing wave of test sheets.
Collage by Vanessa Solis/Education Week (Images: iStock/DigitalVision Vectors/Getty)