Assessment Governing Board Crafts Definition of 'Prepared for College'
Language covers reading, math for 12th grade NAEP
The governing board for the tests known as "the nation's report card" has marked its own definition of what makes a student academically prepared for college.
At a meeting here this month, members of the National Assessment Governing Board, which supervises the National Assessment of Educational Progress, voted 17-2 to adopt language that will define the new "college prepared" scores in reading and mathematics on the assessment.
The language will be used for reporting in the next 12th grade NAEP, whose results will be announced in spring 2014. Those results will include a nationally representative sample of seniors as well as state-level results for 13 states that volunteered to give the test to more students.
The final definition is fairly limited, with members voting to say that the percentages of students performing at or above 163 out of 300 in mathematics and 302 out of 500 in reading on the 12th grade assessments would be "plausible estimate[s] of the percentage of students who possess the knowledge, skills, and abilities [in those subjects] to make them academically prepared for college."
The two members who voted against the measure Aug. 3 were Andrés A. Alonso and Rebecca Gagnon.
Mr. Alonso, a professor of practice at the Harvard Graduate School of Education and a former chief executive of the Baltimore public schools, argued that the research was not strong enough to set particular cutoff scores for college preparation. By contrast, Ms. Gagnon, the director of the Minneapolis board of education, argued that NAGB should take a firmer stance. She said the cutoff scores are "reasonable" estimates rather than "plausible."
The new definitions are based on more than 30 studies, including several comparing the content and predictive value of the federally sponsored NAEP with those of college-placement assessments, such as the SAT, the ACT, and Accuplacer, as well as longitudinal studies in Florida of how students who performed at different levels on NAEP later fared in freshman-level college courses.
Researchers used the federal High School Transcript Study, a 2009 study linking outcomes between NAEP and the SAT, and a longitudinal study of Florida students to compare performance on NAEP in reading and math with the test scores considered "college-readiness benchmarks" in the act and the SAT in 2005 and 2009.
In both subjects, the researchers found students who met the "proficient" achievement level on NAEP—176 out of 300 in math and 302 out of 500 in reading—also scored at or above the college-readiness benchmark scores on the SAT and the ACT. In 2009, 38 percent of 12th graders scored at or above proficient in reading; only 26 percent reached proficiency in math.
Chester E. Finn Jr., who was the chairman of NAGB when the NAEP achievement levels were first approved, said at a symposium this summer in Washington that the "proficient" level was always intended to be "aspirational," while " 'basic' was supposed to show you were literate and could make your way through the subway system."
"Now, 23 years later, when college and career readiness is on everyone's lips, ... lo and behold, the pretty-clear conclusion reached is NAEP 'proficient' comes pretty darn close to college preparedness," said Mr. Finn, the president of the Thomas B. Fordham Institute, a Washington-based research group.
To get a more nuanced look at how students of different performance levels fared in college, the researchers tracked students by using Florida's K-20 student longitudinal system.
Based on the Florida data, students who earned at least a 298 out of 500 in reading or 162 out of 300 in math—reading nearly at the "proficient" level and math in the "basic" range—also at least met the ACT or SAT college-placement benchmarks, had a first-year college grade point average of at least 2.67, and were placed in nonremedial courses in math and literature.
In a parallel effort to set career-readiness benchmarks within NAEP, the governing board also had studied ways to connect NAEP to readiness for work as an automotive master mechanic, computer-support specialist, heating and air-conditioning technician, licensed practical nurse, and pharmacy technician, but it was not able to draw conclusions about how performance on NAEP would relate to such careers.
For example, among NAEP's math-framework objectives, 64 percent to 74 percent were "not evident as prerequisite" in any of the training required for the careers studied, a finding Cornelia Orr, the board's executive director, called "quite shocking."
"This was a very hard task, but it was very revealing," Ms. Orr said. "We found no evidence that someone prepared for job training is academically prepared for college. That said, someone prepared for college is certainly prepared for job training."
The governing board plans to conduct more linking studies between NAEP and the SAT and the ACT; longitudinal studies in Florida, Illinois, Massachusetts, Michigan, and Texas; and a linking study with the act's Explore, an assessment for 8th graders, in Kentucky and Tennessee.
However, Achieve, a Washington-based college-readiness advocacy group, wrote in a July 30 letter to NAGB that NAGB's college-preparedness benchmarks don't gauge how well students are prepared for college-level writing and questions in that subject released by NAEP "do not come close to assessing [the] skill set" involved in writing based on multiple sources.
NAGB has not officially responded to the letter, but there are also moves to develop more-detailed descriptions of the skills and NAEP questions that the "college-prepared" cutoff scores represent. "The one thing I've been concerned about from the very beginning of this research is its applicability to real life," said board member W. James Popham, a professor emeritus of education and information studies at the University of California, Los Angeles. "Real examples for real people would be useful."
Vol. 33, Issue 01, Page 9