Assessment Governing Board Defines 'College Prepared'
The governing board for the tests known as “the nation’s report card” has marked its own definition of what makes a student academically prepared for college.
At a meeting here this month, members of the National Assessment Governing Board, which supervises the National Assessment of Educational Progress, voted 17-2 to adopt language that will define the new “college prepared” scores in reading and mathematics on the assessment.
The language will be used for reporting in the next 12th grade NAEP, whose results will be announced in 2014. Those results will include a nationally representative sample of seniors as well as state-level results for 13 states that volunteered to give the test to more students.
The final definition is fairly limited, with members voting to say that the percentages of students performing at or above 163 out of 300 in mathematics and 302 out of 500 in reading on the 12th grade assessments would be “plausible estimate[s] of the percentage of students who possess the knowledge, skills, and abilities [in those subjects] to make them academically prepared for college.”
The two members who voted against the measure Aug. 3 were Andrés A. Alonso and Rebecca Gagnon.
Mr. Alonso, a professor of practice at the Harvard Graduate School of Education and a former chief executive of the Baltimore public schools, argued that the research was not strong enough to set particular cutoff scores for college preparation. By contrast, Ms. Gagnon, the director of the Minneapolis board of education, argued that NAGB should take a firmer stance. She said the cutoff scores are “reasonable” estimates rather than “plausible.”
Locating the Levels
The new definitions are based on more than 30 studies, including several comparing the content and predictive value of the federally sponsored NAEP with those of college-placement assessments, such as the SAT, the ACT, and Accuplacer, as well as longitudinal studies in Florida of how students who performed at different levels on NAEP later fared in freshman-level college courses.
Researchers used the federal High School Transcript Study, a 2009 study linking outcomes between NAEP and the SAT, and a longitudinal study of Florida students to compare performance on NAEP in reading and math with the test scores considered “college-readiness benchmarks” in the ACT and the SAT in 2005 and 2009.
In both subjects, the researchers found students who met the “proficient” achievement level on NAEP—176 out of 300 in math and 302 out of 500 in reading—also scored at or above the college-readiness benchmark scores on the SAT and the ACT. In 2009, 38 percent of 12th graders scored at or above proficient in reading; only 26 percent reached proficiency in math.
Chester E. Finn Jr., who was the chairman of NAGB when the NAEP achievement levels were first approved, said at a symposium this summer in Washington that the “proficient” level was always intended to be “aspirational,” while “ ‘basic’ was supposed to show you were literate and could make your way through the subway system.”
“Now, 23 years later, when college and career readiness is on everyone’s lips, ... lo and behold, the pretty-clear conclusion reached is NAEP ‘proficient’ comes pretty darn close to college preparedness,” said Mr. Finn, now the president of the Thomas B. Fordham Institute, a Washington-based research group.
To get a more nuanced look at how students of different performance levels fared in college, the researchers tracked students by using Florida’s K-20 student longitudinal system.
Based on the Florida data, students who earned at least a 162 out of 300 in math or 298 out of 500 in reading—both scores in the upper half of the “basic” achievement range—also at least met the ACT or SAT college-placement benchmarks, had a first-year college grade point average of at least 2.67, and were placed in nonremedial courses in math and literature.
In a parallel effort to set career-readiness benchmarks within NAEP, the governing board also had studied ways to connect NAEP to readiness for work as an automotive master mechanic, computer support specialist, heating and air conditioning technician, licensed practical nurse, and pharmacy technician, but it was not able to draw conclusions about how performance on NAEP would relate to such careers.
For example, among NAEP’s math-framework objectives, 64 percent to 74 percent were “not evident as prerequisite” in any of the training required for the careers studied, a finding Cornelia Orr, the board’s executive director, called “quite shocking.”
“This was a very hard task, but it was very revealing,” Ms. Orr said. “We found no evidence that someone prepared for job training is academically prepared for college. That said, someone prepared for college is certainly prepared for job training.”
The governing board next plans to conduct additional linking studies between NAEP and the SAT and the ACT, as well as longitudinal studies in Florida, Illinois, Massachusetts, Michigan, and Texas, and a linking study with the ACT’s Explore, an earlier assessment for 8th graders, in Kentucky and Tennessee.
“Since NAEP is the gold-standard assessment, it will be enormously helpful to have NAEP benchmarks for academic preparedness for college as both assessment consortia [for the Common Core State Standards], as well as individual states, work to develop or purchase measures of college readiness,” Michael Cohen, the president of Achieve, a Washington-based college-readiness advocacy group, wrote in a July 30 letter to NAGB Chairman David P. Driscoll.
However, Achieve expressed concern that NAGB’s college-preparedness benchmarks don’t gauge how well students are prepared for college-level writing—and, moreover, that questions in that subject released by NAEP “do not come close to assessing [the] skill set” involved in writing from multiple sources.
NAGB has not officially responded to the letter, but there are also moves to develop more-detailed description of the skills and NAEP questions that the “college-prepared” cutoff scores represent.
“The one thing I’ve been concerned about from the very beginning of this research is its applicability to real life,” said board member W. James Popham, a professor emeritus of education and information studies at the University of California, Los Angeles. “Real examples for real people would be useful.”
The 2009 NAEP gives a few such examples. Someone with a score of 295 in reading (just below proficient) would be expected to draw details from across a story to explain a character’s motivation, but might not be able to provide and explain information from an article (associated with a score of 305, in the proficient range).
In math, a student earning a score of 163 would be expected to solve a nonequivalent expression in algebra.
Vol. 33, Issue 1
Get more stories and free e-newsletters!
- Modern & Classical Languages Department Chair
- New Trier Township High School District 203, Winnetka, IL
- Superintendent of Schools
- Fremont County School District #14, Ethete, WY
- Dean of Students
- Diman Regional Vocational-Technical High School, Fall River, MA
- Chief Information Officer and Special Projects Manager
- Randolph Public Schools, MA
- Glenbrook North High School, Glenview, IL