Published Online: January 8, 2013
Published in Print: January 16, 2013, as Multiple Gauges Best for Teachers
Updated: January 18, 2013
Includes correction(s): January 18, 2013

Combined Measures Better at Gauging Teacher Effectiveness, Study Finds

Study probes scores, observations, surveys

Student feedback, test-score growth calculations, and observations of practice appear to pick up different but complementary information that, combined, can provide a balanced and accurate picture of teacher performance, according to research recently released from the Bill & Melinda Gates Foundation.

A composite measure on teacher effectivenessRequires Adobe Acrobat Reader drawing on all three of those measures, and tested through a random-assignment experiment, closely predicted how much a high-performing group of teachers would successfully boost their students’ standardized-test scores, concludes the series of new papers, part of the massive Measures of Effective Teaching study launched more than three years ago.

“If you select the right measures, you can provide teachers with an honest assessment of where they stand in their practice that, hopefully, will serve as the launching point for their development,” said Thomas J. Kane, a professor of education and economics at the Harvard Graduate School of Education, who headed the study.

Basing more than half a teacher’s evaluation on test-score-based measures of student achievement seemed to compromise it, the researchers also found.

Another piece suggests that teachers should be observed by more than one person to ensure that scores are reliable.

The findings, released Jan. 8, are among dozens from the final work products of MET. Together, they are billed as a proof point for the three measures the foundation has spent years studying.

Multiple Yardsticks

Researchers compared a number of different schemes for weighting the three indicators in an evaluation system. In general, more weight on“value added” made the systems more predictive of achievement growth on state tests, but less reliable. Results differed by grade and subjects; those depicted are for middle school teachers of english/language arts

Even as they praised the project’s other insights, some scholars debated the strength of the findings from the random experiment. One glitch: Teachers and administrators didn't always comply with the randomization, making it harder to interpret the results.

“We can only be certain that it’s a valid predictor of future test scores for those teachers who complied with the assignments,” said Jonah E. Rockoff, an associate professor of finance and economics at Columbia Business School, who has studied teacher-quality issues using economic techniques. Mr. Rockoff was not involved in the study, but reviewed early drafts of the findings.

Taken as a whole, the final MET findings provide much food for thought about how teacher evaluations might best be structured. But they are not likely to end a contentious, noisy debate about evaluation systems, and they are almost certain to be intensely scrutinized, in part because of Gates’ separate support for advocacy organizations that have already staked out positions on teacher evaluations.

(The Gates Foundation also provides support for coverage of business and innovation in Education Week.)

Weighing Measures

The $45 million study, in progress since 2009, is one of the largest and most extensive research projects ever undertaken on the question of how to identify and measure high-quality teaching. It involved some 3,000 teachers in six districts: Charlotte-Mecklenberg, N.C.; Dallas; Denver; Hillsborough County, Fla.; Memphis, Tenn.; and New York City.

Earlier studies released by the MET project had examined three potential measures of teacher quality: observations of teachers keyed to teaching frameworks, surveys of students’ perceptions of their teachers, and a value-added method, which attempts to isolate teachers’ contributions to their students’ academic achievement. Researchers examined the relationship of each measure to students’ scores on state standardized tests as well as on a more complex, project-based series of tasks; and to students' feelings of effort and engagement in class.

Each of those measures, the earlier papers stated, had tradeoffs in terms of their reliability and their correlation to the academic and nonacademic outcomes.

One of the four new papers released by the Gates Foundation goes the next step: It examines different ways of weighting those three measures.

It found that those that relied the most heavily on state standardized-test scores appeared to be counterproductive. Those composites tended to be volatile and were also the least predictive of how students taught by those teachers would fare on the more cognitively demanding tasks.

Yet weighting schemes that put the most emphasis on teacher observations were the least predictive of gains on the state test scores, it says.

In all, the study indicates, those that use a more equal mix of components, including between a third and half based on value-added, couple better correlations to the outcome measures with improved reliability.

In a way, the findings indicate that there is no one “best” way to weight the measures; instead, that decision will depend on what policymakers most value, whether state test scores or other outcomes.

Randomized Experiment

From the beginning, one of the foundation’s key goals was to subject promising measures to “validation” through a randomized experiment.

Though infrequently conducted in K-12 education because of logistical problems and expense, random assignment allows researchers to eliminate sources of bias, such as the sorting of students into particular classes, not accounted for using traditional statistical techniques.

The Gates project, with its reach across six districts and thousands of teachers, offered an unusual chance to test the ideas at a scale not seen previously.

For the randomization, researchers in 2009-10 generated estimates of teachers’ performance based on composite measures using data from the surveys, prior test scores, and observation scores. Within individual schools, the study randomly assigned a class of students to each of the participating teachers in particular grades and subjects. After a year, then, researchers compared those teachers’ actual performance to the estimates.

The results were examined in groups based on the teachers’ predicted performance.

Related Blog

In general, the groups of teachers identified as being more effective did, in fact, help the assigned classes of students learn more, producing results on par with what the measures had predicted. They also improved student performance not just on traditional standardized tests but also on the deeper, project-based tasks.

“Because of the random assignment, we can be confident that we identified a subgroup of teachers who caused achievement to happen,” regardless of student characteristics, Harvard’s Mr. Kane said. “It’s sort of a big deal to be able to say that.”

Student attrition and other factors, including the refusal of several schools to carry out the randomization despite agreeing to do so, led to relatively high rates of noncompliance. About 66 percent of students in Dallas stayed with their assigned teacher, but only 27 percent of students in Memphis did.

To account for the noncompliance, researchers used a statistical technique known as “instrumental variables” to adjust the results. The technique is widely used in the social sciences.

Scholars had different opinions about how far the findings could be extrapolated.

“These results could still be based on a very selective group of teachers,” said Jesse M. Rothstein, an assistant professor of economics at the University of California, Berkeley, who has often been critical of the MET findings. “I would love to see a lot more investigating of just who was and wasn’t complying, and why they were left out.”

Douglas N. Harris, a professor of economics at Tulane University, in New Orleans, added that the study didn’t address some other potential sources of bias. For example, it’s possible that bias in the value-added estimates for each individual teacher might have been averaged out in the group estimates. (The averaging was done in order to obtain a sufficient sample size, a limit of the random-assignment method.) But most school districts and states using value-added approaches are using individual, not group-level results, he noted.

The study’s authors also acknowledge that the experiment is limited to comparisons of teachers within, but not across, schools.

“There are a lot of ways in which there could be a nonrandom assignment of students to teachers,” Mr. Harris said. “They’re studying some elements of that, but not others.”

Teacher Observations

In yet another new finding, the researchers dug deeper into observations of teachers. They examined lessons from a subset of 67 teachersin the Hillsborough, Fla., district, investigating ways to improve the scoring of those lessons.

The researchers found that having different raters score observations of teachers’ practice may be a key component for the observations systems: Raters’ first perception of a teacher's practice tended to influence how they scored additional lessons taught by that same teacher, the study found.

Nearly all teachers scored in the middle categories on the framework studied, the four-tiered Framework for Teaching, a popular tool created in 1996 by consultant Charlotte Danielson, rather than at the top or bottom ones. The researchers struggled to interpret that finding.

“It could be that observers are simply uncomfortable making absolute distinctions between teachers,” that paper says. “It could be that the performance-level standards need to make finer distinctions. Or it could simply be that underlying practice on the existing scales does not vary that much.”

Mixed Reception?

Nearly every work product released by the MET researchers thus far has been contested to some degree by observers, and the most recent results are likely to be no exception.

“They see this as proof that the more equally weighted, combined measure is superior, but they omit all discussion of the expense and difficulty of collecting the classroom observations and student surveys,” said Jay P. Greene, a professor of education policy at the University of Arkansas. Mr. Greene contends that earlier reports from Gates have veered too far into advocacy.

By contrast, the American Federation of Teachers, whose leader has had an on-again-off-again rapport with Mr. Gates and with the MET project, embraced the final studies.

“The MET findings reinforce the importance of evaluating teachers based on a balance of multiple measures of teaching effectiveness, in contrast to the limitations of focusing on student test scores, value-added scores, or any other single measure,” AFT President Randi Weingarten said in a statement.

Vol. 32, Issue 17, Pages 1,16

You must be logged in to leave a comment. Login | Register
Ground Rules for Posting
We encourage lively debate, but please be respectful of others. Profanity and personal attacks are prohibited. By commenting, you are agreeing to abide by our user agreement.
All comments are public.

Back to Top Back to Top

Correction: 
This story has been updated with the current title for Jay P. Greene, a professor of education policy at the University of Arkansas.

Most Popular Stories

Viewed

Emailed

Recommended

Commented