Top researchers studying new "value added" or "growth index" models for measuring a teacher's contribution to student achievement completely agree on one thing: These methods should be used in staff-evaluation systems with more caution than they have been so far.
That area of agreement emerged in an Aug. 9 meeting that drew together a who's who of a dozen of the nation's top education researchers on value-added methods--in areas from education to economics--to build, if not consensus, at least familiarity within a disparate research community for value-added systems. The U.S. Department of Education's research agency, which organized the forum, today released the proceedings of the meeting, as well as individual briefs from each of the experts.
"There's been a huge amount of research in this field in recent years, but it tends to be really siloed," John Q. Easton, the director of the Institute of Education Sciences, told members of the National Board for Education Sciences, IES's advisory group, during a briefing earlier this month. "People don't seem to read each other's work, and it's published in totally different journals. It was so typical to read somebody's study who was not citing all the others."
Pros and Cons
Value-added methods, which attempt to measure teachers' performance based on their students' test scores, have gained support in the last decade, as studies by Stanford University economist Eric A. Hanushek and others found inconclusive evidence to support a link between a teacher's effectiveness and his or her degree credentials--the latter of which is the traditional basis for teacher pay. Massive federal support, in the form of the $290 million Teacher Incentive Fund and the $4 billion Race to the Top competition has led to rapid growth in the number of states and districts adopting these methods in their teacher evaluation systems.
Advocates argue that value-added methods can be more objective than principal observations alone, and if done well can provide information about areas a teacher needs to beef up instruction. Critics contend these scores can only be used for teachers of mathematics and English/language arts in tested grades, leaving out both a large proportion of district teachers and any contribution a teacher makes to untested subjects or skills, be they science or self-control.
An influential series of studies by Jesse Rothstein, a public policy and economics professor at the University of California, Berkeley, and a participant in the meeting, found a standard value-added model was biased because it did not take into account that parents and principals often push teachers to take certain students, rather than assigning them at random.
"[Value-added measures] will deteriorate--will become less reliable and less closely tied to true effectiveness--if they are used for high-stakes individual decisions," Mr. Rothstein wrote in a brief for the meeting. "How much will teachers change their content coverage, neglect nontested subjects and topics, lobby for the right students, teach test-taking strategies, and cheat outright? ... We simply don't know."
Tools for Improvement
The Measures of Effective Teaching Project, funded by the Seattle-based Bill and Melinda Gates Foundation, is expected to release a report later this year in which class rosters were randomly assigned to clusters of teachers by school, grade, and subject area. This may help identify how the selection bias Mr. Rothstein found takes place and can be prevented, according to Thomas Kane, an education and economics professor at the Harvard Graduate School of Education, and a meeting participant.
Mr. Kane and fellow Harvard assistant education professor Andrew Ho, contended that district leaders should focus less on using value-added systems to rank teachers, which Mr. Ho likened to hospital intake questionnaires that identify initial symptoms. "Medicine (and education) is not only about symptoms (and even less so about one-dimensional rankings of symptoms), but, far more critically, diagnosis and ultimately treatment," Mr. Ho said. "How can we use VAM results to improve teaching and the teacher corps?"
Education officials' tendency to average multiple measures or years of data into a single composite score worried many researchers.
From one year to the next, a teacher's ratings under some of the value-added systems now in use can vary by 4 percent to 25 percent, according to Linda Darling-Hammond, an education professor and faculty co-director of the Stanford Center for Opportunity Policy in Education at Stanford University in Palo Alto, Calif. She argued that researchers and policymakers must take into account the range of scores available on their state's tests when developing a value-added system. For example, a teacher of gifted students may not show up as very effective, because his or her students are already performing near the top of the test's ability to measure their progress.
Mr. Kane countered that teachers have such a strong effect on student achievement that if value-added measures help identify teachers in the bottom 5 percent of performance and bring them up to the district average, they can lead to an average increase in lifetime earnings for each student of $52,000 as a result of being taught by that teacher for one year.
Many of the experts see both promise and peril in the rollout of the Common Core State Standards and their effect on existing and emerging teacher evaluation systems.
In most districts, researchers voiced concern that evaluation systems do not take into account the time it will take for even the most effective teachers to adapt to new areas of focus in the standards--not to mention that the common core deliberately omits guidance on specific teaching strategies to meet the new requirements.
For example, Henry Braun, the director of the Center for the Study of Testing, Evaluation, and Education Policy at Boston College and a consultant with the Partnership for Assessment of Readiness for College and Careers, or PARCC, one of the two consortia developing tests for the common core, has been struggling with how to design an assessment which likely will end up being used for teacher evaluation. He worried that if the teacher accountability "tail" wags the student assessment "dog," tests won't be designed appropriately to measure students' learning rather than teacher behavior.
Experts called for state policy leaders to consider how their individual state tests will affect the validity of individual districts' evaluation systems.
Most Popular Stories
- Boehner Lashes Out Against Ted Cruz, Far Right
- TFA Recruiting DACA Recipients
- Hawaii Official Who Release Obama Certificate Only Victim of Plane Crash
- Holiday Shopping Off to a Slow Start This Season
- Ford Plans New Cars, Jobs in 2014
- Gold, Silver Slide on Prospects of Fed Exit
- 'Rape Insurance' Bill Passes in Michigan
- Ted Cruz Coloring Book Selling Briskly
- Kim Jong Un's Uncle Executed
- Grizzly Bears Could Be Taken Off Endangered List