First, they would have to embrace the comprehensive
use of test score growth data (through Value - Added Measurement)-- and ultimately, the standardized tests they loathe — in evaluating districts, teachers, and school leaders.
Because state legislators, at the behest of the National Education Association's affiliate there, refused to pass a law back in February allowing
the use of test score growth data in teacher evaluations.
Not exact matches
Later that same day, Gov. Andrew Cuomo's Common Core task force released its recommendations, including a four - year moratorium on the
use of state - provided
growth scores based on state
tests in evaluations.
Using student - level data from two states, Harvard Professor Martin West and I found that 40 to 60 percent
of schools serving mostly low - income or underrepresented minority students would fall into the bottom 15 percent
of schools statewide based on their average
test scores, but only 15 to 25 percent
of these same schools would be classified as low performing based on their
test -
score growth.
After extensive research on teacher evaluation procedures, the Measures
of Effective Teaching Project mentions three different measures to provide teachers with feedback for
growth: (1) classroom observations by peer - colleagues
using validated scales such as the Framework for Teaching or the Classroom Assessment
Scoring System, further described in Gathering Feedback for Teaching (PDF) and Learning About Teaching (PDF), (2) student evaluations
using the Tripod survey developed by Ron Ferguson from Harvard, which measures students» perceptions
of teachers» ability to care, control, clarify, challenge, captivate, confer, and consolidate, and (3)
growth in student learning based on standardized
test scores over multiple years.
The most sophisticated approach
uses a statistical technique known as a value - added model, which attempts to filter out sources
of bias in the
test -
score growth so as to arrive at an estimate
of how much each teacher contributed to student learning.
Fortunately, we have a recent study that examined whether the criteria
used by regulators in New Orleans are predictive
of test score growth — even if we accept
test gains as a reliable indicator
of quality.
They include nearly 1.6 million
test -
score growth records for students (where a
growth record consists
of a linked current and prior
score) covering the five - year time span from 2007 to 2011 (2006
scores are
used as prior - year
scores for the 2007 cohort).
So, he asks «whether regulators are any good at identifying which schools will contribute to
test score gains» and then says this: «The bottom line is that none
of the factors
used by authorizers to open or renew charter schools in New Orleans were predictive
of how much
test score growth these schools could produce later on.»
To be eligible for that program, states had to adopt Common Core (or similarly rigorous standards and assessments), and they had to put into place teacher evaluation systems that
use student
test score growth as a «significant» part
of both teacher and school principal evaluations.
As explained in a guest blog this year by by FairTest's Lisa Guisbond, these measures
use student standardized
test scores to track the
growth of individual students as they progress through the grades and see how much «value» a teacher has added.
Value - added measures
use test scores to track the
growth of individual students as they progress through the grades and see how much «value» a teacher has added.
The most - positive aspect
of Kline's plan lies with its requirement that states develop teacher evaluation systems that
use student
test score growth data (along with other «multiple measures) in evaluating teacher performance.
Districts can
use an Indiana Department
of Education - approved evaluation system or design their own, but all schools must include student
growth data — think
test scores — as part
of a teacher's rating.
What reformers should do is develop the tools that can allow families to make school overhauls successful; this includes building comprehensive school data systems that can be
used in measuring success, and continuing to advance teacher quality reforms (including comprehensive teacher and principal evaluations based mostly on value - added analysis
of student
test score growth data, a subject
of this week's Dropout Nation Podcast) that can allow school operators
of all types to select high - quality talents.
You wrote, «We should not destroy our schools to create a bell curve
of accountability performance, which is created when we compare teachers to each other
using student
test score growth.»
In a nutshell, she points out that the MET study asked whether actual observation
of teaching, student surveys, or VAM
test score measures did a better job
of predicting future student
test score growth, which «privileges»
test scores by
using it both as a variable being
tested and as the outcome reflecting gains.
Principals are also evaluated based on PARCC data: Principals
of schools with any grade from 4 - 8 taking the PARCC
tests will also have a median student
growth score used as a 10 percent weight in their evaluations.
In addition, Hespe said the state will add an appeal process for the current year around the
use of so - called «student
growth objectives,» a separate measure that
uses assessments other than standardized
test scores.
Some schools thought
of as high or low performers in the past based on
test scores could have ratings that show the opposite because
of other factors being
used in the ratings, including
test score growth over time, readiness for graduation and progress on closing achievement gaps between student groups.
While the Department will likely add more academic performance measures in the future, for 2014 officials also included the level
of participation in state assessments, achievement gaps between students with disabilities and the general population as well as
scores on the National Assessment
of Educational Progress, a standardized
test used to gauge academic
growth across the country.
Positions long held by MORE, like strenuous opposition to high stakes
testing and the
use of VAM
growth scores to evaluate teachers, were until very recently considered by the power structure to be extreme.
With
testing increasingly in the news — from the release
of state
test scores to the federal Department
of Education's recent announcement in support
of short delays for teacher evaluation
using assessment
growth — now is the time set a new course.
According to the report, «value - added models» refer to a variety
of sophisticated statistical techniques that measure student
growth and
use one or more years
of prior student
test scores, as well as other background data, to adjust for pre-existing differences among students when calculating contributions to student
test performance.
When Adair County Elementary School (ACES) staff members were brainstorming ways to get their students excited about
using Study Island to prepare for the NWEA MAP
test in math, and ultimately increase
growth scores, they decided to tap into the spirit
of competition and hold their own creative Math Bowl contest.
To accommodate these requirements, state departments
of education commonly
use state
test scores to calculate measures
of student learning, which we refer to as
growth scores or value - added measures.
Governor Brown and the state board were correct in balking at the requirement to
use test scores as a measure
of student
growth in appraising teachers.
This detailed information about student academic
growth should be
used instead
of AGT
scores or any other measurements based on a single
test, as teachers and administrators seek to
use data to inform best practices that will improve student achievement;» [emphasis ours]
One new link is to a video featuring Ritz speaking into the camera about evaluation and dropping another bombshell — that her staff plans to revise Bennett - created rules that would have assigned teachers ratings
of 1 through 4 based on the ISTEP
test score growth of their students that districts could
use as part
of their evaluations.
This is an important question because it appears that the Obama administration is essentially allowing any evaluation system to gain its blessing long as it has unspecified
use of longitudinal student
test score growth data as one
of the main components.
The deal specifically prohibits the
use of individual teacher
growth scores other than to «give perspective and to assist in reviewing the past CST (California Standards
Test) results
of the teacher.»
Brown and the State Board balked at the stipulation that the state require districts to
use standardized
test scores as a measure
of student academic
growth when evaluating teachers.
And considering the low - quality
of subjective classroom observations that are the norm for traditional teacher evaluation systems, the state laws and collective bargaining agreements governing teacher performance management discourage school leaders from providing more - ample feedback, and that the
use of objective student
test score growth data is just coming into play, few teachers have gotten the kind
of feedback needed to build such expertise in the first place.
After all, Obama has until recently made a compelling case thanks to efforts such as Race to the Top, which has managed to force states to lift or modify restrictions on the expansion
of charter schools, and allow for the
use of student
test score growth data in teacher evaluations.
Because current state policy requires that refused
tests be given the lowest possible
score, the
scores of 1 given to refused
tests are calculated into the
growth rates
used to evaluate individual teachers.
For example, in order to address concerns about the fairness
of using student
test scores to evaluate teachers, Hillsborough County Public Schools, in Tampa, Florida, decided early on to focus on the
growth in
test scores between two points in time rather than a static achievement measure captured only once a year.
Building on research presented during the Kinder Institute's October KIForum, Reardon's working paper
uses a measure
of educational opportunity meant to track student
growth from grades three through eight utilizing standardized
test scores for roughly 45 million students in more than 11,000 school districts across the country.
So, when Adair County Elementary School (ACES) staff members were brainstorming ways to get their students excited about
using Study Island to prepare for the NWEA MAP
test in math, and ultimately increase
growth scores, they decided to tap into the spirit
of competition.
Graduation rates (along with
test score growth data) is a critical component
of No Child's AYP system, while any overhaul
of No Child should include
using school discipline data along with chronic truancy rates (ideally, based on 10 or more days
of unexcused absence, as
used in Indiana) as a component
of accountability.
He also reiterated the union's opposition to the district's
use of Academic
Growth over Time data, which is based on state standardized
test scores and is being
used to evaluate teachers and principals in a voluntary program.
Yet while the idea
of using student
test scores for teacher evaluations may be conceptually appealing, there is no universally accepted methodology for translating student
growth into a measure
of teacher performance.
The letter grade is based 80 percent on the school's achievement
score (which
uses various data including student performance on end -
of - grade and end -
of - course standardized
test scores) and 20 percent on students» academic
growth (a measure
of students» performance in relation to their expected performance based on the prior year's
test results), resulting in a grade
of A, B, C, D, or F. «Low - performing districts» are those with over 50 percent
of their schools identified as low - performing.
Moreover, in many states, up to half
of each evaluation
score is based on value - added analysis, a complex statistical method that seeks to determine each teacher's contribution to student
growth using student
test score data.
On this note, and «[i] n sum, recent research on value added tells us that, by
using data from student perceptions, classroom observations, and
test score growth, we can obtain credible evidence [albeit weakly related evidence, referring to the Bill & Melinda Gates Foundation's MET studies]
of the relative effectiveness
of a set
of teachers who teach similar kids [emphasis added] under similar conditions [emphasis added]... [Although] if a district administrator
uses data like that collected in MET, we can anticipate that an attempt to classify teachers for personnel decisions will be characterized by intolerably high error rates [emphasis added].
«Multimetric accountability systems should
use formative assessments, evidence
of student learning, and progress toward personal
growth objectives to measure student and teacher success rather than rely on standardized
test scores as the primary reference point.»
Student achievement
growth may have a role to play, but
using test scores of these students to determine that
growth presents unique challenges to value - added modeling.
His recent research includes the study
of how student mobility rates affect the rate
of learning
growth, the
use of surveys
of student perceptions in evaluation classroom environments, the effects
of homogenous ability grouping and tracking, and the interpretation
of value - added
test scores.
Unfortunately, in contract negotiations SEA allowed Seattle to became the only city in the entire state to allow two measures
of student
growth in educators» evaluations, including the
use of state standardized
tests scores.
Suburban districts, after all, also have to deal with quality - blind seniority - based privileges such as reverse - seniority layoff rules, pay scales that favor seat time over performance, and restrictions on the
use of objective student
test score growth data from
use in teacher evaluations.
It took more than five months and the intervention
of a mediator to craft an evaluation that factors in standardized
test scores, as well as Academic
Growth Over Time, a controversial mathematical formula
used to measure student progress.