SlideShare uma empresa Scribd logo
1 de 56
Implementing the Connecticut teacher evaluation
system

John Cronin, Ph.D.
Director
The Kingsbury Center @ NWEA
Helping teacher’s set reasonable and rigorous
goals

Presenter - John Cronin, Ph.D.
Contacting us:
Rebecca Moore: 503-548-5129
E-mail: rebecca.moore@nwea.org
What you’ll learn
• The purposes of teacher evaluation. There can be different purposes for
different educators.
• The value of differentiating educator evaluations and the risks associated
with not differentiating.
• The value of multiple measurements, and the importance of defining a
purpose for each measurement.
• The information needed to determine whether a goal is attainable
• The difference between “aspirational” and “evaluative” goals and the
value of each.
• Thoughts on strategies for addressing the difficult to measure.
What’s the purpose?
• The Connecticut system requires a collaborative process.
• For most educators the purpose of evaluation is formative.
• For a small minority of educators the purpose is to summative
and goals may involve demonstrating basic competence.
• Leaders should be transparent about the purpose of the
process for each educator.
• Perfect consistency isn’t necessarily a requirement.
Differences between principal and
teacher evaluation

Principals
Thus principals should bestaff.
• Inherit a pre-existing
evaluated on their ability to improve
• Have limited control over
growth or maintain high levels of
staffing conditions
growth over time rather than their
students’with this intact school
• Work growth within a
year.
group from year to year.
Differences between principal and
teacher evaluation

Teachers

• They have new groups of
students each year
• They generally work with
those students for one
school year only.
• New teachers generally
become more effective in
their first three years
The difference between formative and
summative evaluation
• Formative evaluation is intended to give educators
useful feedback to help them improve their job
performance. For most educators formative
evaluation should be the focus of the process.
• Summative evaluation is a judgment of educator
performance that informs future decisions about
employment including granting of tenure,
performance pay, protection from layoff.
Purposes of summative evaluation
• An accurate and defensible judgment of an
educator’s job performance.
• Ratings of performance that provide
meaningful differentiation across educators.
• Goals of evaluation
– Support professional improvement
– Retain your top educators
– Dismiss ineffective educators
If evaluators do not
differentiate their
ratings, then all
differentiation
comes from the test.
If performance
ratings aren’t
consistent with
school growth, that
will probably be
public information.
Results of Tennessee Teacher Evaluation
Pilot
60%
50%
40%
30%

Value-added result
Observation Result

20%
10%
0%
1

2

3

4

5
Results of Georgia Teacher Evaluation Pilot
Evaluator Rating

ineffective
Minimally Effective
Effective
Highly Effective
Connecticut expectations around
teacher observation
• First and Second year teachers –
– Required – 3 formal observations
– Recommended – 3 formal observations and 3 informal
observations
• Below standard and developing –
– Required – 3 formal evaluations
– Recommended – 3 formal evaluations and 5 informal
evaluations
• Proficient and exemplary –
– Recommended – 3 formal evaluations – 1 in class
– Required – 3 formal evaluations – 1 in class
Reliability of evaluation weights in predicted
stability of student growth gains year to year
Observation by

Reliability coefficient
(relative to state test
value-added gain)

Proportion of test
variance
explained

Model 1 – State test – 81%
Student surveys 17% Classroom
Observations – 2%

.51

26.0%

Model 2 – State test – 50%
Student Surveys – 25%
Classroom Observation – 25%

.66

43.5%

Model 3 – State test – 33% Student Surveys – 33%
Classroom Observations – 33%

.76

57.7%%

Model 4 – Classroom Observation
50%
State test – 25%
Student surveys – 25%

.75

56.2%

Bill and Melina Gates Foundation (2013, January). Ensuring Fair and Reliable Measures of Effective
Teaching: Culminating Findings from the MET Projects Three-Year Study
Reliability of a variety of teacher observation
implementations
Observation by

Reliability coefficient
(relative to state test
value-added gain)

Proportion of test
variance
explained

Principal – 1

.51

26.0%

Principal – 2

.58

33.6%

Principal and other administrator

.67

44.9%

Principal and three short
observations by peer observers

.67

44.9%

Two principal observations and
two peer observations

.66

43.6%

Two principal observations and
two different peer observers

.69

47.6%

Two principal observations one
peer observation and three short
observations by peers

.72

51.8%

Bill and Melina Gates Foundation (2013, January). Ensuring Fair and Reliable Measures of Effective
Teaching: Culminating Findings from the MET Projects Three-Year Study
Why should we care about goal
setting in education?
Because we want students
to learn more!

• Research view
–Setting goals improves performance
The testing to teacher evaluation process

Testing
Metric (Growth Score)

Analysis (Value-Added)
Evaluation (Rating)
The difference between growth
and improvement
One district’s change in 5th grade math performance
relative to Kentucky cut scores

Mathematics

Number of Students

No Change
Down
Up

Fall RIT
Number of students who achieved the normal
mathematics growth in that district

Mathematics
Failed growth target

Number of Students

Met growth target

Student’s score in fall
Issues in the use of growth measures

Measurement design of the
instrument
Many assessments are not
designed to measure growth.
Others do not measure growth
equally well for all students.
Tests are not equally accurate for all
students
California STAR

NWEA MAP
Issues with rubric based instruments
• Rubrics should be granular enough to show
growth. Four point rubrics may not be
adequate.
• High and low scores should be written in a
manner that sets a reasonable floor and
ceiling.
Expect consistent inconsistency!
Inconsistency occurs because
• Of differences in test design.
• Differences in testing conditions.
• Differences in models being applied to
evaluate growth.
The reliability problem –
Inconsistency in testing conditions

Test

Test 1
Time 1

Test 2
Time 1

Test 1
Time 2

Retest

Test 2
Time 2
The reliability problem –
Inconsistency in testing conditions
Test 1
Time 1

Test 1
Time 1

Test 1
Time 1

Test 2
Time 1

Test 2
Time 1

Test 2
Time 1

Test 1
Time 2

Test 1
Time 2

Test 1
Time 2

Test 2
Time 2

Test 2
Time 2

Test 2
Time 2
The problem with spring-spring testing

Teacher 1
3/11

4/11

5/11

Summer
6/11

7/11

8/11

Teacher 2
9/11

10/11

11/11

12/11

1/12

2/12

3/12
The testing to teacher evaluation process

Testing
Metric (Growth Score)

Analysis (Value-Added)
Evaluation (Rating)
Issues in the use of growth and valueadded measures

Differences among value-added
models
Los Angeles Times Study
Los Angeles Times Study #2
Issues in the use of value-added measures

Control for statistical error
All models attempt to address this
issue. Nevertheless, many teachers
value-added scores will fall within
the range of statistical error.
Issues in the use of growth and valueadded measures

Control for statistical error
New York City

New York City #2
Range of teacher value-added
estimates
12.00
11.00

Mathematics Growth Index Distribution by Teacher - Validity Filtered

10.00
9.00
8.00
7.00

Each line in this display represents a single teacher. The graphic
shows the average growth index score for each teacher (green
line), plus or minus the standard error of the growth index estimate
(black line). We removed students who had tests of questionable
validity and teachers with fewer than 20 students.

Average Growth Index Score and Range

6.00
5.00
4.00
3.00
2.00

Q5
Q4

1.00
0.00

Q3

-1.00
-2.00

-3.00
-4.00
-5.00
-6.00
-7.00
-8.00
-9.00
-10.00
-11.00

-12.00

Q2
Q1
What’s a SMART goal?
•
•
•
•
•

Specific
Measurable
Attainable
Relevant
Time-Bound
Specific
• What: What do I want to accomplish?
• Why: What are the reasons or purpose for
pursuing the goal?
• Which: What are the requirements and
constraints for achieving the goal?
SMART goal resources
• National Staff Development Council. Provides
a nice process for developing SMART goals.
• Arlington Public Schools. Excellent and
detailed examples of SMART goals across
subject disciplines, including art and music.
• The Handbook for Smart School Teams. Anne
Conzemius and Jan O’Neill.
Issues with local tests and goal
setting
• Validity and reliability of assessments.
• Teachers/administrators are unlikely to set
goals that are inconsistent with their current
performance.
• It is difficult to set goals without prior
evidence or context.
Specific

The goal should ALWAYS be
improvement in a domain (subject)!
Specific

There should ALWAYS be multiple
data sources and metrics.
Data should be triangulated
• Classroom assessment data to standardized
test data.
• Domain data (mathematics) to sub-domain
data (fractions and decimals) to granular data
(division with fractions).
Specific

All students should be “in play” relative
to the goal.
Measurable
Types of Goals
• Performance – 75% of the students in my 7th grade
mathematics class will achieve the qualifying score
needed for placement in 8th grade Algebra.
• Growth – 65% of my students will show growth on the
OAKS mathematics test that is greater than the state
reported norm.
• Improvement – Last year 40% of my students showed
growth on the OAKS mathematics test that was greater
than the norm. This year 50% of my students will show
greater than normal growth.
Attainable

The goals set should be reasonable and
rigorous. At minimum, they should
represent a level of performance that a
competent educator could be expected
to achieve.
An analogy to baseball
Center Fielders - WAR (Wins Above Replacement)
6.4
6.1

Superstar – Mike Trout – Los Angeles Angels

1.7

Median Major Leaguer – Gregor Blanco, San Francisco Giants

0.0

Marginal Major Leaguer – Chris Young, Oakland As

-1.3
The difference between aspirational and
evaluative goals
Aspirational – I will meet my target weight by losing
50 pounds during the next year and sustain that
weight for one year.
Proficient – I intend to lose 15 pounds in the next six
months, which will move me from the “obese” to the
“overweight” category, and sustain that weight for one
year.
Marginal – I will lose weight in the next six months.
Ways to evaluate the attainability of a goal
• Prior performance
• Performance of peers within the system
• Performance of a norming group
One approach to evaluating the attainment
of goals.
Students in La Brea Elementary School show
mathematics growth equivalent to only 2/3 of the
average for students in their grade.
Level 4 – (Aspirational) – Students in La Brea Elementary School will
improve their mathematics growth equivalent to 1.5 times the average
for their grade.
Level 3 – (Proficient) Students in La Brea Elementary School will
improve their mathematics growth to be equivalent to the average for
their grade.
Level 2 – (Marginal) Students in La Brea Elementary School will
improve their mathematics growth relative to last year.
Level 1 – (Unacceptable) Students in La Brea Elementary School
do not improve their mathematics growth relative to last year.
Is this goal attainable?
62% of students at John Glenn Elementary met or exceeded
proficiency in Reading/Literature last year. Their goal is to improve
their rate to 82% this year. Is the goal attainable?

Oregon schools – change in
Reading/Literature proficiency 2009-10 to
2010-11 among schools that started with
60% proficiency rates
400
300

200
100

362

351

291

173

0
Growth > -20% > -10% > 0%
> -30%

73

14

3

> 10% > 20% > 30%
Is this goal attainable and
rigorous?
45% of the students at La Brea elementary showed average growth or
better last year. Their goal is to improve that rate to 50% this year. Is
their goal reasonable?

Students with average or better annual
growth in Repus school district
100%
80%
60%

40%
20%
0%
The selection of metrics matters

Students at LaBrea Elementary School
will show growth equivalent to 150% of
grade level.

Students at Etsaw Middle School will
show growth equivalent to 150% of grade
level.
Scale score growth relative to NWEA’s
growth norm in mathematics
Growth Index
7.0

Scale Score Growth

6.0
5.0
4.0
3.0
2.0
1.0
0.0
-1.0

2

3

4

5

6

7

8

9
Percent of a Year’s Growth

Percent of a year’s growth in
mathematics
Mathematics

200%
180%
160%
140%
120%
100%
80%
60%
40%
20%
0%
2

3

4

5

6

7

8

9
Assessing the difficult to measure
• Encourage use of performance assessment and rubrics.
• Encourage outside scoring
– Use of peers in other buildings, professionals in the field,
contest judges
• Make use of resources
– Music educator, art educator, vocational professional
associations
– Available models – AP art portfolio.
– Use your intermediate agency
– Work across buildings
• Make use of classroom observation.
The outcome is important, but why the
outcome occurred is equally important!
Success can’t be replicated if you don’t know why you
succeeded.
Failure can’t be reversed if you don’t know why you failed.
The outcome is important, but why the
outcome occurred is equally important!
• Establish checkpoints and check-in beyond
what’s required when possible.
• Collect implementation information
– Classroom observations - visits
– Teacher journal
– Student work – artifacts
These processes can be done by teacher peers!
Thank you for attending

Presenter - John Cronin, Ph.D.
Contacting us:
NWEA Main Number: 503-624-1951
E-mail: rebecca.moore@nwea.org

Mais conteúdo relacionado

Mais procurados

Item analysis and validation
Item analysis and validationItem analysis and validation
Item analysis and validation
KEnkenken Tan
 
Patuxent High School Collaborative Action Research Plan
Patuxent High School Collaborative Action Research PlanPatuxent High School Collaborative Action Research Plan
Patuxent High School Collaborative Action Research Plan
Matt Stone
 
Achievement tests
Achievement testsAchievement tests
Achievement tests
Manu Sethi
 
Ma sampletest-hs 2010-13
Ma sampletest-hs 2010-13Ma sampletest-hs 2010-13
Ma sampletest-hs 2010-13
Erlinda Rey
 
Qualities of a good test (1)
Qualities of a good test (1)Qualities of a good test (1)
Qualities of a good test (1)
kimoya
 

Mais procurados (20)

Characteristics of Good Evaluation Instrument
Characteristics of Good Evaluation InstrumentCharacteristics of Good Evaluation Instrument
Characteristics of Good Evaluation Instrument
 
Internal and external examination
Internal and external  examinationInternal and external  examination
Internal and external examination
 
Characteristics of a good measuring tool
Characteristics of a good measuring toolCharacteristics of a good measuring tool
Characteristics of a good measuring tool
 
5 achievement
5 achievement5 achievement
5 achievement
 
Selecting and using standardized achievement test
Selecting and using standardized achievement testSelecting and using standardized achievement test
Selecting and using standardized achievement test
 
Item analysis and validation
Item analysis and validationItem analysis and validation
Item analysis and validation
 
Collaborative Examination Item Review Process in a Team-Taught Course
Collaborative Examination Item Review Process in a Team-Taught CourseCollaborative Examination Item Review Process in a Team-Taught Course
Collaborative Examination Item Review Process in a Team-Taught Course
 
Item analysis in MCQs
Item analysis in MCQsItem analysis in MCQs
Item analysis in MCQs
 
Evaluation in education
Evaluation in educationEvaluation in education
Evaluation in education
 
Patuxent High School Collaborative Action Research Plan
Patuxent High School Collaborative Action Research PlanPatuxent High School Collaborative Action Research Plan
Patuxent High School Collaborative Action Research Plan
 
Quantitative & Qualitative Assessment
Quantitative & Qualitative AssessmentQuantitative & Qualitative Assessment
Quantitative & Qualitative Assessment
 
Achievement Test
Achievement TestAchievement Test
Achievement Test
 
Achievement tests
Achievement testsAchievement tests
Achievement tests
 
Ma sampletest-hs 2010-13
Ma sampletest-hs 2010-13Ma sampletest-hs 2010-13
Ma sampletest-hs 2010-13
 
Language testing and evaluation
Language testing and evaluationLanguage testing and evaluation
Language testing and evaluation
 
Instructional leardership chpt.8
Instructional leardership chpt.8Instructional leardership chpt.8
Instructional leardership chpt.8
 
Qualities of a good test (1)
Qualities of a good test (1)Qualities of a good test (1)
Qualities of a good test (1)
 
Item analysis
Item analysisItem analysis
Item analysis
 
Evaluation methods
Evaluation methodsEvaluation methods
Evaluation methods
 
achievement test
achievement testachievement test
achievement test
 

Destaque

Teacher evaluation presentation mississippi
Teacher evaluation presentation mississippiTeacher evaluation presentation mississippi
Teacher evaluation presentation mississippi
John Cronin
 
Mobile Strategy @ 20FourLabs
Mobile Strategy @ 20FourLabsMobile Strategy @ 20FourLabs
Mobile Strategy @ 20FourLabs
Matthew Buckland
 
01 introdução e revisão
01   introdução e revisão01   introdução e revisão
01 introdução e revisão
confidencial
 
Job competence for process operator
Job competence for process operatorJob competence for process operator
Job competence for process operator
Fariz Hafiz
 
Circular marzo2015web
Circular marzo2015webCircular marzo2015web
Circular marzo2015web
ClubTrevinca
 
Guia de lagartos da reserva adolpho ducke, amazonia central 2008
Guia de lagartos da reserva adolpho ducke, amazonia central 2008Guia de lagartos da reserva adolpho ducke, amazonia central 2008
Guia de lagartos da reserva adolpho ducke, amazonia central 2008
Coonass
 

Destaque (20)

How far will the language go with gender?
How far will the language go with gender?How far will the language go with gender?
How far will the language go with gender?
 
Teacher evaluation presentation mississippi
Teacher evaluation presentation mississippiTeacher evaluation presentation mississippi
Teacher evaluation presentation mississippi
 
Immobest artikel svit_lernende
Immobest artikel svit_lernendeImmobest artikel svit_lernende
Immobest artikel svit_lernende
 
Contaminación acústica CMC
Contaminación acústica  CMCContaminación acústica  CMC
Contaminación acústica CMC
 
Bloque l libro 1
Bloque l libro 1Bloque l libro 1
Bloque l libro 1
 
Tipos de software
Tipos de softwareTipos de software
Tipos de software
 
EL PODER DE RE CREAR
EL PODER DE RE CREAREL PODER DE RE CREAR
EL PODER DE RE CREAR
 
İLKE Presentation
İLKE PresentationİLKE Presentation
İLKE Presentation
 
The perceived effectiveness of Assessment Centers for Indian IT and call cent...
The perceived effectiveness of Assessment Centers for Indian IT and call cent...The perceived effectiveness of Assessment Centers for Indian IT and call cent...
The perceived effectiveness of Assessment Centers for Indian IT and call cent...
 
Brochure
BrochureBrochure
Brochure
 
Mobile Strategy @ 20FourLabs
Mobile Strategy @ 20FourLabsMobile Strategy @ 20FourLabs
Mobile Strategy @ 20FourLabs
 
Guachapp
GuachappGuachapp
Guachapp
 
Manual del hombre nuevo
Manual del hombre nuevoManual del hombre nuevo
Manual del hombre nuevo
 
Cuttings garden 2011
Cuttings garden   2011Cuttings garden   2011
Cuttings garden 2011
 
01 introdução e revisão
01   introdução e revisão01   introdução e revisão
01 introdução e revisão
 
Paper de iw
Paper de iwPaper de iw
Paper de iw
 
Job competence for process operator
Job competence for process operatorJob competence for process operator
Job competence for process operator
 
Taller práctico autónomos 2013
Taller práctico autónomos 2013Taller práctico autónomos 2013
Taller práctico autónomos 2013
 
Circular marzo2015web
Circular marzo2015webCircular marzo2015web
Circular marzo2015web
 
Guia de lagartos da reserva adolpho ducke, amazonia central 2008
Guia de lagartos da reserva adolpho ducke, amazonia central 2008Guia de lagartos da reserva adolpho ducke, amazonia central 2008
Guia de lagartos da reserva adolpho ducke, amazonia central 2008
 

Semelhante a Teacher evaluation and goal setting connecticut

Ed Reform Lecture - University of Arkansas
Ed Reform Lecture - University of ArkansasEd Reform Lecture - University of Arkansas
Ed Reform Lecture - University of Arkansas
John Cronin
 
Maximizing student assessment systems cronin
Maximizing student assessment systems   croninMaximizing student assessment systems   cronin
Maximizing student assessment systems cronin
NWEA
 
2010 ohio tif meeting creating a comprehensive teacher effectiveness system
2010 ohio tif meeting  creating a comprehensive teacher effectiveness system2010 ohio tif meeting  creating a comprehensive teacher effectiveness system
2010 ohio tif meeting creating a comprehensive teacher effectiveness system
Christopher Thorn
 
Connecticut mesuring and modeling growth
Connecticut   mesuring and modeling growthConnecticut   mesuring and modeling growth
Connecticut mesuring and modeling growth
John Cronin
 
Connecticut mesuring and modeling growth
Connecticut   mesuring and modeling growthConnecticut   mesuring and modeling growth
Connecticut mesuring and modeling growth
John Cronin
 
Principal and-teacher-evaluation-key-ideas-your-role-and-your-school's-leadin...
Principal and-teacher-evaluation-key-ideas-your-role-and-your-school's-leadin...Principal and-teacher-evaluation-key-ideas-your-role-and-your-school's-leadin...
Principal and-teacher-evaluation-key-ideas-your-role-and-your-school's-leadin...
Winnie de Leon
 
Using tests for teacher evaluation texas
Using tests for teacher evaluation texasUsing tests for teacher evaluation texas
Using tests for teacher evaluation texas
NWEA
 

Semelhante a Teacher evaluation and goal setting connecticut (20)

Ed Reform Lecture - University of Arkansas
Ed Reform Lecture - University of ArkansasEd Reform Lecture - University of Arkansas
Ed Reform Lecture - University of Arkansas
 
Colorado assessment summit_teacher_eval
Colorado assessment summit_teacher_evalColorado assessment summit_teacher_eval
Colorado assessment summit_teacher_eval
 
Maximizing student assessment systems cronin
Maximizing student assessment systems   croninMaximizing student assessment systems   cronin
Maximizing student assessment systems cronin
 
2010 ohio tif meeting creating a comprehensive teacher effectiveness system
2010 ohio tif meeting  creating a comprehensive teacher effectiveness system2010 ohio tif meeting  creating a comprehensive teacher effectiveness system
2010 ohio tif meeting creating a comprehensive teacher effectiveness system
 
Maximizing student assessment systems cronin
Maximizing student assessment systems   croninMaximizing student assessment systems   cronin
Maximizing student assessment systems cronin
 
Using Assessment Data for Educator and Student Growth
Using Assessment Data for Educator and Student GrowthUsing Assessment Data for Educator and Student Growth
Using Assessment Data for Educator and Student Growth
 
Taking control of the South Carolina Teacher Evaluation framework
Taking control of the South Carolina Teacher Evaluation frameworkTaking control of the South Carolina Teacher Evaluation framework
Taking control of the South Carolina Teacher Evaluation framework
 
Connecticut mesuring and modeling growth
Connecticut   mesuring and modeling growthConnecticut   mesuring and modeling growth
Connecticut mesuring and modeling growth
 
Connecticut mesuring and modeling growth
Connecticut   mesuring and modeling growthConnecticut   mesuring and modeling growth
Connecticut mesuring and modeling growth
 
Connecticut mesuring and modeling growth
Connecticut   mesuring and modeling growthConnecticut   mesuring and modeling growth
Connecticut mesuring and modeling growth
 
Principal and-teacher-evaluation-key-ideas-your-role-and-your-school's-leadin...
Principal and-teacher-evaluation-key-ideas-your-role-and-your-school's-leadin...Principal and-teacher-evaluation-key-ideas-your-role-and-your-school's-leadin...
Principal and-teacher-evaluation-key-ideas-your-role-and-your-school's-leadin...
 
NYSCOSS Conference Superintendents Training on Assessment 9 14
NYSCOSS Conference Superintendents Training on Assessment 9 14NYSCOSS Conference Superintendents Training on Assessment 9 14
NYSCOSS Conference Superintendents Training on Assessment 9 14
 
Using tests for teacher evaluation texas
Using tests for teacher evaluation texasUsing tests for teacher evaluation texas
Using tests for teacher evaluation texas
 
Assessment 101 Parts 1 & 2
Assessment 101 Parts 1 & 2Assessment 101 Parts 1 & 2
Assessment 101 Parts 1 & 2
 
NWEA Growth and Teacher evaluation VA 9-13
NWEA Growth and Teacher evaluation VA 9-13NWEA Growth and Teacher evaluation VA 9-13
NWEA Growth and Teacher evaluation VA 9-13
 
Teacher evaluation presentation oregon
Teacher evaluation presentation   oregonTeacher evaluation presentation   oregon
Teacher evaluation presentation oregon
 
teacher made test Vs standardized test
 teacher made test Vs standardized test teacher made test Vs standardized test
teacher made test Vs standardized test
 
ED544422.pdf
ED544422.pdfED544422.pdf
ED544422.pdf
 
Summative and Formative Assessments1.ppt
Summative and Formative Assessments1.pptSummative and Formative Assessments1.ppt
Summative and Formative Assessments1.ppt
 
Summative and Formative Assessments.ppt
Summative and Formative Assessments.pptSummative and Formative Assessments.ppt
Summative and Formative Assessments.ppt
 

Mais de John Cronin

Colorado assessment summit_oct12
Colorado assessment summit_oct12Colorado assessment summit_oct12
Colorado assessment summit_oct12
John Cronin
 

Mais de John Cronin (15)

Nycoss presentation
Nycoss presentationNycoss presentation
Nycoss presentation
 
California administrator symposium nwea
California administrator symposium nweaCalifornia administrator symposium nwea
California administrator symposium nwea
 
Seven purposes presentation
Seven purposes presentationSeven purposes presentation
Seven purposes presentation
 
Chief accountability officers presentation
Chief accountability officers presentationChief accountability officers presentation
Chief accountability officers presentation
 
Triggers for college success cr
Triggers for college success crTriggers for college success cr
Triggers for college success cr
 
College
CollegeCollege
College
 
Nyinst
NyinstNyinst
Nyinst
 
New ways to think about framing accountability to your community
New ways to think about framing accountability to your communityNew ways to think about framing accountability to your community
New ways to think about framing accountability to your community
 
Valid data for school improvement final
Valid data for school improvement finalValid data for school improvement final
Valid data for school improvement final
 
College readiness presentation
College readiness presentationCollege readiness presentation
College readiness presentation
 
Tasa presentation version 2
Tasa presentation version 2Tasa presentation version 2
Tasa presentation version 2
 
The purpose driven assessment system
The purpose driven assessment systemThe purpose driven assessment system
The purpose driven assessment system
 
Teacher evaluation present
Teacher evaluation presentTeacher evaluation present
Teacher evaluation present
 
Parent conferencing with map
Parent conferencing with mapParent conferencing with map
Parent conferencing with map
 
Colorado assessment summit_oct12
Colorado assessment summit_oct12Colorado assessment summit_oct12
Colorado assessment summit_oct12
 

Último

Making and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdfMaking and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdf
Chris Hunter
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
QucHHunhnh
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
ciinovamais
 

Último (20)

microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
Sociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning ExhibitSociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning Exhibit
 
Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docx
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdf
 
Making and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdfMaking and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdf
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
 
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy Consulting
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104
 
Role Of Transgenic Animal In Target Validation-1.pptx
Role Of Transgenic Animal In Target Validation-1.pptxRole Of Transgenic Animal In Target Validation-1.pptx
Role Of Transgenic Animal In Target Validation-1.pptx
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 

Teacher evaluation and goal setting connecticut

  • 1. Implementing the Connecticut teacher evaluation system John Cronin, Ph.D. Director The Kingsbury Center @ NWEA
  • 2. Helping teacher’s set reasonable and rigorous goals Presenter - John Cronin, Ph.D. Contacting us: Rebecca Moore: 503-548-5129 E-mail: rebecca.moore@nwea.org
  • 3. What you’ll learn • The purposes of teacher evaluation. There can be different purposes for different educators. • The value of differentiating educator evaluations and the risks associated with not differentiating. • The value of multiple measurements, and the importance of defining a purpose for each measurement. • The information needed to determine whether a goal is attainable • The difference between “aspirational” and “evaluative” goals and the value of each. • Thoughts on strategies for addressing the difficult to measure.
  • 4. What’s the purpose? • The Connecticut system requires a collaborative process. • For most educators the purpose of evaluation is formative. • For a small minority of educators the purpose is to summative and goals may involve demonstrating basic competence. • Leaders should be transparent about the purpose of the process for each educator. • Perfect consistency isn’t necessarily a requirement.
  • 5. Differences between principal and teacher evaluation Principals Thus principals should bestaff. • Inherit a pre-existing evaluated on their ability to improve • Have limited control over growth or maintain high levels of staffing conditions growth over time rather than their students’with this intact school • Work growth within a year. group from year to year.
  • 6. Differences between principal and teacher evaluation Teachers • They have new groups of students each year • They generally work with those students for one school year only. • New teachers generally become more effective in their first three years
  • 7. The difference between formative and summative evaluation • Formative evaluation is intended to give educators useful feedback to help them improve their job performance. For most educators formative evaluation should be the focus of the process. • Summative evaluation is a judgment of educator performance that informs future decisions about employment including granting of tenure, performance pay, protection from layoff.
  • 8. Purposes of summative evaluation • An accurate and defensible judgment of an educator’s job performance. • Ratings of performance that provide meaningful differentiation across educators. • Goals of evaluation – Support professional improvement – Retain your top educators – Dismiss ineffective educators
  • 9. If evaluators do not differentiate their ratings, then all differentiation comes from the test.
  • 10. If performance ratings aren’t consistent with school growth, that will probably be public information.
  • 11. Results of Tennessee Teacher Evaluation Pilot 60% 50% 40% 30% Value-added result Observation Result 20% 10% 0% 1 2 3 4 5
  • 12. Results of Georgia Teacher Evaluation Pilot Evaluator Rating ineffective Minimally Effective Effective Highly Effective
  • 13. Connecticut expectations around teacher observation • First and Second year teachers – – Required – 3 formal observations – Recommended – 3 formal observations and 3 informal observations • Below standard and developing – – Required – 3 formal evaluations – Recommended – 3 formal evaluations and 5 informal evaluations • Proficient and exemplary – – Recommended – 3 formal evaluations – 1 in class – Required – 3 formal evaluations – 1 in class
  • 14. Reliability of evaluation weights in predicted stability of student growth gains year to year Observation by Reliability coefficient (relative to state test value-added gain) Proportion of test variance explained Model 1 – State test – 81% Student surveys 17% Classroom Observations – 2% .51 26.0% Model 2 – State test – 50% Student Surveys – 25% Classroom Observation – 25% .66 43.5% Model 3 – State test – 33% Student Surveys – 33% Classroom Observations – 33% .76 57.7%% Model 4 – Classroom Observation 50% State test – 25% Student surveys – 25% .75 56.2% Bill and Melina Gates Foundation (2013, January). Ensuring Fair and Reliable Measures of Effective Teaching: Culminating Findings from the MET Projects Three-Year Study
  • 15. Reliability of a variety of teacher observation implementations Observation by Reliability coefficient (relative to state test value-added gain) Proportion of test variance explained Principal – 1 .51 26.0% Principal – 2 .58 33.6% Principal and other administrator .67 44.9% Principal and three short observations by peer observers .67 44.9% Two principal observations and two peer observations .66 43.6% Two principal observations and two different peer observers .69 47.6% Two principal observations one peer observation and three short observations by peers .72 51.8% Bill and Melina Gates Foundation (2013, January). Ensuring Fair and Reliable Measures of Effective Teaching: Culminating Findings from the MET Projects Three-Year Study
  • 16. Why should we care about goal setting in education? Because we want students to learn more! • Research view –Setting goals improves performance
  • 17. The testing to teacher evaluation process Testing Metric (Growth Score) Analysis (Value-Added) Evaluation (Rating)
  • 18. The difference between growth and improvement
  • 19. One district’s change in 5th grade math performance relative to Kentucky cut scores Mathematics Number of Students No Change Down Up Fall RIT
  • 20. Number of students who achieved the normal mathematics growth in that district Mathematics Failed growth target Number of Students Met growth target Student’s score in fall
  • 21. Issues in the use of growth measures Measurement design of the instrument Many assessments are not designed to measure growth. Others do not measure growth equally well for all students.
  • 22. Tests are not equally accurate for all students California STAR NWEA MAP
  • 23. Issues with rubric based instruments • Rubrics should be granular enough to show growth. Four point rubrics may not be adequate. • High and low scores should be written in a manner that sets a reasonable floor and ceiling.
  • 25. Inconsistency occurs because • Of differences in test design. • Differences in testing conditions. • Differences in models being applied to evaluate growth.
  • 26. The reliability problem – Inconsistency in testing conditions Test Test 1 Time 1 Test 2 Time 1 Test 1 Time 2 Retest Test 2 Time 2
  • 27. The reliability problem – Inconsistency in testing conditions Test 1 Time 1 Test 1 Time 1 Test 1 Time 1 Test 2 Time 1 Test 2 Time 1 Test 2 Time 1 Test 1 Time 2 Test 1 Time 2 Test 1 Time 2 Test 2 Time 2 Test 2 Time 2 Test 2 Time 2
  • 28. The problem with spring-spring testing Teacher 1 3/11 4/11 5/11 Summer 6/11 7/11 8/11 Teacher 2 9/11 10/11 11/11 12/11 1/12 2/12 3/12
  • 29. The testing to teacher evaluation process Testing Metric (Growth Score) Analysis (Value-Added) Evaluation (Rating)
  • 30. Issues in the use of growth and valueadded measures Differences among value-added models Los Angeles Times Study Los Angeles Times Study #2
  • 31. Issues in the use of value-added measures Control for statistical error All models attempt to address this issue. Nevertheless, many teachers value-added scores will fall within the range of statistical error.
  • 32. Issues in the use of growth and valueadded measures Control for statistical error New York City New York City #2
  • 33. Range of teacher value-added estimates 12.00 11.00 Mathematics Growth Index Distribution by Teacher - Validity Filtered 10.00 9.00 8.00 7.00 Each line in this display represents a single teacher. The graphic shows the average growth index score for each teacher (green line), plus or minus the standard error of the growth index estimate (black line). We removed students who had tests of questionable validity and teachers with fewer than 20 students. Average Growth Index Score and Range 6.00 5.00 4.00 3.00 2.00 Q5 Q4 1.00 0.00 Q3 -1.00 -2.00 -3.00 -4.00 -5.00 -6.00 -7.00 -8.00 -9.00 -10.00 -11.00 -12.00 Q2 Q1
  • 34. What’s a SMART goal? • • • • • Specific Measurable Attainable Relevant Time-Bound
  • 35. Specific • What: What do I want to accomplish? • Why: What are the reasons or purpose for pursuing the goal? • Which: What are the requirements and constraints for achieving the goal?
  • 36. SMART goal resources • National Staff Development Council. Provides a nice process for developing SMART goals. • Arlington Public Schools. Excellent and detailed examples of SMART goals across subject disciplines, including art and music. • The Handbook for Smart School Teams. Anne Conzemius and Jan O’Neill.
  • 37. Issues with local tests and goal setting • Validity and reliability of assessments. • Teachers/administrators are unlikely to set goals that are inconsistent with their current performance. • It is difficult to set goals without prior evidence or context.
  • 38. Specific The goal should ALWAYS be improvement in a domain (subject)!
  • 39. Specific There should ALWAYS be multiple data sources and metrics.
  • 40. Data should be triangulated • Classroom assessment data to standardized test data. • Domain data (mathematics) to sub-domain data (fractions and decimals) to granular data (division with fractions).
  • 41. Specific All students should be “in play” relative to the goal.
  • 42. Measurable Types of Goals • Performance – 75% of the students in my 7th grade mathematics class will achieve the qualifying score needed for placement in 8th grade Algebra. • Growth – 65% of my students will show growth on the OAKS mathematics test that is greater than the state reported norm. • Improvement – Last year 40% of my students showed growth on the OAKS mathematics test that was greater than the norm. This year 50% of my students will show greater than normal growth.
  • 43. Attainable The goals set should be reasonable and rigorous. At minimum, they should represent a level of performance that a competent educator could be expected to achieve.
  • 44. An analogy to baseball Center Fielders - WAR (Wins Above Replacement) 6.4 6.1 Superstar – Mike Trout – Los Angeles Angels 1.7 Median Major Leaguer – Gregor Blanco, San Francisco Giants 0.0 Marginal Major Leaguer – Chris Young, Oakland As -1.3
  • 45. The difference between aspirational and evaluative goals Aspirational – I will meet my target weight by losing 50 pounds during the next year and sustain that weight for one year. Proficient – I intend to lose 15 pounds in the next six months, which will move me from the “obese” to the “overweight” category, and sustain that weight for one year. Marginal – I will lose weight in the next six months.
  • 46. Ways to evaluate the attainability of a goal • Prior performance • Performance of peers within the system • Performance of a norming group
  • 47. One approach to evaluating the attainment of goals. Students in La Brea Elementary School show mathematics growth equivalent to only 2/3 of the average for students in their grade. Level 4 – (Aspirational) – Students in La Brea Elementary School will improve their mathematics growth equivalent to 1.5 times the average for their grade. Level 3 – (Proficient) Students in La Brea Elementary School will improve their mathematics growth to be equivalent to the average for their grade. Level 2 – (Marginal) Students in La Brea Elementary School will improve their mathematics growth relative to last year. Level 1 – (Unacceptable) Students in La Brea Elementary School do not improve their mathematics growth relative to last year.
  • 48. Is this goal attainable? 62% of students at John Glenn Elementary met or exceeded proficiency in Reading/Literature last year. Their goal is to improve their rate to 82% this year. Is the goal attainable? Oregon schools – change in Reading/Literature proficiency 2009-10 to 2010-11 among schools that started with 60% proficiency rates 400 300 200 100 362 351 291 173 0 Growth > -20% > -10% > 0% > -30% 73 14 3 > 10% > 20% > 30%
  • 49. Is this goal attainable and rigorous? 45% of the students at La Brea elementary showed average growth or better last year. Their goal is to improve that rate to 50% this year. Is their goal reasonable? Students with average or better annual growth in Repus school district 100% 80% 60% 40% 20% 0%
  • 50. The selection of metrics matters Students at LaBrea Elementary School will show growth equivalent to 150% of grade level. Students at Etsaw Middle School will show growth equivalent to 150% of grade level.
  • 51. Scale score growth relative to NWEA’s growth norm in mathematics Growth Index 7.0 Scale Score Growth 6.0 5.0 4.0 3.0 2.0 1.0 0.0 -1.0 2 3 4 5 6 7 8 9
  • 52. Percent of a Year’s Growth Percent of a year’s growth in mathematics Mathematics 200% 180% 160% 140% 120% 100% 80% 60% 40% 20% 0% 2 3 4 5 6 7 8 9
  • 53. Assessing the difficult to measure • Encourage use of performance assessment and rubrics. • Encourage outside scoring – Use of peers in other buildings, professionals in the field, contest judges • Make use of resources – Music educator, art educator, vocational professional associations – Available models – AP art portfolio. – Use your intermediate agency – Work across buildings • Make use of classroom observation.
  • 54. The outcome is important, but why the outcome occurred is equally important! Success can’t be replicated if you don’t know why you succeeded. Failure can’t be reversed if you don’t know why you failed.
  • 55. The outcome is important, but why the outcome occurred is equally important! • Establish checkpoints and check-in beyond what’s required when possible. • Collect implementation information – Classroom observations - visits – Teacher journal – Student work – artifacts These processes can be done by teacher peers!
  • 56. Thank you for attending Presenter - John Cronin, Ph.D. Contacting us: NWEA Main Number: 503-624-1951 E-mail: rebecca.moore@nwea.org