SlideShare uma empresa Scribd logo
1 de 3
Baixar para ler offline
COMMENTARY / COMMENTAIRE
Clinical and Scientific Considerations in Progress Monitoring: When Is a
Measure Too Long?
Barry L. Duncan
The Heart and Soul of Change Project, Jensen Beach, Florida
Robert J. Reese
University of Kentucky
In their reaction to Duncan (2012), Halstead, Youn, and Armijo (2013) pose a useful psychometric
question regarding how brief is too brief when considering progress measures. They suggest that
measures should be of sufficient length to provide reliability and validity but provide no definition of
what constitutes sufficient reliability or validity. Moreover, Halstead et al. overlook the important clinical
issue of feasibility, whether the measure will be routinely used by front line clinicians. We assert that
there is no doubt that the increased reliability and validity of longer measures likely result in better
detection, prediction, and ultimate measurement of outcome, but suggest that empirical investigation is
required to determine if these differences are clinically meaningful and offset the low compliance rates.
We also assert that while empirical investigation is required to determine how brief is too brief, the
answer to the question regarding when a measure is too long is simple: When clinicians won’t use it.
Keywords: PCOMS, reliability, validity, feasibility, ORS
In their thoughtful commentary in reaction to the special issue of
Canadian Psychology on progress tracking (2012, Vol. 53, No. 2),
and specifically to the article by Duncan (2012), Halstead, Youn,
and Armijo (2013) posed a useful psychometric question regarding
how brief is too brief when considering progress measures and
raised important issues relevant to the divide between research and
practice. They also presented a very interesting distinction con-
cerning existing measures, namely the Normative and the Com-
municative models.
Halstead et al. (2013) made several good points: Longer measures
of outcome are more reliable and valid than shorter measures; we
should use reliable measures that are sensitive to change; two mea-
sures are better than one; every session measurement allows more
complete information given the unpredictable nature of treatment
endings; and perhaps, measures that allow both normative and com-
municative feedback offer the best choice for clinical practice.
Halstead et al. (2013) concluded that measures should be of
sufficient length to provide reliability but provide no definition of
what constitutes sufficient reliability. They suggested that the
Outcome Rating Scale (ORS) has “relatively low alpha” (averag-
ing .85 across studies, but if the lowest [.79] is removed that used
an oral version in a telephonic service, the average is .87), and
performs “quite poorly” on concurrent validity, for example, with
the Outcome Questionnaire 45.2 (OQ; Lambert et al., 1996),
ranging from .53 to .74, but offer no comparison to a standard of
relatively high alpha or performing well on concurrent validity.
This lack of comparison is understandable given the absence of
agreed on thresholds for reliability and validity. The interpretations
offered by Halstead et al., however, are debatable. For example,
Cicchetti (1994) proposed that reliability estimates between .80
and .90 are “good” for clinical significance. Of course, others have
suggested more stringent criteria (Nunnally & Bernstein, 1994).
When it comes to the reliability and validity of measures, their
“beauty” is in the eye of the beholder. Not only do Halstead et al.
not specify what constitutes acceptable psychometrics but they
also don’t explain how the purported low alpha and poor concur-
rent validity translate to any meaningful clinical differences.
The two unpublished studies cited do little to support their psycho-
metric point given there are so many unknowns regarding how the
analyses were conducted, and they are not available for review. For
example, although they suggested that the sPaCE detected deteriora-
tion at different rates than the CORE, it is unknown which measure
achieved the “correct” rate or how their findings related to final
outcomes. Similarly, they reported that the ORS was less stable in
predicting change in the first five sessions than the sPaCE, but we are
without information regarding what clinical significance that made or
how it impacted ultimate outcome.
Halstead et al. (2013) have trouble with the rhetoric “in the real
world” and suggested that the divide between research and practice
need not exist. Perhaps it need not exist, but it does. Their com-
Barry L. Duncan, The Heart and Soul of Change Project, Jensen Beach,
Florida; Robert J. Reese, Department of Educational, School, & Counseling
Psychology, University of Kentucky.
Duncan is a coholder of the copyright of the Outcome Rating Scale/Session
Rating Scale family of measures. The measures are free for individual use, but
Duncan receives royalties from licenses issued to groups and organisations. In
addition, the Web-based system, MyOutcomes.com is a commercial product,
and he receives royalties based on sales.
Correspondence concerning this article should be addressed to Barry L.
Duncan, P.O. Box 6157, Jensen Beach, FL 34957. E-mail: barrylduncan@
comcast.net
Canadian Psychology / Psychologie canadienne © 2013 Canadian Psychological Association
2013, Vol. 54, No. 2, 135–137 0708-5591/13/$12.00 DOI: 10.1037/a0032362
135
mentary, in many ways, provides a ready explanation for the
divide. They suggest:
. . .we think that where it is possible to use more reliable measures that
give us normative information, we should attempt to use them. . .The
more reliable measures are better at detecting early improvement and
more importantly early deterioration and allow us to track change in
a scientifically meaningful manner. (p. 84)
Although “more reliable” is not defined, “where it is possible”
represents the crux of the issue, and it is where the division
between research and practice occurs. Or in other words, can the
science of measurement be feasible for everyday clinical use? The
brevity of the ORS, with its attending lower reliability and validity
(although we contend are far from unacceptable), makes a differ-
ence, because, as is news to no one on the front lines, and
especially in the public sector, the number of forms and other
oversight procedures has exploded. Few have the time to devote to
the repeated administration, scoring, and interpretation of lengthy
measures—feasibility is critical.
Low compliance rates are the most frequent result of longer
measures. For example, comparison of two similar sites, one
implementing the ORS and the other the OQ revealed a compli-
ance rate for the ORS of 86% at the end of one year, and despite
ongoing encouragement, the use of the OQ was just 25% (Miller,
Duncan, Brown, Sparks, & Claud, 2003). Furthermore, longer
measures often wind up being used as periodic or prepost mea-
sures, which result in poor data integrity, not representative of
actual practice. For example, a benchmarking study conducted in
a managed care setting requiring the 30-item OQ at the first, third,
fifth, and every fifth session thereafter lost over 55% of the data for
lack of two data points (Minami et al., 2008). Similarly, a study of
the CORE 34 resulted in only 9,703 clients with pre- and post
information from a database of over 33,000 (Stiles, Barkham,
Connell, & Mellor-Clark, 2008).
Measures that are perceived as too long by psychotherapists pre-
vent many from even considering monitoring outcome. For example,
in reaction to a managed care company’s introduction of a 30-item
outcome questionnaire, the New England Psychologist (Hanlon,
2005) reported that providers complained about its length and fre-
quent administration. The response by clinicians was so severe that
the State Psychological Association president said, “I have never seen
such negative reaction from providers” (Hanlon, 2005, p. 11).
Intimately related to feasibility is the issue of the utility of the
feedback—whether the measure has an intended clinical use to
improve the effectiveness of rendered services. Most outcome
measures are used primarily as prepost and/or periodic assessment
devices. Such instruments measure program effectiveness but are
not feasible to administer frequently, and therefore, they do not
provide real-time feedback for immediate treatment modification
before clients drop out or suffer a negative outcome—in short, they
are not clinical tools as much as they are management or oversight
tools. The ORS was designed as a clinical and outcome tool to
provide real-time feedback to improve the effectiveness of services
and as a way to measure outcomes.
Perhaps this speaks to the normative versus communicative
distinction made by Halstead et al. The communicative aspects of
the ORS are critical to enhancing outcomes, but the normative
aspects provide the credibility for the discussion. There are now
over 400,000 administrations of the ORS resulting in algorithms
for expected treatment response. It is not surprising that the tra-
jectories are not unlike those reported by other outcome measures.
There is an unfortunate lack of data from where most mental
health services are provided—in public behavioural health set-
tings. In many ways, the lack of available data in the real world
speaks to the very heart of the divide between research and
practice. Wolfe (2012), in a clever dialogue between his researcher
and practitioner selves, suggested that practical outcome tools for
everyday clinical practice, like the ORS, can serve to build the
bridge between research and practice.
There is no doubt that 45 items, 30 items, or even 19 items is
better than 4 items, and that the increased reliability and validity of
longer measures likely result in better detection, prediction, and
ultimate measurement of outcome. But how much better is the
reliability and validity and more important, how much better is the
detection, prediction, and ultimate measurement of outcome? Are
these differences clinically meaningful, and do they offset the low
compliance rates and resulting data integrity issues? These are the
questions that require empirical investigation to determine how
brief is too brief.
But when is a measure too long? The answer is simple: When
clinicians won’t use it.
Résumé
Dans leur réaction a` Duncan (2012), Halstead, Youn et Armijo
(2013) posent la question d’ordre psychométrique suivante :
qu’est-ce qui constitue une mesure trop courte pour l’évaluation du
progrès? Ils suggèrent que les mesures doivent être suffisamment
longues pour assurer fiabilité et validité, sans toutefois offrir de
définition de ce qui constitue une fiabilité ou une validité suff-
isante. En outre, Halstead et al. passent sous silence l’importante
question dans le domaine clinique qu’est la faisabilité, a` savoir si
la mesure sera couramment utilisée par les cliniciens de première
ligne. Nous affirmons qu’il ne fait aucun doute que la fiabilité et
la validité accrues de mesures plus longues donneront probable-
ment lieu a` de meilleurs résultats en ce qui a trait a` la détection, a`
la prédiction et a` l’évaluation finale des résultats, mais nous
suggérons qu’il faut avoir recours a` l’enquête empirique pour
déterminer si ces différences sont significatives sur le plan clinique
et si elles compensent les faibles taux de conformité. Nous affir-
mons de plus que s’il faut recourir a` l’enquête empirique pour
déterminer ce qui constitue une mesure trop brève, la réponse a` la
question « Quand une mesure est-elle trop longue? » est simple :
lorsque les cliniciens ne veulent pas l’utiliser.
Mots-clés : PCOMS, fiabilité, validité, faisabilité, ORS.
References
Cicchetti, D. V. (1994). Guidelines, criteria, and rules of thumb for eval-
uating normed and standardized assessment instruments in psychology.
Psychological Assessment, 6, 284–290. doi:10.1037/1040-3590.6.4.284
Duncan, B. (2012). The Partners for Change Outcome Management Sys-
tem (PCOMS): The Heart and Soul of Change Project. Canadian Psy-
chology/Psychologie canadienne, 53, 93–104. doi:10.1037/a0027762
Halstead, J., Youn, S. J., & Armijo, I. (2013). Scientific and clinical
considerations in progress monitoring: When is a brief measure too
brief? Canadian Psychology, 54, 83–85.
136 DUNCAN AND REESE
Hanlon, P. (2005). PacifiCare screening tool, policies raise concerns. New
England Psychologist, 13, 11–12.
Lambert, M. J., Hansen, N. B., Umphress, V., Lunnen, K., Okiishi, J.,
Burlingame, G. M., . . . Reisinger, C. (1996). Administration and scoring
manual for the OQ 45.2. Stevenson, MD: American Professional Cre-
dentialing Services.
Miller, S. D., Duncan, B. L., Brown, J., Sparks, J., & Claude, D. (2003).
The outcome rating scale: A preliminary study of the reliability, validity,
and feasibility of a brief visual analog measure. Journal of Brief Ther-
apy, 2, 91–100.
Minami, T., Wampold, B., Serlin, R., Hamilton, E., Brown, G., & Kircher,
J. (2008). Benchmarking the effectiveness of psychotherapy treatment
for adult depression in a managed care environment: A preliminary
study. Journal of Consulting and Clinical Psychology, 76, 116–124.
doi:10.1037/0022-006X.76.1.116
Nunnally, J. C., & Bernstein, I. H. (1994). Psychometric theory (3rd ed.).
New York, NY: McGraw-Hill.
Stiles, W. B., Barkham, M., Connell, J., & Mellor-Clark, J. (2008). Re-
sponsive regulation of treatment duration in routine practice in United
Kingdom primary care settings: Replication in a larger sample. Journal
of Consulting and Clinical Psychology, 76, 298–305. doi:10.1037/0022-
006X.76.2.298
Wolfe, B. E. (2012). Healing the research-practice split: Let’s start with
me. Psychotherapy, 49, 101–108. doi:10.1037/a0027114
Received February 14, 2013
Accepted February 21, 2013 Ⅲ
137CONSIDERATIONS IN PROGRESS MONITORING

Mais conteúdo relacionado

Mais procurados

Duncan & Sparks Ch 5 of Cooper & Dryden
Duncan & Sparks Ch 5 of Cooper & DrydenDuncan & Sparks Ch 5 of Cooper & Dryden
Duncan & Sparks Ch 5 of Cooper & DrydenBarry Duncan
 
Outcome Rating Scale (ORS)
Outcome Rating Scale (ORS)Outcome Rating Scale (ORS)
Outcome Rating Scale (ORS)Barry Duncan
 
Norway Feedback Project
Norway Feedback ProjectNorway Feedback Project
Norway Feedback ProjectBarry Duncan
 
GroupSRSValidationStudy
GroupSRSValidationStudyGroupSRSValidationStudy
GroupSRSValidationStudyBarry Duncan
 
ReeseUsherBowmanetetal
ReeseUsherBowmanetetalReeseUsherBowmanetetal
ReeseUsherBowmanetetalBarry Duncan
 
PCOMS: A Viable Quality Improvement Strategy for Public Behavioral Health
PCOMS: A Viable Quality Improvement Strategy for Public Behavioral HealthPCOMS: A Viable Quality Improvement Strategy for Public Behavioral Health
PCOMS: A Viable Quality Improvement Strategy for Public Behavioral HealthBarry Duncan
 
PCOMS as an Alternative to Psychiatric Diagnosis (Duncan, Sparks, & Timimi, 2...
PCOMS as an Alternative to Psychiatric Diagnosis (Duncan, Sparks, & Timimi, 2...PCOMS as an Alternative to Psychiatric Diagnosis (Duncan, Sparks, & Timimi, 2...
PCOMS as an Alternative to Psychiatric Diagnosis (Duncan, Sparks, & Timimi, 2...Barry Duncan
 
FootprintsCoupleTherapy
FootprintsCoupleTherapyFootprintsCoupleTherapy
FootprintsCoupleTherapyBarry Duncan
 
GettingBetterAtWhatWeDo
GettingBetterAtWhatWeDoGettingBetterAtWhatWeDo
GettingBetterAtWhatWeDoBarry Duncan
 
RFL Feedback Study
RFL Feedback StudyRFL Feedback Study
RFL Feedback StudyBarry Duncan
 
Duncan2013AUMasterTherapist
Duncan2013AUMasterTherapistDuncan2013AUMasterTherapist
Duncan2013AUMasterTherapistBarry Duncan
 
HafkenscheidDutchORS
HafkenscheidDutchORSHafkenscheidDutchORS
HafkenscheidDutchORSBarry Duncan
 
CooperStewartSparksBunting2013
CooperStewartSparksBunting2013CooperStewartSparksBunting2013
CooperStewartSparksBunting2013Barry Duncan
 
On becomingabettertherapist
On becomingabettertherapistOn becomingabettertherapist
On becomingabettertherapistBarry Duncan
 
TeachingAccountability
TeachingAccountabilityTeachingAccountability
TeachingAccountabilityBarry Duncan
 
DeSantisJacksonDuncanReese2016
DeSantisJacksonDuncanReese2016DeSantisJacksonDuncanReese2016
DeSantisJacksonDuncanReese2016Barry Duncan
 
Feedback informed treatment (fit) achieving(apa ip miller hubble seidel chow ...
Feedback informed treatment (fit) achieving(apa ip miller hubble seidel chow ...Feedback informed treatment (fit) achieving(apa ip miller hubble seidel chow ...
Feedback informed treatment (fit) achieving(apa ip miller hubble seidel chow ...Scott Miller
 

Mais procurados (20)

Duncan & Sparks Ch 5 of Cooper & Dryden
Duncan & Sparks Ch 5 of Cooper & DrydenDuncan & Sparks Ch 5 of Cooper & Dryden
Duncan & Sparks Ch 5 of Cooper & Dryden
 
Outcome Rating Scale (ORS)
Outcome Rating Scale (ORS)Outcome Rating Scale (ORS)
Outcome Rating Scale (ORS)
 
ORS Replication
ORS ReplicationORS Replication
ORS Replication
 
Norway Feedback Project
Norway Feedback ProjectNorway Feedback Project
Norway Feedback Project
 
GroupSRSValidationStudy
GroupSRSValidationStudyGroupSRSValidationStudy
GroupSRSValidationStudy
 
ReeseUsherBowmanetetal
ReeseUsherBowmanetetalReeseUsherBowmanetetal
ReeseUsherBowmanetetal
 
PCOMS: A Viable Quality Improvement Strategy for Public Behavioral Health
PCOMS: A Viable Quality Improvement Strategy for Public Behavioral HealthPCOMS: A Viable Quality Improvement Strategy for Public Behavioral Health
PCOMS: A Viable Quality Improvement Strategy for Public Behavioral Health
 
PCOMS as an Alternative to Psychiatric Diagnosis (Duncan, Sparks, & Timimi, 2...
PCOMS as an Alternative to Psychiatric Diagnosis (Duncan, Sparks, & Timimi, 2...PCOMS as an Alternative to Psychiatric Diagnosis (Duncan, Sparks, & Timimi, 2...
PCOMS as an Alternative to Psychiatric Diagnosis (Duncan, Sparks, & Timimi, 2...
 
FootprintsCoupleTherapy
FootprintsCoupleTherapyFootprintsCoupleTherapy
FootprintsCoupleTherapy
 
GettingBetterAtWhatWeDo
GettingBetterAtWhatWeDoGettingBetterAtWhatWeDo
GettingBetterAtWhatWeDo
 
Slone et al. 2105
Slone et al. 2105Slone et al. 2105
Slone et al. 2105
 
RFL Feedback Study
RFL Feedback StudyRFL Feedback Study
RFL Feedback Study
 
Duncan2013AUMasterTherapist
Duncan2013AUMasterTherapistDuncan2013AUMasterTherapist
Duncan2013AUMasterTherapist
 
HafkenscheidDutchORS
HafkenscheidDutchORSHafkenscheidDutchORS
HafkenscheidDutchORS
 
CooperStewartSparksBunting2013
CooperStewartSparksBunting2013CooperStewartSparksBunting2013
CooperStewartSparksBunting2013
 
JustSayNo
JustSayNoJustSayNo
JustSayNo
 
On becomingabettertherapist
On becomingabettertherapistOn becomingabettertherapist
On becomingabettertherapist
 
TeachingAccountability
TeachingAccountabilityTeachingAccountability
TeachingAccountability
 
DeSantisJacksonDuncanReese2016
DeSantisJacksonDuncanReese2016DeSantisJacksonDuncanReese2016
DeSantisJacksonDuncanReese2016
 
Feedback informed treatment (fit) achieving(apa ip miller hubble seidel chow ...
Feedback informed treatment (fit) achieving(apa ip miller hubble seidel chow ...Feedback informed treatment (fit) achieving(apa ip miller hubble seidel chow ...
Feedback informed treatment (fit) achieving(apa ip miller hubble seidel chow ...
 

Destaque

AUReplicationORSSRS
AUReplicationORSSRSAUReplicationORSSRS
AUReplicationORSSRSBarry Duncan
 
ClientDirectedWraparound
ClientDirectedWraparoundClientDirectedWraparound
ClientDirectedWraparoundBarry Duncan
 
The outcome of psychotherapy yesterday, today and tomorrow (psychotherapy in ...
The outcome of psychotherapy yesterday, today and tomorrow (psychotherapy in ...The outcome of psychotherapy yesterday, today and tomorrow (psychotherapy in ...
The outcome of psychotherapy yesterday, today and tomorrow (psychotherapy in ...Daryl Chow
 
OnBecoming2ndedChapter1
OnBecoming2ndedChapter1OnBecoming2ndedChapter1
OnBecoming2ndedChapter1Barry Duncan
 
Sparks & Duncan Ch 6 Cooper & Dryden
Sparks & Duncan Ch 6 Cooper & DrydenSparks & Duncan Ch 6 Cooper & Dryden
Sparks & Duncan Ch 6 Cooper & DrydenBarry Duncan
 
Do people fill out the SRS differently IF the therapist is in the room?
Do people fill out the SRS differently IF the therapist is in the room?Do people fill out the SRS differently IF the therapist is in the room?
Do people fill out the SRS differently IF the therapist is in the room?Scott Miller
 

Destaque (9)

AUReplicationORSSRS
AUReplicationORSSRSAUReplicationORSSRS
AUReplicationORSSRS
 
ClientDirectedWraparound
ClientDirectedWraparoundClientDirectedWraparound
ClientDirectedWraparound
 
TheoryofChange
TheoryofChangeTheoryofChange
TheoryofChange
 
Duncan2014
Duncan2014Duncan2014
Duncan2014
 
The outcome of psychotherapy yesterday, today and tomorrow (psychotherapy in ...
The outcome of psychotherapy yesterday, today and tomorrow (psychotherapy in ...The outcome of psychotherapy yesterday, today and tomorrow (psychotherapy in ...
The outcome of psychotherapy yesterday, today and tomorrow (psychotherapy in ...
 
OnBecoming2ndedChapter1
OnBecoming2ndedChapter1OnBecoming2ndedChapter1
OnBecoming2ndedChapter1
 
Sparks & Duncan Ch 6 Cooper & Dryden
Sparks & Duncan Ch 6 Cooper & DrydenSparks & Duncan Ch 6 Cooper & Dryden
Sparks & Duncan Ch 6 Cooper & Dryden
 
PCOMS at a Glance
PCOMS at a GlancePCOMS at a Glance
PCOMS at a Glance
 
Do people fill out the SRS differently IF the therapist is in the room?
Do people fill out the SRS differently IF the therapist is in the room?Do people fill out the SRS differently IF the therapist is in the room?
Do people fill out the SRS differently IF the therapist is in the room?
 

Semelhante a DuncanReese2013

Advanced Regression Methods For Single-Case Designs Studying Propranolol In ...
Advanced Regression Methods For Single-Case Designs  Studying Propranolol In ...Advanced Regression Methods For Single-Case Designs  Studying Propranolol In ...
Advanced Regression Methods For Single-Case Designs Studying Propranolol In ...Stephen Faucher
 
qualitative portfolio
qualitative portfolioqualitative portfolio
qualitative portfolioMathew Aspey
 
Reply DB5 w9 researchReply discussion boards 1-jauregui.docx
Reply DB5 w9 researchReply discussion boards 1-jauregui.docxReply DB5 w9 researchReply discussion boards 1-jauregui.docx
Reply DB5 w9 researchReply discussion boards 1-jauregui.docxcarlt4
 
Evidence Translation and ChangeWeek 7What are the common.docx
Evidence Translation and ChangeWeek 7What are the common.docxEvidence Translation and ChangeWeek 7What are the common.docx
Evidence Translation and ChangeWeek 7What are the common.docxturveycharlyn
 
Detecting flawed meta analyses
Detecting flawed meta analysesDetecting flawed meta analyses
Detecting flawed meta analysesJames Coyne
 
Research and Program.docx
Research and Program.docxResearch and Program.docx
Research and Program.docxaudeleypearl
 
Responding to new scientific objections to the ors
Responding to new scientific objections to the orsResponding to new scientific objections to the ors
Responding to new scientific objections to the orsScott Miller
 
Beyond measures and monitoring
Beyond measures and monitoringBeyond measures and monitoring
Beyond measures and monitoringScott Miller
 
Organizational Culture and Readiness.docx
Organizational Culture and Readiness.docxOrganizational Culture and Readiness.docx
Organizational Culture and Readiness.docx4934bk
 
Notes for question please no plag use references to cite wk 2 .docx
Notes for question please no plag use references to cite wk 2 .docxNotes for question please no plag use references to cite wk 2 .docx
Notes for question please no plag use references to cite wk 2 .docxcherishwinsland
 
PPT-Asking Resaerch Question.pptx
PPT-Asking Resaerch Question.pptxPPT-Asking Resaerch Question.pptx
PPT-Asking Resaerch Question.pptxssuser064316
 
Asking Resaerch Question.pptx
Asking Resaerch Question.pptxAsking Resaerch Question.pptx
Asking Resaerch Question.pptxssuser064316
 
DEBATE Open AccessComparative effectiveness research what.docx
DEBATE Open AccessComparative effectiveness research what.docxDEBATE Open AccessComparative effectiveness research what.docx
DEBATE Open AccessComparative effectiveness research what.docxedwardmarivel
 
Evidence based social work
Evidence based social workEvidence based social work
Evidence based social workStephen Webb
 
Evidence-Based Practice in PsychologyImplications for Resea.docx
Evidence-Based Practice in PsychologyImplications for Resea.docxEvidence-Based Practice in PsychologyImplications for Resea.docx
Evidence-Based Practice in PsychologyImplications for Resea.docxgitagrimston
 
Evidence-Based Practice in PsychologyImplications for Resea.docx
Evidence-Based Practice in PsychologyImplications for Resea.docxEvidence-Based Practice in PsychologyImplications for Resea.docx
Evidence-Based Practice in PsychologyImplications for Resea.docxSANSKAR20
 
710B_Akansha Vaswani & Diego Flores
710B_Akansha Vaswani & Diego Flores710B_Akansha Vaswani & Diego Flores
710B_Akansha Vaswani & Diego FloresDiego Flores
 

Semelhante a DuncanReese2013 (20)

Advanced Regression Methods For Single-Case Designs Studying Propranolol In ...
Advanced Regression Methods For Single-Case Designs  Studying Propranolol In ...Advanced Regression Methods For Single-Case Designs  Studying Propranolol In ...
Advanced Regression Methods For Single-Case Designs Studying Propranolol In ...
 
qualitative portfolio
qualitative portfolioqualitative portfolio
qualitative portfolio
 
Reply DB5 w9 researchReply discussion boards 1-jauregui.docx
Reply DB5 w9 researchReply discussion boards 1-jauregui.docxReply DB5 w9 researchReply discussion boards 1-jauregui.docx
Reply DB5 w9 researchReply discussion boards 1-jauregui.docx
 
Sample Of Research Essay
Sample Of Research EssaySample Of Research Essay
Sample Of Research Essay
 
Evidence Translation and ChangeWeek 7What are the common.docx
Evidence Translation and ChangeWeek 7What are the common.docxEvidence Translation and ChangeWeek 7What are the common.docx
Evidence Translation and ChangeWeek 7What are the common.docx
 
Course project ntr_5503
Course project ntr_5503Course project ntr_5503
Course project ntr_5503
 
Reeseetal2013
Reeseetal2013Reeseetal2013
Reeseetal2013
 
Detecting flawed meta analyses
Detecting flawed meta analysesDetecting flawed meta analyses
Detecting flawed meta analyses
 
Research and Program.docx
Research and Program.docxResearch and Program.docx
Research and Program.docx
 
Responding to new scientific objections to the ors
Responding to new scientific objections to the orsResponding to new scientific objections to the ors
Responding to new scientific objections to the ors
 
Beyond measures and monitoring
Beyond measures and monitoringBeyond measures and monitoring
Beyond measures and monitoring
 
Organizational Culture and Readiness.docx
Organizational Culture and Readiness.docxOrganizational Culture and Readiness.docx
Organizational Culture and Readiness.docx
 
Notes for question please no plag use references to cite wk 2 .docx
Notes for question please no plag use references to cite wk 2 .docxNotes for question please no plag use references to cite wk 2 .docx
Notes for question please no plag use references to cite wk 2 .docx
 
PPT-Asking Resaerch Question.pptx
PPT-Asking Resaerch Question.pptxPPT-Asking Resaerch Question.pptx
PPT-Asking Resaerch Question.pptx
 
Asking Resaerch Question.pptx
Asking Resaerch Question.pptxAsking Resaerch Question.pptx
Asking Resaerch Question.pptx
 
DEBATE Open AccessComparative effectiveness research what.docx
DEBATE Open AccessComparative effectiveness research what.docxDEBATE Open AccessComparative effectiveness research what.docx
DEBATE Open AccessComparative effectiveness research what.docx
 
Evidence based social work
Evidence based social workEvidence based social work
Evidence based social work
 
Evidence-Based Practice in PsychologyImplications for Resea.docx
Evidence-Based Practice in PsychologyImplications for Resea.docxEvidence-Based Practice in PsychologyImplications for Resea.docx
Evidence-Based Practice in PsychologyImplications for Resea.docx
 
Evidence-Based Practice in PsychologyImplications for Resea.docx
Evidence-Based Practice in PsychologyImplications for Resea.docxEvidence-Based Practice in PsychologyImplications for Resea.docx
Evidence-Based Practice in PsychologyImplications for Resea.docx
 
710B_Akansha Vaswani & Diego Flores
710B_Akansha Vaswani & Diego Flores710B_Akansha Vaswani & Diego Flores
710B_Akansha Vaswani & Diego Flores
 

Último

Mythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITWMythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITWQuiz Club NITW
 
Scientific Writing :Research Discourse
Scientific  Writing :Research  DiscourseScientific  Writing :Research  Discourse
Scientific Writing :Research DiscourseAnita GoswamiGiri
 
Blowin' in the Wind of Caste_ Bob Dylan's Song as a Catalyst for Social Justi...
Blowin' in the Wind of Caste_ Bob Dylan's Song as a Catalyst for Social Justi...Blowin' in the Wind of Caste_ Bob Dylan's Song as a Catalyst for Social Justi...
Blowin' in the Wind of Caste_ Bob Dylan's Song as a Catalyst for Social Justi...DhatriParmar
 
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...DhatriParmar
 
How to Fix XML SyntaxError in Odoo the 17
How to Fix XML SyntaxError in Odoo the 17How to Fix XML SyntaxError in Odoo the 17
How to Fix XML SyntaxError in Odoo the 17Celine George
 
Measures of Position DECILES for ungrouped data
Measures of Position DECILES for ungrouped dataMeasures of Position DECILES for ungrouped data
Measures of Position DECILES for ungrouped dataBabyAnnMotar
 
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptxQ4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptxlancelewisportillo
 
4.11.24 Poverty and Inequality in America.pptx
4.11.24 Poverty and Inequality in America.pptx4.11.24 Poverty and Inequality in America.pptx
4.11.24 Poverty and Inequality in America.pptxmary850239
 
ROLES IN A STAGE PRODUCTION in arts.pptx
ROLES IN A STAGE PRODUCTION in arts.pptxROLES IN A STAGE PRODUCTION in arts.pptx
ROLES IN A STAGE PRODUCTION in arts.pptxVanesaIglesias10
 
Mental Health Awareness - a toolkit for supporting young minds
Mental Health Awareness - a toolkit for supporting young mindsMental Health Awareness - a toolkit for supporting young minds
Mental Health Awareness - a toolkit for supporting young mindsPooky Knightsmith
 
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdfGrade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdfJemuel Francisco
 
Multi Domain Alias In the Odoo 17 ERP Module
Multi Domain Alias In the Odoo 17 ERP ModuleMulti Domain Alias In the Odoo 17 ERP Module
Multi Domain Alias In the Odoo 17 ERP ModuleCeline George
 
Reading and Writing Skills 11 quarter 4 melc 1
Reading and Writing Skills 11 quarter 4 melc 1Reading and Writing Skills 11 quarter 4 melc 1
Reading and Writing Skills 11 quarter 4 melc 1GloryAnnCastre1
 
ClimART Action | eTwinning Project
ClimART Action    |    eTwinning ProjectClimART Action    |    eTwinning Project
ClimART Action | eTwinning Projectjordimapav
 
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptx
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptxINTRODUCTION TO CATHOLIC CHRISTOLOGY.pptx
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptxHumphrey A Beña
 
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnv
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnvESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnv
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnvRicaMaeCastro1
 

Último (20)

Paradigm shift in nursing research by RS MEHTA
Paradigm shift in nursing research by RS MEHTAParadigm shift in nursing research by RS MEHTA
Paradigm shift in nursing research by RS MEHTA
 
Mythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITWMythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITW
 
Scientific Writing :Research Discourse
Scientific  Writing :Research  DiscourseScientific  Writing :Research  Discourse
Scientific Writing :Research Discourse
 
Blowin' in the Wind of Caste_ Bob Dylan's Song as a Catalyst for Social Justi...
Blowin' in the Wind of Caste_ Bob Dylan's Song as a Catalyst for Social Justi...Blowin' in the Wind of Caste_ Bob Dylan's Song as a Catalyst for Social Justi...
Blowin' in the Wind of Caste_ Bob Dylan's Song as a Catalyst for Social Justi...
 
INCLUSIVE EDUCATION PRACTICES FOR TEACHERS AND TRAINERS.pptx
INCLUSIVE EDUCATION PRACTICES FOR TEACHERS AND TRAINERS.pptxINCLUSIVE EDUCATION PRACTICES FOR TEACHERS AND TRAINERS.pptx
INCLUSIVE EDUCATION PRACTICES FOR TEACHERS AND TRAINERS.pptx
 
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...
 
How to Fix XML SyntaxError in Odoo the 17
How to Fix XML SyntaxError in Odoo the 17How to Fix XML SyntaxError in Odoo the 17
How to Fix XML SyntaxError in Odoo the 17
 
Measures of Position DECILES for ungrouped data
Measures of Position DECILES for ungrouped dataMeasures of Position DECILES for ungrouped data
Measures of Position DECILES for ungrouped data
 
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptxQ4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
 
4.11.24 Poverty and Inequality in America.pptx
4.11.24 Poverty and Inequality in America.pptx4.11.24 Poverty and Inequality in America.pptx
4.11.24 Poverty and Inequality in America.pptx
 
ROLES IN A STAGE PRODUCTION in arts.pptx
ROLES IN A STAGE PRODUCTION in arts.pptxROLES IN A STAGE PRODUCTION in arts.pptx
ROLES IN A STAGE PRODUCTION in arts.pptx
 
Mental Health Awareness - a toolkit for supporting young minds
Mental Health Awareness - a toolkit for supporting young mindsMental Health Awareness - a toolkit for supporting young minds
Mental Health Awareness - a toolkit for supporting young minds
 
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdfGrade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
 
Multi Domain Alias In the Odoo 17 ERP Module
Multi Domain Alias In the Odoo 17 ERP ModuleMulti Domain Alias In the Odoo 17 ERP Module
Multi Domain Alias In the Odoo 17 ERP Module
 
prashanth updated resume 2024 for Teaching Profession
prashanth updated resume 2024 for Teaching Professionprashanth updated resume 2024 for Teaching Profession
prashanth updated resume 2024 for Teaching Profession
 
Reading and Writing Skills 11 quarter 4 melc 1
Reading and Writing Skills 11 quarter 4 melc 1Reading and Writing Skills 11 quarter 4 melc 1
Reading and Writing Skills 11 quarter 4 melc 1
 
Mattingly "AI & Prompt Design: Large Language Models"
Mattingly "AI & Prompt Design: Large Language Models"Mattingly "AI & Prompt Design: Large Language Models"
Mattingly "AI & Prompt Design: Large Language Models"
 
ClimART Action | eTwinning Project
ClimART Action    |    eTwinning ProjectClimART Action    |    eTwinning Project
ClimART Action | eTwinning Project
 
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptx
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptxINTRODUCTION TO CATHOLIC CHRISTOLOGY.pptx
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptx
 
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnv
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnvESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnv
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnv
 

DuncanReese2013

  • 1. COMMENTARY / COMMENTAIRE Clinical and Scientific Considerations in Progress Monitoring: When Is a Measure Too Long? Barry L. Duncan The Heart and Soul of Change Project, Jensen Beach, Florida Robert J. Reese University of Kentucky In their reaction to Duncan (2012), Halstead, Youn, and Armijo (2013) pose a useful psychometric question regarding how brief is too brief when considering progress measures. They suggest that measures should be of sufficient length to provide reliability and validity but provide no definition of what constitutes sufficient reliability or validity. Moreover, Halstead et al. overlook the important clinical issue of feasibility, whether the measure will be routinely used by front line clinicians. We assert that there is no doubt that the increased reliability and validity of longer measures likely result in better detection, prediction, and ultimate measurement of outcome, but suggest that empirical investigation is required to determine if these differences are clinically meaningful and offset the low compliance rates. We also assert that while empirical investigation is required to determine how brief is too brief, the answer to the question regarding when a measure is too long is simple: When clinicians won’t use it. Keywords: PCOMS, reliability, validity, feasibility, ORS In their thoughtful commentary in reaction to the special issue of Canadian Psychology on progress tracking (2012, Vol. 53, No. 2), and specifically to the article by Duncan (2012), Halstead, Youn, and Armijo (2013) posed a useful psychometric question regarding how brief is too brief when considering progress measures and raised important issues relevant to the divide between research and practice. They also presented a very interesting distinction con- cerning existing measures, namely the Normative and the Com- municative models. Halstead et al. (2013) made several good points: Longer measures of outcome are more reliable and valid than shorter measures; we should use reliable measures that are sensitive to change; two mea- sures are better than one; every session measurement allows more complete information given the unpredictable nature of treatment endings; and perhaps, measures that allow both normative and com- municative feedback offer the best choice for clinical practice. Halstead et al. (2013) concluded that measures should be of sufficient length to provide reliability but provide no definition of what constitutes sufficient reliability. They suggested that the Outcome Rating Scale (ORS) has “relatively low alpha” (averag- ing .85 across studies, but if the lowest [.79] is removed that used an oral version in a telephonic service, the average is .87), and performs “quite poorly” on concurrent validity, for example, with the Outcome Questionnaire 45.2 (OQ; Lambert et al., 1996), ranging from .53 to .74, but offer no comparison to a standard of relatively high alpha or performing well on concurrent validity. This lack of comparison is understandable given the absence of agreed on thresholds for reliability and validity. The interpretations offered by Halstead et al., however, are debatable. For example, Cicchetti (1994) proposed that reliability estimates between .80 and .90 are “good” for clinical significance. Of course, others have suggested more stringent criteria (Nunnally & Bernstein, 1994). When it comes to the reliability and validity of measures, their “beauty” is in the eye of the beholder. Not only do Halstead et al. not specify what constitutes acceptable psychometrics but they also don’t explain how the purported low alpha and poor concur- rent validity translate to any meaningful clinical differences. The two unpublished studies cited do little to support their psycho- metric point given there are so many unknowns regarding how the analyses were conducted, and they are not available for review. For example, although they suggested that the sPaCE detected deteriora- tion at different rates than the CORE, it is unknown which measure achieved the “correct” rate or how their findings related to final outcomes. Similarly, they reported that the ORS was less stable in predicting change in the first five sessions than the sPaCE, but we are without information regarding what clinical significance that made or how it impacted ultimate outcome. Halstead et al. (2013) have trouble with the rhetoric “in the real world” and suggested that the divide between research and practice need not exist. Perhaps it need not exist, but it does. Their com- Barry L. Duncan, The Heart and Soul of Change Project, Jensen Beach, Florida; Robert J. Reese, Department of Educational, School, & Counseling Psychology, University of Kentucky. Duncan is a coholder of the copyright of the Outcome Rating Scale/Session Rating Scale family of measures. The measures are free for individual use, but Duncan receives royalties from licenses issued to groups and organisations. In addition, the Web-based system, MyOutcomes.com is a commercial product, and he receives royalties based on sales. Correspondence concerning this article should be addressed to Barry L. Duncan, P.O. Box 6157, Jensen Beach, FL 34957. E-mail: barrylduncan@ comcast.net Canadian Psychology / Psychologie canadienne © 2013 Canadian Psychological Association 2013, Vol. 54, No. 2, 135–137 0708-5591/13/$12.00 DOI: 10.1037/a0032362 135
  • 2. mentary, in many ways, provides a ready explanation for the divide. They suggest: . . .we think that where it is possible to use more reliable measures that give us normative information, we should attempt to use them. . .The more reliable measures are better at detecting early improvement and more importantly early deterioration and allow us to track change in a scientifically meaningful manner. (p. 84) Although “more reliable” is not defined, “where it is possible” represents the crux of the issue, and it is where the division between research and practice occurs. Or in other words, can the science of measurement be feasible for everyday clinical use? The brevity of the ORS, with its attending lower reliability and validity (although we contend are far from unacceptable), makes a differ- ence, because, as is news to no one on the front lines, and especially in the public sector, the number of forms and other oversight procedures has exploded. Few have the time to devote to the repeated administration, scoring, and interpretation of lengthy measures—feasibility is critical. Low compliance rates are the most frequent result of longer measures. For example, comparison of two similar sites, one implementing the ORS and the other the OQ revealed a compli- ance rate for the ORS of 86% at the end of one year, and despite ongoing encouragement, the use of the OQ was just 25% (Miller, Duncan, Brown, Sparks, & Claud, 2003). Furthermore, longer measures often wind up being used as periodic or prepost mea- sures, which result in poor data integrity, not representative of actual practice. For example, a benchmarking study conducted in a managed care setting requiring the 30-item OQ at the first, third, fifth, and every fifth session thereafter lost over 55% of the data for lack of two data points (Minami et al., 2008). Similarly, a study of the CORE 34 resulted in only 9,703 clients with pre- and post information from a database of over 33,000 (Stiles, Barkham, Connell, & Mellor-Clark, 2008). Measures that are perceived as too long by psychotherapists pre- vent many from even considering monitoring outcome. For example, in reaction to a managed care company’s introduction of a 30-item outcome questionnaire, the New England Psychologist (Hanlon, 2005) reported that providers complained about its length and fre- quent administration. The response by clinicians was so severe that the State Psychological Association president said, “I have never seen such negative reaction from providers” (Hanlon, 2005, p. 11). Intimately related to feasibility is the issue of the utility of the feedback—whether the measure has an intended clinical use to improve the effectiveness of rendered services. Most outcome measures are used primarily as prepost and/or periodic assessment devices. Such instruments measure program effectiveness but are not feasible to administer frequently, and therefore, they do not provide real-time feedback for immediate treatment modification before clients drop out or suffer a negative outcome—in short, they are not clinical tools as much as they are management or oversight tools. The ORS was designed as a clinical and outcome tool to provide real-time feedback to improve the effectiveness of services and as a way to measure outcomes. Perhaps this speaks to the normative versus communicative distinction made by Halstead et al. The communicative aspects of the ORS are critical to enhancing outcomes, but the normative aspects provide the credibility for the discussion. There are now over 400,000 administrations of the ORS resulting in algorithms for expected treatment response. It is not surprising that the tra- jectories are not unlike those reported by other outcome measures. There is an unfortunate lack of data from where most mental health services are provided—in public behavioural health set- tings. In many ways, the lack of available data in the real world speaks to the very heart of the divide between research and practice. Wolfe (2012), in a clever dialogue between his researcher and practitioner selves, suggested that practical outcome tools for everyday clinical practice, like the ORS, can serve to build the bridge between research and practice. There is no doubt that 45 items, 30 items, or even 19 items is better than 4 items, and that the increased reliability and validity of longer measures likely result in better detection, prediction, and ultimate measurement of outcome. But how much better is the reliability and validity and more important, how much better is the detection, prediction, and ultimate measurement of outcome? Are these differences clinically meaningful, and do they offset the low compliance rates and resulting data integrity issues? These are the questions that require empirical investigation to determine how brief is too brief. But when is a measure too long? The answer is simple: When clinicians won’t use it. Résumé Dans leur réaction a` Duncan (2012), Halstead, Youn et Armijo (2013) posent la question d’ordre psychométrique suivante : qu’est-ce qui constitue une mesure trop courte pour l’évaluation du progrès? Ils suggèrent que les mesures doivent être suffisamment longues pour assurer fiabilité et validité, sans toutefois offrir de définition de ce qui constitue une fiabilité ou une validité suff- isante. En outre, Halstead et al. passent sous silence l’importante question dans le domaine clinique qu’est la faisabilité, a` savoir si la mesure sera couramment utilisée par les cliniciens de première ligne. Nous affirmons qu’il ne fait aucun doute que la fiabilité et la validité accrues de mesures plus longues donneront probable- ment lieu a` de meilleurs résultats en ce qui a trait a` la détection, a` la prédiction et a` l’évaluation finale des résultats, mais nous suggérons qu’il faut avoir recours a` l’enquête empirique pour déterminer si ces différences sont significatives sur le plan clinique et si elles compensent les faibles taux de conformité. Nous affir- mons de plus que s’il faut recourir a` l’enquête empirique pour déterminer ce qui constitue une mesure trop brève, la réponse a` la question « Quand une mesure est-elle trop longue? » est simple : lorsque les cliniciens ne veulent pas l’utiliser. Mots-clés : PCOMS, fiabilité, validité, faisabilité, ORS. References Cicchetti, D. V. (1994). Guidelines, criteria, and rules of thumb for eval- uating normed and standardized assessment instruments in psychology. Psychological Assessment, 6, 284–290. doi:10.1037/1040-3590.6.4.284 Duncan, B. (2012). The Partners for Change Outcome Management Sys- tem (PCOMS): The Heart and Soul of Change Project. Canadian Psy- chology/Psychologie canadienne, 53, 93–104. doi:10.1037/a0027762 Halstead, J., Youn, S. J., & Armijo, I. (2013). Scientific and clinical considerations in progress monitoring: When is a brief measure too brief? Canadian Psychology, 54, 83–85. 136 DUNCAN AND REESE
  • 3. Hanlon, P. (2005). PacifiCare screening tool, policies raise concerns. New England Psychologist, 13, 11–12. Lambert, M. J., Hansen, N. B., Umphress, V., Lunnen, K., Okiishi, J., Burlingame, G. M., . . . Reisinger, C. (1996). Administration and scoring manual for the OQ 45.2. Stevenson, MD: American Professional Cre- dentialing Services. Miller, S. D., Duncan, B. L., Brown, J., Sparks, J., & Claude, D. (2003). The outcome rating scale: A preliminary study of the reliability, validity, and feasibility of a brief visual analog measure. Journal of Brief Ther- apy, 2, 91–100. Minami, T., Wampold, B., Serlin, R., Hamilton, E., Brown, G., & Kircher, J. (2008). Benchmarking the effectiveness of psychotherapy treatment for adult depression in a managed care environment: A preliminary study. Journal of Consulting and Clinical Psychology, 76, 116–124. doi:10.1037/0022-006X.76.1.116 Nunnally, J. C., & Bernstein, I. H. (1994). Psychometric theory (3rd ed.). New York, NY: McGraw-Hill. Stiles, W. B., Barkham, M., Connell, J., & Mellor-Clark, J. (2008). Re- sponsive regulation of treatment duration in routine practice in United Kingdom primary care settings: Replication in a larger sample. Journal of Consulting and Clinical Psychology, 76, 298–305. doi:10.1037/0022- 006X.76.2.298 Wolfe, B. E. (2012). Healing the research-practice split: Let’s start with me. Psychotherapy, 49, 101–108. doi:10.1037/a0027114 Received February 14, 2013 Accepted February 21, 2013 Ⅲ 137CONSIDERATIONS IN PROGRESS MONITORING