ingramlab.files.wordpress.com€¦  · web viewthe restructured clinical (rc) scales of the mmpi-2...

58
Running Head: PERSONALITY ASSESSMENT TRAINING Trends in Training and Trainee Competence in Personality Assessment across Health Service Psychology Doctoral Students: A Pilot Study Paul B. Ingram, PhD 1 Matthew R. Cribbet, PhD 2 Adam T. Schmidt 1 , PhD Texas Tech University 1 University of Alabama 2 Author Notes. Portions of this paper were presented at the 2018 MMPI-2-RF/MMPI-2/MMPI-A-RF/MMPI-A Research Symposium. Paul Ingram receives research funding from Pearson, the distributor of several assessment instruments reported on within this paper. Correspondence concerning this article should be addressed to Paul B. Ingram, PhD., Texas Tech University, 2810 18 th Street,

Upload: nguyendung

Post on 19-Aug-2019

220 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Trends in Training and Trainee Competence in Personality Assessment across Health Service

Psychology Doctoral Students: A Pilot Study

Paul B. Ingram, PhD1 Matthew R. Cribbet, PhD2 Adam T. Schmidt1, PhD

Texas Tech University1

University of Alabama2

Author Notes. Portions of this paper were presented at the 2018 MMPI-2-RF/MMPI-2/MMPI-A-

RF/MMPI-A Research Symposium. Paul Ingram receives research funding from Pearson, the

distributor of several assessment instruments reported on within this paper. Correspondence

concerning this article should be addressed to Paul B. Ingram, PhD., Texas Tech University,

2810 18th Street, Lubbock, TX 79423. Email: [email protected] Phone: (806) 834-3354

© 2019, American Psychological Association. This paper is not the copy of record and

may not exactly replicate the final, authoritative version of the article. Please do not copy

or cite without authors' permission. This article is in press with Training and Education in

Professional Psychology. The final article will be available, upon publication, via its DOI:

10.1037/tep0000249 

Page 2: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

PERSONALITY ASSESSMENT TRAINING 2

Abstract

This investigation surveyed students (n = 91) from 16 American Psychological Association

accredited doctoral programs in Clinical and Counseling Psychology about knowledge and

training in personality assessment. We report self-perceived competency on specific instruments

as well as training trends in coursework and instrument exposure in clinical settings. We also

evaluate skill at interpretation on a popular personality instrument using two tasks, a narrative

interpretation where trainees estimate an originating score profile using a standardized

interpretive report and a symptom probability task where trainees predict the likelihood of

symptoms based on a score profile. The Minnesota Multiphasic Personality Inventory-2 and

Personality Assessment Inventory were the most frequently trained and utilized and had the

highest self-perceived competence. When given interpretation tasks to evaluate assessment skills

using the MMPI-2-RF, trainee performance was variable and discrepant from a comparison

expert panel given the same tasks. Overall, our results suggest that there is a need for further, and

more comprehensive, study on competence and training according to the experience of trainees.

We note that there is variability across instruction on instrument use, exposure to instruments in

practice, and practical skill level. We highlight our findings across four conceptual areas and

discuss the implications for the observed trends: frequency of instrument exposure, trainee

beliefs of competence, trainee interpretation skills, and reporting of testing interpretation.

Public Significance Statement: This pilot study provide the first evaluation of assessment

training patterns as reported by health service psychology trainees. Results indicate that, in

general, trainee instrument exposure and use mirror those of practicing psychologists; however,

there is notable variability in how assessment results are reported and in trainee competency.

Page 3: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Additional research evaluating health service psychology training should utilize student

participants so that direct evaluations of competency and outcomes are possible.

Keywords: Personality, Psychological Assessment, Training, Competence

Page 4: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Trends in Training and Trainee Competence in Personality Assessment across Health Service

Psychology Doctoral Students: A Pilot Study

Assessment is an integral and consistent part of a psychologist’s work and professional

identity. It has been identified by the American Psychological Association (APA) as a critical

component of clinical training (Kaslow et al., 2004; Krishnamurthy et al., 2004; Rodolfa et al.,

2005; Rodolfa et al., 2013) and represents a core area needed for program accreditation (APA,

2017). Given its central role during training, it is unsurprising that a recent survey found that a

small majority (58%) of practicing psychologists report performing psychological assessments

(Norcross & Karpiak, 2012). In fact, practicing psychologists spend, on average, 24% of their

time engaged in some form of assessment (Wright et al., 2016).

Given the wide use of assessment in clinical practice and its cornerstone in professional

psychology, even training in specialized assessment has seen an increase (e.g., neuropsychology;

Ready & Veague, 2014) despite a historical tendency towards limited coverage during training

(Krishnamurthy et al., 2004; Pidano & Whitcomb, 2012). An increasing emphasis on training in

assessment has coincided with the development of a set of competencies set forth by the

Psychological assessment work group of the competencies conference (Krishnamurthy et al.,

2004). Personality assessment has been previously recognized as a health service proficiency

within the field and is currently under review for renewal (Commission for Recognitions of

Specialties and Proficiencies in Professional Psychology, 2014). The Society of Personality

Assessment (SPA) (2006) has also identified key domains that should be covered during

assessment training. This includes having two or more courses in graduate education (which

includes coverage of psychometric theory, instrument selection and use, and data integration) as

well as supervised practice. However, training components are vaguely described and primarily

Page 5: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

serve to reinforce the role of doctoral training in assessment competence without providing clear

competency benchmarks for such training (SPA, 2006). In turn, assessment training practices,

particularly within personality assessment, need further advancement (Kaslow & Egan, 2017)

and are understudied (Smith, 2017). Accordingly, internship directors view only a minority of

their training candidates as having sufficient experience for the critical assessment practice of

report writing (Ready, Santorelli, Lundquist, & Romano, 2016; Stedman, Hatch, & Schoenfeld,

2001). While general domains for a competence framework have been identified (e.g.,

knowledge, skills, and attitudes about their integration; Kaslow, 2018), little has been done to

evaluate outcomes on the fusion of training program components and those of the field. Domains

of needed competence in assessment exist (APA, 2017; SPA, 2006); however, they are often

vague and describe students needing competency without describing what competency entails.

This vagueness allows for latitude in training and subsequent variability in outcomes.

Presently, personality assessment lacks a framework for evaluating competency (Kaslow,

Finklea, & Chan, 2018) and still needs conceptual frameworks through which skills are taught

and conceptual knowledge is used to inform treatment in a unified manner (Blais & Hopwood,

2017). Accordingly, quality and consistency of personality assessment training is difficult to

evaluate reliably and may be declining relative to other areas of assessment (e.g.,

neuropsychological testing, self-report measures, etc.; Ready & Veague, 2014). Fortunately,

there is some work being done to evaluate assessment training practices. However, that work is

based on reports by training directors about what is typically offered, required, or obtained as

part of the program of study and is not based on trainee responses (e.g., Mihura, Roy, Graceffo,

2017; Ready & Veague, 2014).

Page 6: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

While part of the difficulty with promoting assessment competence may relate to

insufficient training exposure (Cook et al., 2017), another issue may be in how test performance

is evaluated and communicated in reports (Cox, Cox, & Caplan, 2013). Some research has been

done to improve and standardize assessment reporting and interpretation; however, this work has

focused almost exclusively on neuropsychological assessment. For example, Guilmette, Hagan,

and Giuliano (2008) noted that when board-certified neuropsychologists were asked to assign

descriptive (qualitative) labels to test scores representing cognitive functioning, there was

substantial variability among individuals. They also found that roughly 33% of

neuropsychological reports excluded any form of explicit score information (i.e., raw item

endorsement, scaled/standardized score, or percentile rank) which made interpretation more

prone to error. Likewise, Schoenberg and Rum (2017) described a pattern of inconsistent use of

qualitative descriptors due to different score interpretation approaches. While some variability is

expected because of the debate between what constitutes pathological and normal functioning

(e.g., Axelrod, & Wall, 2007; Binder, Iverson, & Brooks, 2009), variation in interpretation can

lead to different diagnoses and treatment recommendations (Hagan & Guilmette, 2015). Thus,

despite some efforts to increase awareness of how results are reported and to promote a more

standardized approach, there remains limited work in this area. Moreover, the work that has been

done has focused exclusively on the areas of cognition and memory using samples of

neuropsychologists who have long been in practice.

While work in the interpretation standardization for neuropsychology is helpful to

promoting the assessment proficiency, the guidance offered for that form of testing does not

translate well to the practice of personality measurement. The types of diagnostic labels and

descriptions common to neuropsychological reports (e.g., consistent with IQ, intact, grossly

Page 7: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

intact, borderline impaired; Guilmette et al., 2008) vary from how test scores are interpreted in

personality testing, where clinicians are usually attempting to ascertain the likelihood of a given

set of behaviors, sensations, emotions, or perceptions. Thus, not only is it unclear what

constitutes standard training practices for personality assessment, there are no standard

interpretive schemas to guide students as they develop skills in this critical area of practice. Not

having a standardized basis for the interpretation, or even an understanding for what is a typical

skillset for doctoral level trainees, is concerning since most internship training programs

(frequently the final required formal training stop before a psychologist can engage in

independent practice) consider personality assessment as an essential competency for trainees to

develop (Stedman et al., 2017). It is difficult to ensure knowledge is measured reliably: (1) if

standard training practices are unknown, (2) if outcome expectations are inconsistent or poorly

defined, or (3) if outcomes and perceptions of trainees themselves are not assessed.

This pilot study examines assessment training practices and outcomes in a sample of

APA-approved clinical and counseling psychology doctoral programs. Specifically, we assess

exposure to personality instruments during doctoral training, perceptions of competence, testing

data reporting practices, and personality assessment interpretation skill compared to a panel of

expert raters. We also examine what information trainees believe should be included in

psychological reports and how frequently supervisors have directed them to include certain

information (e.g., raw scores, scaled scores, percentile, and qualitative or narrative descriptions).

Such an investigation is warranted because it develops the awareness of current training practices

for personality assessment and thus offers a way to improve educational standards.

Method

Participants

Page 8: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Survey invitations were sent to the directors, as well as associate directors when

available, of clinical training (DCT and ADCT, respectively) at 16 APA-approved doctoral

(Ph.D.) training programs in clinical and counseling psychology (8 = Clinical, 8 = Counseling).

These invitations requested that the DCT/ADCT forward a recruitment e-mail to their students,

requesting that individuals at each of the programs participate in a brief survey on personality

assessment training practices. In exchange for their participation, each graduate student was

awarded a five-dollar gift card to Amazon. Seven programs (2 Counseling and 5 Clinical; 44% of

the sample) DCT/ADCT had prior professional relationships with one of the authors. The

remaining 9 programs were identified based on their similarity in focus (e.g., clinical or

counseling), perspective (e.g., scientist practitioner or clinical science), and training outcomes to

the convenience sample. They were also intentionally targeted in order to increase the

geographic diversity of the programs and trainees sampled (e.g., multiple programs were targeted

for recruitment from the southeast, southwest, northeast, and midwest regions).

In general, the programs targeted were drawn from a variety of geographic regions

(programs in nine states from the east to west coast) and, as of the summer of 2018, represent

approximately 5% of all APA-accredited health service programs and 6.5% of non-combined

type PhD programs. Programs training models were scientist-practitioner (n = 13; 81.3%) and

clinical science training models (n = 3; 18.8%). The programs sampled are also at or above the

national mean on major markers of successful student training (e.g., accredited internship match

rate, Examination of Professional Practice in Psychology [EPPP] pass and licensure rate). Many

were also recognized by outside educational groups as providing particularly strong heath service

psychology training, as noted by program website descriptions. Thus, the programs sampled are

likely representative of training experiences across the country amongst high quality programs in

Page 9: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

health service psychology. The final sample of the study included 91 trainee participants (Table

1). Of those who did not complete the entire survey, dropout occurred almost exclusively when

participants were asked to complete assessment interpretation tasks.

Survey and Analysis Plan

The survey contained questions about basic demographic information. The survey also

included questions about the number of formal (i.e., standardized graduate coursework) and

informal (i.e., CE workshops, practicum, internship, etc.) training experiences in assessment. We

also assessed participants’ opportunity to use specific personality measures in clinical practice

during training, exposure to specific personality instruments in both formal and informal

training, number of written reports, and semesters of practicum. Information was also gathered

about how students have been trained to incorporate personality assessments results into reports.

We examined self-assessed competency with instruments on which trainees had exposure.

Competency on specific instruments was rated on a 0-100 scale with the following anchors (0 =

Not at all competent, 50 = Average competency, 100 = Extremely Competent). We also examined

performance-based competency by asking participants to complete several interpretation tasks.

Of personality assessment instruments, the Minnesota Multiphasic Personality Inventory

(MMPI) is the most frequently used in practice (Wright et al., 2017) and is covered in most

graduate training programs (Ready & Veague, 2014). The MMPI has also consistently been

described as inseparable from the process, practice, and history of testing (Benjamin, 2005;

Buchanan, 1994; Craik, 1986). Accordingly, clinicians are also likely to regularly encounter

reports that include the MMPI (Wright et al., 2016). The Restructured Clinical (RC) scales of the

MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Page 10: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

retention of historic conceptual pathology constructs (Ben-Porath, 2012,) and coverage in major

texts on the MMPI-2 (Ben-Porath & Tellegen, 2018).

Participants were asked to complete interpretation tasks using a score profile and a

qualitative narrative selected from a sample MMPI-2-RF interpretive report on the Pearson

website. They were not instructed to use, or not use, any materials for the task. Trainees were

instructed to assume that the respondent had provided a valid profile and that the elevations and

descriptors were valid indicators of the individual’s personality. Participants were asked to

estimate the probability of twenty-four different symptom and behavioral problems on a five-

point Likert-type scale (Extremely Likely to Extremely Unlikely) using an interpretive profile

composed of only the Higher Order (H-O) and Restructured Clinical (RC) scales (available as

supplemental material). The problem list utilized for this task was generated to reflect an array of

clinical concerns and was drawn from wording used to describe various MMPI-2-RF scales in

various interpretive texts (e.g., Ben-Porath, 2012; Ben-Porath & Tellegen, 2008/2011; Graham,

2016). Participants were also asked to estimate T-scores for the H-O and RC scales based on a

standardized interpretive report of the MMPI-2-RF. Next, a panel of three experts recruited from

the 2018 MMPI research symposium were asked to complete the same two tasks

(symptom/behavior likelihood and T-score estimation). Each expert rater has a lengthy history of

scholarship with the MMPI and are recognized broadly within the field. Trainee T-score

estimates were compared to both expert estimates and to the scale scores included in the sample

interpretive report. Participant estimates of symptom probability were compared to expert rating.

Results

Most participants endorsed needing (n = 57; 62.6%) and wanting (n = 79; 86.8%) more

training than they currently have. Of those who endorsed needing training, most indicated

Page 11: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

needing either a substantial (n = 13; 22.8%) or a moderate amount (n =34; 59.6%). A minority

endorsed needing either a great deal (n = 3; 5.3%) or minimal (n = 7; 12.3%) training.

Additionally, participants were asked how likely they were to include personality assessment in

their future career and indicated that they were: extremely likely (n = 17; 18.7%), moderately

likely (n = 29; 31.9%), slightly likely (n =25; 27.5%), slightly unlikely (n =9; 9.9%), moderately

unlikely (n =9; 9.9%), or extremely unlikely (n = 1; 1.1%). Approximately a third (n = 30;

33.0%) of participants had training in the administration and interpretation of projective /

implicit measures of personality. Only a portion (n = 52; 57%) of those who took the survey

completed the interpretive portion while most (87%) completed the survey up to the

interpretation task.

Independent t-tests indicated that no significant differences exist between trainees in

Clinical and Counseling psychology for semesters of clinical practicum or training, t(89) = -1.07,

number of integrative reports authored during training, t(89) = 0.98, number of formal or

informal personality assessment trainings received, t(89) = -0.84, wanting, t(89) = 1.93, or

needing, t(89) = 1.25, additional personality assessment training, or in program year, t(87)=

0.39. Non-significance of these dimensions suggests that assessment training patterns may be

analyzed across disciplines for subsequent analyses. An interclass correlation (ICC) assessing the

absolute agreement of expert raters averaged across measures supported the calculation of a

composite expert rater on the symptom probability task (ICC = .847) and the narrative

interpretation (ICC = .704).

Specific personality instrument training and use (Table 2) suggest that students received

the most training on the MMPI-2 and PAI. These personality inventories are also the most often

utilized instruments within doctoral training programs. The MMPI-2-RF was used slightly less

Page 12: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

frequently than the MMPI-2. Despite less frequent use, training on the MMPI-2-RF is received in

slightly more than half of doctoral training experiences. Trainees feel highly variable in their

competence about instrument use, including both those that they are trained in as well as those

they have no exposure to. A large portion of trainees endorsed a self-assessed competency level

below average, with many indicating that despite their training they felt not at all competent with

the instruments.

The survey asked participants about their use of different reporting standards (i.e., what

type of test information is conveyed in a written report) for personality testing results (Table 3).

Participants were most likely to apply qualitative labels during report writing and about half

included some form of normed comparison data; however, for the remaining portion of the

sample, the frequency of including that information varied. This suggests some variability in

reporting standards for personality instruments; however, raw score data were unlikely to have

been reported. These trends also reflect beliefs held by respondents about what is appropriate for

testing result reporting. Approximately a quarter of the sample believed each of the following

about which scales should be reported within a testing report: all scales on a personality measure

(n = 14; 15.4%), only clinically elevated scales (n = 19; 20.9%), relevant scales to the referral

question (n = 31; 34.1%), and only clinically elevated scales relevant to the referral question (n =

25; 27.5%).

When asked to estimate T-scores based on an interpretive report, trainees provided a

widely varied range of estimated scores and often incorrectly estimated if the content represented

in the report indicated a clinical elevation on its associated profile (Table 4). In contrast to

trainee performance, experts predicted T-scores that were generally consistent with the source

profile’s clinical elevations. A similar pattern of discrepancy was noted between both estimated

Page 13: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

scale scores and identified symptoms (Table 5) by trainees and experts given the same task.

Notably, there was a substantial rate of survey withdrawal that began during the skill assessment

portion. There we no significant differences in self-rated competence among those who

completed and those who dropped out, t(88) = .821, p = .40. Perceived competence was not

correlated with distance from report or expert scores for any of the estimated MMPI-2-RF scales.

Discussion

This study evaluated doctoral trainee exposure and competence with personality

assessment instruments using a sample of PhD students in Clinical and Counseling psychology.

This investigation was conducted to empirically evaluate the state of assessment as a critical

training component within the practice of health service psychology. Trainee competence in

assessment, particularly personality assessment, is understudied, particularly as it relates to the

use of assessment instruments and training. While previous research has evaluated perception of

trainee competency using training director perception, this study is the first to explore trainee

perceptions of competencies directly. Results indicate four distinct trends which are worthy of

further evaluation and are discussed separately below: frequency and exposure to personality

instruments during doctoral training, perceptions of competence, evidence of learned assessment

skills as measured on performance-based tasks, and implications for test reporting and diagnostic

conceptualizations.

Instrument use frequency

One goal of this study was to identify doctoral training trends for personality assessment

instruments within a sample of health service psychology PhD programs. Importantly, these

training trends appear generally consistent with patterns of use by practicing psychologists and

reports of training directions (e.g. Wright et al., 2017), with only a few exceptions. Such

Page 14: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

similarity supports the generalizability of this study on trainee experiences. Trainees receive less

frequent training exposure on the MCMI while students receive more training on the PAI,

relative to professional report of use (Wright et al., 2017). PAI training exposure is high, which

is consistent with rates observed when asking training faculty (Mihura et al., 2017). However,

reported exposure to the Rorschach is less than in use of practicing psychologists (53.6%; Wright

et al., 2016) and by faculty report of required trainee experience (61%; Mihura et al., 2017).

Unlike other personality assessment measures, students demonstrated a strong basis of training

experience on the MMPI, consistent with its reported use (Camara, Nathan, & Puente, 2000;

Mihura et al., 2017; Wright et al., 2016). Even rate of clinical exposure with specific forms of the

MMPI (e.g., MMPI-2, MMPI-2-RF) are relatively consistent with national sales figures (e.g., the

MMPI-2 and MMPI-2-RF account for approximately 60% and 40% of sales, respectively; Ben-

Porath, 2017). Thus, while training experiences and exposure may differ for many instruments,

the MMPI (regardless of version) remains a consistent and central component of personality

assessment training (Benjamin, 2005; Buchanan, 1994; Craik, 1986).

Although results are generally consistent with previous research using training director’s

report, there are some notable variations. The variation in exposure to specific assessment

instruments may function in part because of sampling. For instance, differences in Rorschach

exposure is likely reflective of training program emphasis in either objective or

projective/implicit techniques. In this and other studies of training programs, research has

included only a minority of existing doctoral programs (see Mihura et al., 2017). Given the

general similarity between rates observed here and through faculty report across available

assessment instruments, this study captures a generalized trend of assessment experience. We

believe it is necessary for variability in instruction and training exposure to decline so that there

Page 15: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

is a more explicit assessment training standard for all psychologists. Program efforts to meet

more explicit training exposure guidelines are critical if we are to effectively (and universally)

define the knowledge domain for an assessment competency for psychologists (Kaslow & Egan,

2017). Implicitly this often occurs already, as instructors frequently select instruments for

courses which will prepare trainees for the world of work. Efforts to ensure instrument exposure

which mirrors use does not, however, supplant a need for standardized competence assessment

and merely reflects an opportunity to ensure trainees are adequately prepared for what is

commonly utilized in practice.

Attitudes of Trainee Competence

While training and exposure increases self-assessment of competence, there is notable

variation in the level of self-perceived competence. This variability is evident for both those who

are trained with an instrument as well as for those who are not. We are concerned with the large

range of self-assessed competence. These ranges are problematic as they suggest that many of

those with instrument training exposure are not comfortable in their use and that those without

training may, at the same time, see themselves as extremely competent. This is concerning as it

suggests that those without training in an instrument who see themselves as increasingly

competent may be vulnerable to practicing outside their scope of practice. Such impacts of

problematic over-confidence are well-documented, particularly amongst those with less ability,

and lead to a poorer capacity to determine when they are competent (Kruger & Dunning, 1999).

This may lead to trainees without sufficient assessment competence engaging in poorly

standardized, and perhaps even harmful practice / violations of the ethical code of conduct,

without being aware of it. Promoting trainee awareness of the gap between their perceived and

Page 16: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

actual performance may help to offset this cognitive bias (Dunning, Johnson, Ehrlinger, &

Kruger, 2003).

Interpretation Skills

In this study we piloted the performance-based competence of common core scales that

are part of the MMPI, the most widely used personality assessment instrument. While training

led to higher perceived competence in general, it did little to attenuate variability in competence

scores. When trainees were asked to demonstrate their assessment competency on that

instrument using skill-based interpretive tasks, there were several patterns suggesting lower skill

levels than desired or expected if perceived competency ratings are believed accurate. As such,

variations in self-assessment of competency mirrored the difficulties observed during the

interpretive tasks. That is, perceptions of competency do not appear to translate well into skill-

based assessment tasks. While this may vary across different assessment instruments, the tasks

presented are drawn from a widely used instrument (Mihura et al., 2017; Wright et al., 2016) and

the scales are a core interpretive component that receives training attention (Ben-Porath &

Tellegen, 2018). In general, trainees displayed variability in the degree to which they interpreted

written narratives and estimated symptom probability. Not only were there large amounts of

variability exceeding what is considered clinically significant differences (Rosenthal, Rosnow, &

Rubin, 2000), indications of clinical severity were frequently incorrectly classified. For example,

estimated t-scores have response ranges which varied wildly when compared to the scale scores

associated with the presented written narrative and to expert ratings of that same written

narrative.

The above-mentioned trends need, of course, to be repeated in a national sample of

training programs and use other common assessment instruments to triangulate the broader state

Page 17: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

of assessment competence. However, variability in interpretation skill and participant drop-out

signals to us a potential danger of the current state of how competency is defined. As such, we

believe it is important to promote further study about how curriculum and practicum exposure

translate into generalized practical knowledge. For assessment to emphasize its evidence-based

nature with the greatest impact, interpretation standards should be studied, and this study

emphasizes a need to do so using trainees themselves.

Techniques such as the Q-sort have been described before as providing promise in

training practices (Weed, 2006). It would be beneficial if instructional aids like the Q-sort (e.g.,

those where clear interpretative competency levels that may be set at a national level) were

disseminated and utilized in assessment courses. Such efforts would aid in the standardization of

training outcomes, assist in establishing a consistent basis for skill competency, and promote a

stronger assessment proficiency.

Interpretive Standards

Qualitative labels are the most frequent means of communicating personality testing

results. One need look no further than common training textbooks to see that the field frequently,

and primarily, emphasizes the exclusive use of these labels. Likewise, the panel of experts in this

study also uniformly agreed that quantitative scores should be excluded in lieu of descriptive

labels. Likewise, while some trainees reported supervisors guiding them to include other sources

of data in reports (scaled scores, percentiles, etc.), qualitative labels were the most consistent

way to communicate testing results. This is of curiosity, as such descriptive methods offer

objectively less transparent means of communicating test performance compared to methods

frequently utilized during most other types of testing (e.g., memory, intelligence, aptitude, etc.).

Frequently, the written report (and not the raw data) may be the only information a psychologist

Page 18: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

may have about formal testing and they may wish to compare previous psychological

functioning to current test data. Lower transparency about respondent performance in reports has

implications for how effectively we can communicate with other professionals and may explain

some of the variable performance seen on the skill-based competence tasks. Much as language

for describing the qualitative labels is interpreted across a wide range, it is likely that those not

using interpretive report software vary in the language they use – further increasing room for

interpretive mistakes.

We are unaware of any research that offers empirically-based guidance of objective

assessment reporting standards (e.g., probabilistic language recommended for score ranges or

even what type of reporting information to include, such as qualitative summary statements or

standardized T-scores). Aside from interpretive tradition, we are unaware of any studies which

indicate that the reporting of personality instrument results should include clear indications of

respondent scores to reduce the variability surrounding how descriptive language is interpreted

(Cox et al., 2013). Further study is needed to examine how qualitative labels are applied within

personality assessment, as well as the impacts of including standardized reporting data. This is an

area that has already received some attention in intellectual and neuropsychological assessment

(Guilmette et al., 2008).

If there continues to be evidence that qualitative descriptions (i.e., typically thought of

within personality assessment as a summary of empirically-based correlates) alone are associated

with variation in test interpretation, one method to reduce variability that might decrease that

problem is to shift what is traditionally included in testing reports. For instance, the standardized

inclusion of observed scores (T-scores) for scales utilized during personality interpretation

(perhaps in the form of a table within the report, as is frequently done within neuropsychological

Page 19: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

testing) may lend itself to less interpretive ambiguity. Such a practice could supplement current

test reports that include only narrative interpretations. While consumers are a frequent recipient

of reports and may be unclear how an assessor goes from a score to an interpretive statement, we

believe that including standardized testing scores (whether in an appendix or in a table within the

main report document) remains beneficial. Such information is also regularly included in other

assessments specialties (e.g., neuropsychology and rehabilitation psychology). While including

standardized interpretive data does not ensure a lack of variation or error in interpretation (see

Cox et al., 2013), such a step may facilitate improved interpretation by other qualified providers

who encounter the report.

As a counter-perspective provided by an anonymous reviewer, the inclusion of

standardized testing data into the report may be problematic because interpretation requires

integration of not only test scores but also contextual and individual variables that may influence

test response. As such, this may lead to a test-based, rather than a client-based assessment,

approach. However, we believe that for a test to be utilized it must be validated on the

populations in which it is used. Comparisons between client scores and context-appropriate

comparison groups should therefore be incorporated as part of the report to help curtail potential

for misrepresentation and labeling. Likewise, adaptations based on contextual and individual

factors should be empirically based and clearly justified within the report. Thus, it is our belief

that more information, rather than less, is likely to reduce inappropriate test and report

interpretations. Studies on the influence of various testing information (including respondent

cultural variables) on interpretation accuracy and bias would be helpful in providing guidance on

what should be included in reports. Likewise, it may also be useful to expand this study to other

reporting trends used within the field (e.g., base rates).

Page 20: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Conclusions and Future Directions

This study builds on existing literature and offers the first evaluation of trainee

experience and performance-based competency within assessment. Results are, in many ways,

consistent with previous studies that have outlined assessment instrument exposure during

training (Mihura et al., 2017) and highlighted ways in which trainees are not as prepared to

perform assessment as they may believe or as might be desired by the field (Ready et al., 2016).

In general, there is variability across instruction on instrument use, exposure to instruments in

practice, and practical skill level for the personality instrument evaluated within this study. This

variability offers both areas in which the assessment competency must grow (e.g., decreased

variability in training instrument exposure, increased consistency in outcomes for training

experiences, etc.) and some ways in which that might occur (e.g., inclusion of standardized

respondent score information).

No study is without limitations, however. For instance, this pilot study does not include a

national sample comprising doctoral students from all programs or program types and that some

program factors (such as program training model which differ on required amount of assessment

courses; Ready & Veague, 2014) were, accordingly, unable to be evaluated. Likewise, this study

does not contain PsyD trainees which represent a growing portion of the training landscape.

Despite these limitations, this study provides a novel, trainee-centered evaluation of competency

that is likely to inform future understanding and study of assessment training and trainee

competence. The programs that were approached, and whose students participated, represent a

variety of different training models across health service psychology (e.g., Clinical and

Counseling as well as Scientist-Practitioner and Clinical Science), with those PsyD programs and

scholar-practitioner programs unrepresented within this sample. Likewise, all programs included

Page 21: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

in this study are at or above the national average on indicators of successful training outcomes

(e.g., APA-accredited internship match, EPPP pass and licensure rate, etc.). This, and the diverse

geographic regions from which the schools were recruited, leads us to believe that these results

are likely representative of many doctoral training programs in Clinical and Counseling

psychology and offer the first direct examination of trainee experiences in assessment training.

Therefore, we believe these data offer a reasonable starting point upon which a broader literature

can be developed.

An additional criticism that may be leveled against this study is the way in which

interpretation tasks were utilized to assess competence (e.g., tasks include only one narrative

report, this study assesses the skill of translating narrative description to symptom scores – which

is not trained, etc.). Conversely, it is our position that if clinicians can meaningfully interpret

scores and generate a report (which, we assume, has a purpose of communicating severity of

symptom distress) then that communication (and thus the interpretation of symptom distress)

should be clearest to those training in assessment. This is, in part, why we advocate for more

transparent reporting standards as it is likely to help facilitate communication of symptom levels

and personality testing results. We believe that this is needed as even when scores are displayed

there is notable discrepancy in symptom interpretation (Gilmet & Hagan, 2008). Future research

will, of course, need to expand the number, type, and complexity of interpretation tasks given so

that a fuller picture of competency is established. Such an effort will likely benefit training

programs as they prepare their students for the Enhanced Examination of Professional

Psychology.

Given the pilot nature of this study, findings may be best limited to programs most like

those sampled. In other words, results of this study are likely to be most similar to other

Page 22: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

programs which are APA-approved PhD programs in Clinical or Counseling (comprising

approximately 60% of all approved health service psychology programs) that have average (or

above) training outcomes. Accordingly, future research would benefit from a nationally

representative sample. Such an investigation would also benefit from additional performance-

based competency tasks and a wider array of assessment instruments evaluated (e.g., other

domains of assessment such as intellectual assessment). This wider array of instruments would

benefit from including those assessing both clinical and normal variations in personality. A

broader sample would also enable examination of diversity related differences in training

opportunity and outcome.

In general, we hope that this study inspires more conversation about training

standardization across the country as well as efforts to define and promote competency within

assessment. We also hope that such conversations more directly involve trainees and include

performance-based measures of competence. Accordingly, future research would benefit from

use of a larger, more representative and inclusive sample as well as a greater number of

interpretation tasks across additional assessment instruments.

Page 23: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

References

American Psychological Association. (2017). Standards of accreditation for programs in health

service psychology. Washington, DC: Author.

Axelrod, B.N., & Wall, J.R. (2007). Expectancy of impaired neuropsychological test scores in a

non-clinical sample. International Journal of Neuroscience, 117, 1591-1602. doi:

10.1080/00207450600941189

Benjamin, L.T., Jr. (2005). A history of clinical psychology as a profession in America (and a

glimpse at its future). Annual Review of Clinical Psychology, 1, 1-30,

10.1146/annurev.clinpsy.1.102803.143758.

Ben-Porath, Y.S. (2012). Interpreting the MMPI-2-RF. Minneapolis, MN. University of

Minnesota Press.

Ben-Porath, Y.S. (2017). An update to Williams and Lally’s (2017) analysis of MMPI-2-RF

acceptance. Professional Psychology: Research and Practice, 48, 275-278. doi:

10.1037/pro0000115

Ben-Porath, Y.S., & Tellegen, A. (2008/2011). Minnesota Multiphasic Personality Inventory–

Restructured form: Manual for administration, scoring, and interpretation. Minneapolis:

University of Minnesota Press.

Ben-Porath, Y.S., & Tellegen, A. (2018). Leone, Mosticoni, Ianella, Biondi, and Butcher’s

(2018) Effort to Compare the MMPI-2-RF with the MMPI-2 Falls Well Short. Archives

of Assessment Psychology, 8(1), 23-31.

Binder, L.M., Iverson, G.L., & Brooks, B.L. (2009). To err is human: Abnormal

neuropsychological scores and variability are common in healthy adults. Archives of

clinical Neuropsychology, 25, 31-46.

Page 24: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Blais, M.A., & Hopwood, C.J. (2017). Model-based approaches for teaching and practicing

personality assessment. Journal of Personality Assessment, 99, 136-145. doi:

10.1080/00223891.2016.1195393

Buchanan, R.D. (1994). The development of the Minnesota Multiphasic Personality

Inventory. Journal of the History of the Behavioral Sciences, 30, 148–

61. doi:10.1002/1520-6696(199404)30:23.0.CO;2-9.

Camara, W.J., Nathan, J.S., & Puente, A.E. (2000). Psychological test usage: Implications in

professional psychology. Professional Psychology: Research and Practice, 31, 141-151.

doi: 10.1037/0735-7028.31.2.141

Clemence, A. J., & Handler, L. (2001). Psychological assessment on internship: A survey of

training directors and their expectations for students. Journal of Personality Assessment,

76, 18–47.

Commission for the Recognition of Specialties and Proficiencies in Professional Psychology.

(2014). Principles for the Recognition of Specialties and Proficiencies in Professional

Psychology. Retrieved from http://www.apa.org/about/policy/recognition-principles.pdf

Cook, J.R., Hausman, E.M., Jensen-Doss, A., & Hawley, K.M. (2017). Assessment practices of

child clinicians: Results from a national survey. Assessment, 24, 210-221. doi:

10.1177/1073191115604353

Cox, D., Cox, R., Caplan, B. (2013). Specialty Competencies in Rehabilitation Psychology. New

York, New York: Oxford University Press

Craik, K.H. (1986). Personality research methods: An historical perspective. Journal of

personality, 54(1), 18-51.

Dunning, D., Johnson, K., Ehrlinger, J., & Kruger, J. (2003). Why people fail to recognize their

Page 25: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

own incompetence. Current Directions in Psychological Science, 12(3), 83-87.

Graham, J.R. (2016). MMPI-2: Assessing personality and psychopathology (5th ed.). New York:

Oxford.

Guilmette, T.J., Hagan, L.D., Giuliano, A.J. (2008). Assigning Qualitative Descriptions to Test

Scores in Neuropsychology: Forensic Implications. The Clinical Neuropsychologist, 22,

122-139. doi: 10.1080/13854040601064559

Hagan, L.D., & Guilmette, T.J. (2015). DSM-5: Challenging diagnostic testimony. International

Journal of Law and Psychiatry, 42, 128-134. doi: 10.1016/j.ijlp.2015.08.017

Kaslow, N. J., Borden, K. A., Collins, F. L., Forrest, L., Illfelder-Kaye, J., Nelson, P. D., &

Willmuth, M. E. (2004). Competencies conference: Future directions in education and

credentialing in professional psychology. Journal of Clinical Psychology, 80, 699–712.

doi:10.1002/jclp.20016

Kaslow, N.J., & Egan, G.J. (2017). A Competency-Focused Commentary on the Special Section

on Teaching, Training, and Supervision in Personality and Psychological

Assessment. Journal of Personality Assessment, 99, 189-191. doi:

10.1080/00223891.2016.1226177

Kaslow, N.J., Finklea, T., & Chan, G. (2018). Personality assessment: A competency-capability

perspective. Journal of Personality Assessment, 100, 176-185. doi:

10.1080/00223891.2017.1381970

Krishnamurthy, R., Vandecreek, L., Kaslow, N. J., Tazeau, Y. N., Milville, M. L., Kerns, R., &

Benton, S. A. (2004). Achieving competency in psychological assessment: Directions for

education and training. Journal of Clinical Psychology, 80, 725–740.

doi:10.1002/jclp.20010

Page 26: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Kruger, J., & Dunning, D. (1999). Unskilled and unaware of it: How difficulties in recognizing

one’s own incompetence lead to inflated self assessments. Journal of Personality and

Social Psychology, 77,

1121–1134

Mihura, J.L., Roy, M., Graceffo, R.A. (2017). Psychological assessment training in clinical

psychology doctoral programs. Journal of Personality Assessment, 99, 153-164. doi:

10.1080/00223891.2016.1201978

Norcross, J. C., & Karpiak, C. P. (2012). Clinical psychologists in the 2010s: 50 years of the

APA division of clinical psychology. Clinical Psychology: Science and Practice, 19, 1–

12. doi:10.1111/j.1468-2850.2012.01269.x

Pidano, A.E., & Whitcomb, J.M. (2012). Training to work with children and families: Results

from a survey of psychologists and doctoral students. Training and Education in

Professional Psychology, 6, 8-17. doi: 10.1037/a0026961

Ready, R. E., Santorelli, G. D., Lundquist, T. S., & Romano, F. M. (2016). Psychology

internship directors' perceptions of pre-internship training preparation in assessment.

North American Journal of Psychology, 18, 317.

Ready, R. E., & Veague, H. B. (2014). Training in psychological assessment: Current practices

of clinical psychology programs. Professional Psychology: Research and Practice, 45,

278-282. doi:10.1037/a0037439

Rodolfa, E. R., Bent, R. J., Eisman, E., Nelson, P. D., Rehm, L., & Ritchie, P. (2005). A cube

model for competency development: Implications for psychology educators and

regulators. Professional Psychology: Research and Practice, 36, 347–354.

doi:10.1037/0735-7028.36.4.347

Page 27: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Rodolfa, E., Greenberg, S., Hunsley, J., Smith-Zoeller, M., Cox, D., Sammons, M. T., & Spivak,

J. (2013). A competency model for the practice of psychology. Training and Education

in

Professional Psychology, 7, 71–83. doi:10.1037/a0032415

Rosenthal, R., Rosnow, R. L., & Rubin, D. B. (2000). Contrasts and effect sizes in behavioral

research: A correlational approach. New York, NY: Cambridge University Press.

Schoenberg, M.R., & Rum, R.S. (2017). Towards reporting standards for neuropsychological

study results: A proposal to minimize communication errors with standardized qualitative

descriptors for normalized test scores. Clinical Neurology and Neurosurgery, 162, 72-79.

doi:10.1016/j.clineuro.2017.07.010

Smith, J. D. (2017). Introduction to the special section on teaching, training, and supervision in

personality and psychological assessment. Journal of Personality Assessment, 99, 113–

116. doi:10.1080/00223891.2016.1217873

Society for Personality Assessment (SPA). (2006). Standards for education and training in

psychological assessment: Position of the Society for Personality Assessment. Journal of

Personality Assessment, 87(3), 355-357.

Stedman, J. M. (2007). What we know about predoctoral internship training: A 10-year update.

Training and Education in Professional Psychology, 1, 74–88. doi:10.1037/1931-

3918.1.1.74

Stedman, J. M., Hatch, J. P., & Schoenfeld, L. S. (2001). The current status of psychological

assessment training in graduate and professional schools. Journal of Personality

Assessment, 77, 398–407. doi:10.1207/S15327752JPA7703_02

Weed, N.C. (2006). Syndromal complexity, paradigm shifts, and the future of validation

Page 28: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

research: Comments on Nichols and Rogers, Sewell, Harrison, and Jordan. Journal of

Personality Assessment, 87, 217-222.

Wright, C.V., Beattie, S.G., Galper, D.I., Church, A.S., Bufka, L.F., Brabender, V.M., & Smith,

B.L. (2016). Assessment practices of professional psychologists: Results of a national

survey. Professional Psychology: Research and Practice, 48, 73-78. doi:

10.1047/pro0000086

Page 29: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Table 1.Demographic Information of study sample    Program Type

Sample CharacteristicSample (n = 91)

Clinical (n = 50)

Counseling (n = 41)

Age 27. 8 (3.7) 27.7 (4.0) 30.0 (3.4)Ethnicity - - -

Caucasian 73 (80.2%) 42 (84.0%) 31 (75.6%)Hispanic/Latinx 6 (6.6%) 3 (6.0%) 3 (7.3%)Asian American 6 (6.6%) 4 (8.0)% 2 (4.9%)

Native American 2 (2.2%) - 2 (4.9%)Other 4 (4.4%) 1 (2.0%) 3 (7.3%)

%Male 12 (13.2%) 4 (8.0%) 8 (19.5%)Highest Degree - - -

BA/BS 19 (20.9%) 14 (28.0%) 5 (12.2%)MA/MS 72 (79.1%) 36 (72.0%) 36 (87.8%)

Year in Program - - -1st 5 (5.5%) 3 (6.0%) 2 (4.9%)

2nd 14 (15.4%) 8 (16.0%) 6 (14.6%)3rd 23 (25.3%) 13 (26.0%) 10 (24.4%)4th 17 (18.7%) 4 (8.0%) 13 (31.7%)

5th or Beyond 21 (23.1%) 14 (28.0%) 7 (17.1%)Internship 8 (8.8%) 6 (12.0%) 2 (4.9%)

Integrated Reports 11.5 (14.2) 13.0 (13.9) 9.8 (14.5)Semesters of Practicum 6.3 (3.6) 6.0 (4.0) 6.7 (3.1)Completed Assessment Coursework - - -

Objective Personality 89 (97.8%) 49 (98.0%) 40 (97.6%)Projective Personality 10 (11.0%) 2 (4.0%) 8 (19.5%)

Neuropsychological 40 (44.0%) 32 (64.0%) 8 (19.5%)Child/Developmental 37 (40.7%) 29 (58.0%) 8 (19.5%)

School-based Assessment 10 (11.0%) 7 (14.0%) 3 (7.3%)Intellectual 100% 100% 100%

Other Trainings (Practicum, Workshop etc.) - - -Objective Personality 73 (80.2%) 43 (86.0%) 30 (73.2%)

Projective Personality 27 (29.7%) 13 (26.0%) 14 (34.1%)Neuropsychological 46 (50.5%) 32 (64.0%) 14 (34.1%)

Child/Developmental 43 (47.3%) 34 (68.0%) 9 (22.0%)School-based Evaluation 18 (19.8%) 13 (26.0%) 5 (12.2%)

Intellectual 71 (81.3%) 44 (88.0%) 30 (73.2%)

Page 30: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Table 2.Use and Perceived Competency of Specific Personality Assessment Instruments      Trained Competence Untrained Competence

Personality Instrument Has TrainingUsed

Clinically M (SD) Range M (SD) RangeMMPI-2 84 (92.3%) 62 (68.1%) 66.1 (18.9) 0 - 98 38.5 (26.9) 0 - 68

MMPI-2-RF 50 (54.9%) 37 (40.7%) 66.2 (19.6) 10 - 100 32.7 (27.2) 0 - 88MMPI-A 34 (37.4%) 24 (26.4%) 60.8 (22.8) 9 - 92 22.7 (23.9) 0 - 85

MMPI-A-RF 10 (11.0%) 8 (8.8%) 50.3 (28.8) 7 - 91 26.0 (25.6) 0 - 91PAI 79 (86.8%) 56 (61.5%) 71.5 (17.5) 18 - 100 18.8 (21.1) 0 - 61

MCMI-III 28 (30.8%) 23 (25.6%) 61.9 (22.9) 0 - 98 11.5 (16.7) 0 - 59MCMI-IV 19 (20.9%) 16 (17.6%) 63.8 (17.7) 18 - 96 16.4 (25.9) 0 - 96

Rorschach (Any System) 17 (18.7%) 13 (14.3%) 41.6 (30.9) 2- 87 6.3 (13.6) 0 - 60Note. Participant dropout resulted in only 85 completing competence questions while all 91 provided information on instrument use and training. All participants were asked to rate their competency on personality measures, regardless of training background. Competency was rated on a 0-100 scale with the following anchors (0 = Not at all competent, 50 = Average competency, 100 = Extremely Competent). Competence was calculated on only those with training in the instrument.

Table 3.Predictors of Perceived Competence  Predictor Variables  

InstrumentHighest Degree

Doctoral Type

Year in Program

# Integrated Reports

Semesters of Practice

Has Training

Used in Practice Model Result

MMPI-2 t = 1.40 t = -2.01* t = -.85 t = -2.18* t = -.72 t =3.81*** t = -4.16*** F(78, 7) = 6.37***, R = .60MMPI-2-RF t = 0.47 t = -0.93 t = 0.53 t = -0.46 t = -1.30 t =3.37*** t = 1.94 F (78, 7) = 7.42***, R = .63

MMPI-A t = -0.95 t = 1.38 t = -0.65 t = 1.12 t = 0.06 t =4.29*** t = -5.17*** F(75, 7) = 14.44***, R = .76MMPI-A-RF t = -0.65 t = -0.84 t = 0.23 t = -0.05 t = -1.02 t = -1.42 t = -3.14* F (67, 7) = 2.75***, R = .45

PAI t = 0.27 t = -2.15*** t = -0.33 t = 0.62 t = -1.92 t = 7.69*** t = -2.28* F (77, 7) = 19.54***, R = .80MCMI-III t = 0.13 t = 1.86 t = 1.88 t = 0.79 t = -1.25 t = -1.01 t = -7.53*** F (76, 7) = 11.65***, R = .72MCMI-IV t = -0.55 t = 1.33 t = 1.64 t = -0.19 t = -0.66 t = 3.29** t = -4.41*** F (76, 7) = 14.51***, R = .76

Rorschach t = 1.89 t = 1.964* t = -0.39 t = -0.16 t = -0.18 t = -0.17 t = -10.59*** F (76, 7) = 18.73***, R = .80Note. All participants, regardless of training exposure, were included within these analyses. p < .05*, p < .01**, p < .001***

Page 31: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Table 3.Percent of different types of data communicated within testing reports Frequency supervisors have required included in reportsData Type 0% 1-10% 11-25% 26-50% 51-75% 76-100%

Scaled Score / T-score 16 (18.2%) 5 (5.7%) 2 (2.2%) 8 (8.8%) 12 (13.2%) 45 (49.5%)Qualitative Label 1 (1.1%) 1 (1.1%) 2 (2.2%) 2 (2.2%) 8 (8.8%) 74 (81.3%)

Percentile 20 (22.0%) 6 (6.6%) 4 (4.4%) 7 (7.7%) 12 (13.2%) 39 (42.9%)Raw Score 60 (65.9%) 8 (8.8%) 3 (3.3%) 3 (3.3%) 5 (5.5%) 8 (8.8%)

Note. Percentages calculated based on validate percent responding to each question. Include in reports reflects the trainee belief that a type of data should be included as a standard component of a test report about a personality instrument.

Page 32: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Table 4.Certainty of a symptom-positive presentation based on T-scoresT-score Will have or does Is likely to Is unlikely to Will not or does not

45 0 1 (1.2%) 61 (75.3%) 19 (23.5%)50 3 (3.8%) 1 (1.3%) 57 (62.6%) 18 (19.8%)55 5 (6.3%) 9 (11.4%) 42 (53.2%) 21 (26.6%)60 9 (11.4%) 27 (34.2%) 24 (30.4%) 7 (8.9%)65 6 (7.5%) 41 (51.3%) 2 (2.5%) 2 (2.5%)70 9 (9.9%) 43 (47.3%) 1 (1.3%) 075 14 (17.5%) 36 (45.0%) 0 0

Page 33: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Table 5.Interpretation of MMPI-2-RF Narrative Report  Distance from Interpretive Report Score Distance from Expert Estimate

Scale M (SD) Range Incorrectly Elevated M (SD) RangeEID 17.8 (11.3) 2 - 42 14 (27.5%) 10.7 (6.6) 1 - 29

THD 8.0 (5.7) 0 - 20 14 (27.5%) 6.1 (5.1) 1 - 21BXD 6.2 (4.3) 2 - 17 11 (21.6%) 5.5 (4.1) 0 - 15RCd 10.0 (8.2) 2 - 28 12 (23.5%) 10.0 (7.6) 1 - 31RC1 23.9 (10.9) 4 - 41 4 (7.7%) 10.7 (6.9) 1 - 36RC2 13.8 (7.9) 2 - 33 13 (25.0%) 11.8 (7.4) 2 - 33RC3 13.3 (8.5) 1 - 31 11 (21.2%) 8.9 (6.5) 0 - 30RC4 8.8 (6.7) 2 - 42 30 (57.7%) 6.8 (8.0) 0 - 48RC6 5.7 (4.7) 0 - 25 15 (28.8%) 6.5 (4.7) 0 - 23RC7 10.7 (6.8) 0 - 30 19 (36.5%) 12.0 (7.4) 2 - 32RC8 9.9 (8.8) 1 - 36 35 (67.3%) 11.3 (6.2) 0 - 25RC9 10.3 (9.5) 0 - 40 13 (25.5%) 6.7 (7.5) 0 - 30

Note. Distance from profile and expert ratings are presented as absolute values of difference scores. Distance from profile was calculated out of available responses, with EID, THD, BXD, RCd, and RC9 estimated out of 51 participants as one left those scores blank (total n = 52). Incorrect elevations indicate the frequency that observed means below are above or the T-score = 65 recommended cut-score (Ben-Porath & Tellegen, 2008/2011) observed on the clinical profile. Self-reported competence for those completing this interpretive task was slightly to that observed for the overall sample (M = 54.9, SD = 27.8) but placed most (70%) at or above an average self-reported level of competence.

Page 34: ingramlab.files.wordpress.com€¦  · Web viewThe Restructured Clinical (RC) scales of the MMPI-2 and MMPI-2-RF were selected because of their simplified interpretive approach,

Running Head: PERSONALITY ASSESSMENT TRAINING

Table 6.Prediction of Symptom presentation  Trainee Rating Compared to Expert

Symptom, Behavior, or Attitude M SD Range% Incorrect Direction

Mean Expert

Hallucinations 2.1 0.8 1 - 5 16.9% 1Unstable Moods 2.1 1.0 1 - 5 22.0% 1

Nightmares 2.9 0.6 1 - 4 86.4% 3Angry 2.2 0.7 1 - 4 76.3% 2

Defensive 2.3 0.8 1 - 4 74.6% 2Critical 2.7 0.9 1 - 5 52.5% 2

Insecure 3.0 0.9 1 - 5 69.5% 3Anxious 3.0 1.0 1 - 4 64.4% 3

Pessimistic 3.4 1.1 1 - 5 67.8% 3Appetite problems 3.2 0.9 1 - 5 57.6% 3

Suicidal 3.1 1.0 1 - 5 60.3% 3Passive 3.7 0.7 2 - 5 39.7% 3

Somatic issues 3.6 1.0 1 - 5 41.4% 3Delusions 2.0 0.8 1 - 5 13.6% 1

Bizarre behavior 1.9 0.8 1 - 5 11.9% 1Panic episodes 3.1 0.8 1 - 5 64.4% 3

Inattentive / Easily bored 3.1 1.0 1 - 5 71.2% 1Manic/hypo-manic episodes 1.6 1.0 1 - 5 13.6% 1

Euphoric mood 2.1 1.0 1 - 5 27.1% 1Concentration problems 2.4 0.8 1 - 4 44.1% 1

Impulsivity 1.9 0.7 1 - 4 15.3% 1Paranoid 1.7 0.8 1 - 5 6.8% 1

Aggressive 2.2 0.8 1 - 4 28.8% 1Social 3.1 0.9 1 - 5 47.5% 2

Note. Behavioral manifestations of functioning were rated on a 5-point Likert-type scale (1 = Extremely Likely, 2 = Somewhat Likely, 3 = Neither likely nor unlikely, 4 = Somewhat unlikely, and 5 = Extremely unlikely). The percent of answers in the incorrect direction reflects the portion of trainees who endorsed 'Likely', 'Neither Likely or Unlikely', or 'Unlikely' response stems while expert panel selected responses in one of the other three response categories (e.g., the expert panel rated a symptom 'Extremely Likely' or 'Somewhat Likely' and a trainee did not select either of those frequency categories).