Dis1 jb

https://bncvirtual.com/vb_econtent.php?CSID=AZWACWCJOSSTUQTK2OQKMUSOB

Save Time On Research and Writing
Hire a Pro to Write You a 100% Plagiarism-Free Paper.
Get My Paper

It is important to have an awareness of why a psychologist chooses certain tests to include in an assessment battery. It is also important to understand how a forensic psychological risk assessment is conducted differently from other types of psychological clinical assessments. Understanding these differences sets the stage for much of your learning in the weeks to come.

In this Discussion, you will review the Learning Resources and apply your knowledge to the case.

To prepare

·  Review the case study provided in Chapter 5 of the course text.

Save Time On Research and Writing
Hire a Pro to Write You a 100% Plagiarism-Free Paper.
Get My Paper

·  Review the Learning Resources.

·  Consider the following questions:

o  

What did the psychologist consider when he chose the specific test instruments?

What did the psychologist consider when he chose the specific test instruments?

o  Why did the psychologist utilize the Personality Assessment Inventory (PAI)?

o  Since the counsel on the case did not request a risk assessment, why did the forensic psychologist choose to utilize the Psychopathy Checklist-Revised (PCL-R) along with the Psychopathic Personality Inventory-Revised (PPI-R) and the Static-99?

o  Why is it important that a forensic psychology practitioner, who does not administer test instruments, have an understanding of these test instruments in his or her work?

Post a response to the following, based on the case study:

·  Provide a brief summary of the case.

·  Explain what role the forensic psychology professional played in selecting the forensic risk assessment instrument used in the case.

·  Explain characteristics of the assessment that make it effective for this case.

·  Explain the implications regarding the selection of the assessment instrument and its impact on the outcome of the case.

https://ebookcentral.proquest.com/lib/waldenu/detail.action?docID=1565912

https://ebookcentral.proquest.com/lib/waldenu/detail.action?docID=1565912

https://web-b-ebscohost-com.ezp.waldenulibrary.org/ehost/detail/detail?vid=0&sid=3255cc2e-c073-43b1-8c0b-855a534a9a81%40pdc-v-sessmgr04&bdata=JnNpdGU9ZWhvc3QtbGl2ZSZzY29wZT1zaXRl#AN=2014-13439-005&db=pdh

https://web-b-ebscohost-com.ezp.waldenulibrary.org/ehost/detail/detail?vid=0&sid=b93da482-e5eb-4274-9f1a-758e74daa859%40sessionmgr101&bdata=JnNpdGU9ZWhvc3QtbGl2ZSZzY29wZT1zaXRl#AN=2014-13439-005&db=pdh

https://mym.cdn.laureate-media.com/2dett4d/Walden/FPSY/8200/CH/mm/podcast_repository/index.html

BOOK REVIEW

In Psychological Profiling: An Evolving Forensic Science (Fort Walton
Beach, FL: Precision Influence Technologies Publication. 2011), Dr.
Frank Goldstein aims to provide the reader with an overview of the
theory and practice of forensic psychological profiling, providing
a general foundation of political profiling, criminal profiling, and
forensic psychological profiling. Dr. Goldstein has written exten-
sively on psychological operations, information operations, and
human factors analysis. In fact, his book, Psychological Operations
Principles and Case Studies, is the premier text on psychological op-
erations in the U.S. and abroad. With contributors from the field of
forensic psychology and chapters written by Dr. Goldstein himself,
Psychological Profiling provides both the novice and experienced
psychological profiler with a wealth of useful information.
Divided into four sections, this book gives a broad overview of the
many aspects of psychological profiling, each section providing a win-
dow into specific profiling areas. The first section of the book focuses on
laying a basic foundation for profile assessment, covering what it takes
to develop a successful psychological profile. This includes a refresher
of basic traits, temperaments, and cognitive and affective domains and
how they relate to psychological assessment and profiling.

Book Review by Wendy Briggs
260 pages • ISBN 978-0-9836466-0

Section two gives an outline and methodology for conducting group
psychological assessment, including religious and country profiling.
Dr. Goldstein provides readers with questions the profiler needs to
answer to effectively understand an organization or group, and what
to look for when gathering data to develop an accurate country or
group profile. Chapter seven provides insight into the often overlooked
religious component to profiling by giving a brief history of United
States involvement in other nations, and the role religion has played
in the outcomes of that involvement.
Section three takes the reader through the steps of actually de-
veloping and writing a profile/assessment, providing key questions
for the profiler to use in his or her evaluation. This section focuses
on showing profilers how to create the most common form of
assessment provided to investigators, the “snap-shot” assessment.
Several examples of “snap-shots” of political figures are provided.
It gives a brief explanation of why it is important to measure ef-
fectiveness of the profile after each case, but leaves finding the
strategies on how to do this up to the reader. Chapters eight and
nine are perhaps the most useful chapters of the book, providing
step-by-step instructions on how to create a “snap-shot” analysis
of a subject through typical means.
Section four gives an overview of Psychological Operations
(PSYOPS), also known as Military Information Support Operations
(MISO), and how it has used psychological assessments for group
and country profiles. This section also breaks down the operational
facets of PSYOPS and shows where a psychological profiler fits into
that structure. The last half of this section provides information that
would be of particular interest to those working with local law en-
forcement. Chapters fifteen through seventeen provide foundational
information for jury profiling, interacting with lawyers, testifying
in court, and conducting psychological autopsies.
The book also provides an extensive glossary of profiling lan-
guage suggested by the profiling students at Marymount University
where Dr. Goldstein is currently on staff. The glossary provides the
reader with a fuller explanation of subjects related to the psycho-
logical profiling field. The appendices provide examples of actual
psychological profiles (both group and individual) produced by
professional profilers.
Forensic psychology is a relatively new science. Therefore, few
instructional books on the subject are available. Of those that are
on the market, few give such a practical overview of this burgeon-
ing science as Goldstein’s Psychological Profiling. While some books
focus mainly on criminal investigations or the philosophy of foren-
sic psychology, Psychological Profiling provides the information and
practical instruction needed for actually conducting the many dif-
ferent kinds of psychological profiles. While some prior knowledge
of psychology and criminology seems to be expected of the reader,
Psychological Profiling does a fairly good job of providing an accessible
and easy-to-understand overview of the potentially overwhelming
field of psychological profiling. The practical methodology provided
makes this book a valuable resource for any forensic psychologist to
have on his or her shelf. n

psychological
profiling:
An Evolving Forensic Science

By Frank L. Goldstein, PhD

submit your book for review today!
FOR MORE INFORMATION, CONTACT THE EdITOR AT:
The Forensic Examiner® • 2750 E. Sunshine • Springfield, MO 65804
OR AT: editor@ac fei.com • 800.592.1399

Fall 2013 THE FORENSIC EXAMINER® 73

Reproduced with permission of the copyright owner. Further reproduction prohibited without
permission.

  • Test Usage in Four Common Types of Forensic Mental Health Assessment
  • Jennifer L. McLaughlin and Lisa Y. Kan
    Sam Houston State University

    Without established standards of care for different types of forensic mental health assessment, practice
    surveys can provide information about current trends among evaluators and gauge how “typical” practice
    follows best practices. This study provides an update on the use of assessment tools in evaluations of
    response style/malingering, competency to stand trial, mental state at time of alleged offense, and
    nonsexual violence risk. Almost all forensic evaluators (n � 102) indicated that they use assessment tools
    to some extent when conducting these types of forensic mental health assessment. Of the 4 instrument
    types—multiscale inventories, forensic assessment/relevant instruments, cognitive/neuropsychological
    instruments, and projective techniques—evaluators reported using multiscale inventories at higher rates
    in evaluations of mental state at time of offense and forensic assessment/relevant instruments at higher
    rates for the other 3 issues. Projective techniques were used the least often across all forensic issues. We
    also considered how evaluator variables relate to differences in test usage. Finally, we compare our
    results with those of previous practice surveys and discuss the implications of these findings.

    Keywords: forensic assessment, practice survey, psychological testing

    In forensic mental health assessments (FMHAs), evaluators
    assess relevant psychological and legal constructs with the primary
    goal of assisting the trier of fact, whether it be a judge or jury, in
    addressing specific legal questions (Heilbrun et al., 2003). FMHAs
    involve obtaining, interpreting, and integrating multiple data
    sources, including record review, clinical interviews, and testing
    (Melton, Petrila, Poythress, & Slobogin, 2007; Zapf & Roesch,
    2009). Clinical interviews may be structured or unstructured, and
    testing may use clinical assessment instruments (e.g., cognitive
    and neuropsychological tests, personality inventories), forensically
    relevant instruments (FRIs; i.e., those that measure clinical con-
    structs related to legal issues, such as response style and psychop-
    athy), or forensic assessment instruments (FAIs; i.e., those that
    directly measure legal constructs, such as functional abilities re-
    lated to adjudicative competence; Melton et al., 2007; Otto &
    Heilbrun, 2002). Institutions may require the use of particular
    procedures or tests for specific forensic questions, and some states
    have firm guidelines on particular assessment procedures. For

    example, Texas requires specific components in sex offender risk
    assessments, including a sex offender screening tool (Tex. Gen.
    Laws ch. 62, § 1.01, 2005). Finally, evaluators must be able to
    support their techniques in court according to the jurisdiction’s
    standard of evidence.

    However, most forensic evaluations do not have statutory re-
    quirements that dictate the approach evaluators must take (Zapf &
    Roesch, 2009), not all evaluators work in institutions with proce-
    dural guidelines, and courts often lack sufficient knowledge to
    determine whether evidence is scientifically sound (Melton et al.,
    2007). FMHA scholars and practitioners have provided some
    guidance in terms of identifying training needs and professional
    competencies (e.g., DeMatteo, Marczyk, Krauss, & Burl, 2009;
    Varela & Conroy, 2012) and developing practice resources. There
    are multiple authoritative texts on conducting FMHAs (e.g., Hei-
    lbrun, Grisso, & Goldstein, 2009; Melton et al., 2007; Otto &
    Weiner, 2013; Packer, 2009; Weiner & Hess, 2006), and the
    Oxford University Press publishes the series “Best Practices in
    Forensic Mental Health Assessment” (Heilbrun et al., 2009). The
    American Psychology–Law Society (AP-LS) and the American
    Academy of Forensic Psychology, two primary professional orga-
    nizations in the field, recently updated the Specialty Guidelines for
    Forensic Psychology (American Psychological Association,
    2013b), which provide guidance for the “complete specialty prac-
    tice area” (p. 8), rather than any specific type of FMHA, unlike
    those published by the American Academy of Psychiatry and the
    Law (Giorgi-Guarnieri et al., 2002; Gold et al., 2008; Mossman et
    al., 2007). The only FMHA-specific practice guidelines published
    by the American Psychological Association (APA) are on child
    custody evaluations and evaluations of child protection matters
    (APA, 2010, 2013a; Heilbrun & Brooks, 2010). Importantly, “best
    practices” and guidelines are generally aspirational and nonbind-
    ing versus the minimally acceptable level of practice legally en-
    forced in standards of care (Heilbrun, DeMatteo, Marczyk, &
    Goldstein, 2008; Slobogin, Rai, & Reisner, 2008).

    JENNIFER L. MCLAUGHLIN is a clinical psychology doctoral student at Sam
    Houston State University where she also earned her masters degree. Her
    clinical and research interests include multicultural issues in forensic
    psycholog

    y.

    LISA Y. KAN received her PhD in clinical psychology from Sam Houston
    State University. She is currently an assistant professor and member of the
    doctoral program faculty in the Department of Psychology and Philosophy
    at Sam Houston State University. Her research interests include multicul-
    tural and practice issues in forensic psychology.
    THIS ARTICLE IS BASED on Jennifer L. McLaughlin’s masters thesis. Portions
    of this research were presented at the 2013 annual conference of the
    American Psychology Law Society.
    CORRESPONDENCE CONCERNING THIS ARTICLE should be addressed to Lisa Y.
    Kan, Department of Psychology and Philosophy, Sam Houston State
    University, Campus Box 2447, Huntsville, TX 77341-2447. E-mail:
    kan@shsu.edu

    T
    hi

    s
    do

    cu
    m

    en
    t

    is
    co

    py
    ri

    gh
    te

    d
    by

    th
    e

    A
    m

    er
    ic

    an
    Ps

    yc
    ho

    lo
    gi

    ca
    l

    A
    ss

    oc
    ia

    tio
    n

    or
    on

    e
    of

    its
    al

    lie
    d

    pu
    bl

    is
    he

    rs
    .

    T
    hi

    s
    ar

    tic
    le

    is
    in

    te
    nd

    ed
    so

    le
    ly

    fo
    r

    th
    e

    pe
    rs

    on
    al

    us
    e

    of
    th

    e
    in

    di
    vi

    du
    al

    us
    er

    an
    d

    is
    no

    t
    to

    be
    di

    ss
    em

    in
    at

    ed
    br

    oa
    dl

    y.

    Professional Psychology: Research and Practice © 2014 American Psychological Association
    2014, Vol. 45, No. 2,

    128

    –135 0735-7028/14/$12.00 DOI: 10.1037/a0036318

    128

    mailto:kan@shsu.edu

    http://dx.doi.org/10.1037/a0036318

    Without clear standards of care, how can forensic psychologists
    assure that their practice is “competent enough”? A likely resource
    is other forensic evaluators to identify how colleagues and peers
    typically conduct a specific type of FMHA. This might be done
    through education, training, consultation, supervision, or literature
    review. Practice surveys are particularly relevant, as they identify
    current trends among evaluators, which can inform standards of
    practice and care (Heilbrun et al., 2008), as well as highlight
    potential problems.1 In addition, such surveys can describe what is
    generally acceptable to the field (e.g., Lally, 2003), the basis of
    evidence admissibility in Frye v. United States (1923) and a
    component of the admissibility standards in Daubert v. Merrell
    Dow Pharmaceuticals (1993).

    We focused on test usage because, unlike clinical interviews and
    record review, there is less consensus on whether and which tests
    should be used in some types of FMHA (Giorgi-Guarnieri et al.,
    2002; Heilbrun & Collins, 1995; Lally, 2003; Melton et al., 2007).
    For example, clinical assessment instruments measure a person’s
    current cognitive, personality, or psychological functioning. Such
    information may have limited relevance in evaluations of criminal
    responsibility/mental status at time of alleged offense (MSO). This
    type of FMHA involves the assessment of the defendant’s
    thoughts, beliefs, and behaviors before, during, and immediately
    following the alleged offense to determine whether (and to what
    extent) the defendant’s understanding of the nature and/or wrong-
    fulness of his or her actions was impaired as a result of a mental
    disease or defect (Melton et al., 2007; Packer, 2009). Thus, it is the
    defendant’s prior functioning that is of primary interest in MSO
    evaluations, not his or her current functioning. Competency to
    stand trial (CST) evaluations, while present focused, require a
    specific assessment of the defendant’s current ability to demon-
    strate a factual and rational understanding of the legal proceedings,
    as well as the ability to consult effectively with counsel (Melton et
    al., 2007; Zapf & Roesch, 2009). Clinical assessment instruments
    do not directly measure these functional abilities.

    Self-report studies, however, suggest that evaluators consider
    acceptable, and often incorporate, the use of multiscale inventories
    and intelligence measures in forensic evaluations, including those
    for MSO and CST (Archer, Buffington-Vollum, Stredny, & Han-
    del, 2006; Borum & Grisso, 1995; Lally, 2003). In particular, all
    psychologists in one study (Borum & Grisso, 1995) reported some
    use of clinical instruments; in contrast, 36% and 46% reported
    never using FAIs in their CST and MSO cases, respectively.
    Interestingly, rates of test usage are substantially lower when they
    are based on a review of FMHA reports. Heilbrun and Collins
    (1995) found that only a minority of reports by psychologists
    included any testing (13% of inpatient and 41% of outpatient
    evaluations). Warren and colleagues (Warren, Murrie, Chauhan,
    Dietz, & Morris, 2004; Warren et al., 2006) too found that tests
    were used in only approximately 17% of CST and 22% of MSO
    evaluations conducted in Virginia.

    On the other hand, there is widespread support for the use of
    measures in violence risk assessments, a type of FMHA in which
    evaluators might be asked to estimate the likelihood of future
    violent behaviors within a specified timeframe, identify risk and
    protective factors, and recommend risk management strategies
    (e.g., Conroy & Murrie, 2007; Heilbrun, 2009; Skeem & Mo-
    nahan, 2011). Risk assessments can provide relevant information
    in various decisions, such as those for civil commitment and

    sentencing (Viljoen, McLachlan, & Vincent, 2010). Scholars have
    made significant progress in the development of risk assessment
    approaches and instruments in the past decades, and the fallacies of
    using an unstructured approach are well documented (see Camp-
    bell, French, & Gendreau, 2009). Both actuarial and structured
    professional judgment approaches involve the use of assessment
    tools that specify which and how risk factors are considered
    (Heilbrun, Yasuhara, & Shah, 2010). Indeed, some experts suggest
    that evaluators will likely need to defend why they did not use an
    instrument when conducting a risk assessment (Conroy & Murrie,
    2007; Viljoen et al., 2010). Even when using specific instruments
    is inappropriate (e.g., due to mismatch between evaluee and nor-
    mative sample; APA, 2002, 2013b), a structured approach is so
    important that evaluators should “structure” their assessments
    through systematic consideration of empirically identified or per-
    sonally relevant risk factors (Heilbrun et al., 2009). Recent surveys
    indicate that evaluator opinion and practice are in line with these
    recommendations. Lally (2003) found that an FAI, specifically the
    Psychopathy Checklist—Revised (PCL–R), is the only “recom-
    mended” instrument for violence risk assessments. Likewise, Vil-
    joen and colleagues (2010) reported that more evaluators used risk
    and psychopathy assessment tools almost all or all the time com-
    pared with tests for psychopathology or cognitive abilities.

    In addition, experts suggest that every FMHA, regardless of the
    referral question, should include an assessment of response style or
    malingering (Frederick, 2012; Melton et al., 2007; Zapf & Roesch,
    2009). This type of assessment aims to determine whether evaluees
    are feigning, exaggerating, or minimizing symptoms of cognitive
    deficits, psychopathology, or functional abilities, for an external
    gain they would not otherwise be granted (e.g., lesser sentence,
    mental health treatment, early discharge; Frederick, 2012). The
    potential consequences of misclassifying honest or dishonest re-
    sponders are serious (D’Amato & Denney, 2008; Rogers, Vitacco,
    & Kurus, 2010; Simon, 2007), and researchers have developed
    multiple tools to increase evaluators’ ability to accurately assess
    response style. Currently, a wide range of assessment tools is
    available, either as individual measures or as validity scales within
    other instruments (Boone, 2009; Heilbronner, Sweet, Morgan,
    Larrabee, & Millis, 2009; Rogers et al., 2010), for different target
    behaviors (e.g., psychopathology, cognitive functioning; Freder-
    ick, 2012). Thus, experts recommend the inclusion of empirically
    supported instruments in assessment of response style or malin-
    gering (Frederick, 2012; Heilbronner et al., 2009; Nussbaum,
    Hancock, Turner, Arrowood, & Melodick, 2008). Despite these
    recommendations, it is unclear how often evaluators use instru-
    ments to assess response style or malingering. Lally (2003) found
    that his respondents considered both the Minnesota Multiphasic
    Personality Inventory—2 (MMPI–2) and the Structured Interview
    of Reported Symptoms (SIRS) as “recommended” instruments,
    and Archer et al. (2006) reported that the SIRS and Test of
    Memory Malingering (TOMM) were the most popular among
    specialized tools for malingering.

    This study provides an update on the use of tests or assessment
    tools in four common foci in FMHAs—response style/malinger-
    ing, CST, MSO, and risk for future nonsexual violence. We also

    1 For a thorough discussion of the relationship between standards of
    practice and standards of care, see Heilbrun et al. (2008).

    T
    hi
    s
    do
    cu
    m
    en
    t
    is
    co
    py
    ri
    gh
    te
    d
    by
    th
    e
    A
    m
    er
    ic
    an
    Ps
    yc
    ho
    lo
    gi
    ca
    l
    A
    ss
    oc
    ia
    tio
    n
    or
    on
    e
    of
    its
    al
    lie
    d
    pu
    bl
    is
    he
    rs
    .
    T
    hi
    s
    ar
    tic
    le
    is
    in
    te
    nd
    ed
    so
    le
    ly
    fo
    r
    th
    e
    pe
    rs
    on
    al
    us
    e
    of
    th
    e
    in
    di
    vi
    du
    al
    us
    er
    an
    d
    is
    no
    t
    to
    be
    di
    ss
    em
    in
    at
    ed
    br
    oa
    dl
    y.

    129TEST USAGE IN FMHA

    examined whether and how evaluator variables relate to test usage.
    Specifically, we hypothesized that test usage would be higher
    among evaluators who completed their degree more recently and
    those who completed a postdoctoral fellowship in forensic psy-
    chology because they are more likely to be aware of the recent
    advances in test development for FMHAs (Heilbrun et al., 2008;
    Viljoen et al., 2010). In addition, we hypothesized that test usage
    would be higher among those who practice in one of the 17 states
    that specifically require FMHA training (Heilbrun & Brooks,
    2010).

    Method

    Participants

    Forensic evaluators (N � 115) in professional psychological
    practice consented to participate in the online survey. Of these, 103
    completed at least the demographic section of the survey. One
    respondent indicated that her clinical practice did not involve any
    FMHAs; thus, her responses were dropped from all analyses. In
    the final sample (n � 102), most participants identified as Cauca-
    sian (91.2%) and had earned a PhD (75.5%), on average, 14.24
    years (SD � 11.60) before. The majority (84.3%) reported com-
    pleting some formal training in forensic psychology, either through
    a postdoctoral fellowship (45.1%) or state/institution-required
    training (70.6%). In addition, 52% reported practicing in a state
    that requires training in FMHA and 16.7% reported receiving
    board certification from the American Board of Forensic Psychol-
    ogy (ABFP). Participants reported working in a variety of settings,
    with the most endorsed setting a state/federal institution (44.1%).
    See Table 1 for additional participant demographics.

    Procedure

    The data were collected as part of a larger study on evaluator
    beliefs, personality, and test usage (McLaughlin, 2013). We iden-
    tified potential participants through three methods and recruited
    them through e-mail. First, the primary author contacted the Web-
    master of AP-LS, who sent an initial recruitment e-mail to all
    nonstudent members of AP-LS, followed by a second e-mail 2
    weeks later. Second, the primary author individually e-mailed
    psychologists who received board certification from ABFP, using
    the publicly available e-mail directory. Lastly, we asked potential
    participants to distribute the survey information to their colleagues.
    Potential participants needed to have a graduate degree in psychol-
    ogy, be in professional practice (i.e., not currently completing a
    postdoctoral fellowship), and have conducted an evaluation ad-
    dressing at least one of the four forensic issues in the previous 6
    months.

    The recruitment message provided a brief description of the
    overall study, inclusionary criteria, contact information, and a link
    to the online materials on Survey Monkey. The link directed the
    participants to the confidentiality and consent agreement, and they
    were required to provide informed consent before beginning the
    survey.

    We were unable to estimate the response rate for several rea-
    sons. AP-LS did not provide the number of nonstudent members in
    its e-mail directory, and not all nonstudent members engage in
    forensic practice. We have no information about the number of

    invitations sent by participants, and some likely received multiple
    invitations, for example, from a colleague and e-mails from AP-
    LS. Our final sample size is comparable to those of other practice
    surveys (Borum & Grisso, 1995; Lally, 2003).

    Materials

    The online survey included three sections. The first section
    included demographic questions regarding the participants’ per-
    sonal characteristics (e.g., age, race, gender), graduate training,
    and training and practice in forensic psychology. Specifically,
    participants estimated the portion of their practice devoted to
    conducting forensic evaluations and the number of times they
    addressed each of the four forensic issues (i.e., response style/
    malingering, CST, MSO, risk for nonsexual violence), as well as
    other types of FMHA, in the past year. The second section (not
    presented here) consisted of the Goldberg Five-Factor Markers
    (Goldberg, 1992) personality measure. The third section included
    four subsections, one for each forensic issue, that asked about the
    evaluators’ perceptions of the use of four types of instruments.
    Examples of multiscale inventories (e.g., MMPI–2, MMPI–2—
    Restructured Form, Millon Clinical Multiaxial Inventory, Person-

    Table 1
    Characteristics of Participants

    Variable Value

    Mean (SD) age (years) 46.41 (13.44)
    Gender: Female, % 51.0
    Ethnicity, %

    African American 2.00
    Asian American 3.90
    Caucasian 91.2
    Hispanic 2.90
    Other 2.00

    Highest degree, %
    PhD 75.50
    PsyD 23.50
    MA 1.00

    Mean (SD) years since degree attainment 14.24 (11.60)
    Work environment, %

    Academic 16.70
    State/federal institution 44.10
    Hospital 24.50
    Independent practice 37.30

    Any forensic training, % 84.30
    Forensic postdoctoral training 45.10
    State/institution training 70.60

    Practice in a state that requires forensic training, % 52.00
    Board certification in forensic psychology, % 16.70
    Mean (SD) percentage of practice devoted to

    forensic evaluations (n � 100) 67.39 (32.80)
    Mean (SD) number of times forensic issue was

    addressed in past year
    Response style/malingering (n � 99) 39.61 (107.16)
    CST (n � 101) 34.12 (46.62)
    MSO (n � 98) 14.12 (25.61)
    Nonsexual violence risk (n � 96) 20.28 (38.11)
    Other types of FMHA (n � 43) 27.53 (63.05)

    Note. n � 102, unless otherwise noted. Evaluators could choose more
    than one category for the variables ethnicity and work environment;
    therefore, the percentages do not add to 100%. CST � competence to stand
    trial; MSO � mental state at time of alleged offense; FMHA � forensic
    mental health assessment.

    T
    hi
    s
    do
    cu
    m
    en
    t
    is
    co
    py
    ri
    gh
    te
    d
    by
    th
    e
    A
    m
    er
    ic
    an
    Ps
    yc
    ho
    lo
    gi
    ca
    l
    A
    ss
    oc
    ia
    tio
    n
    or
    on
    e
    of
    its
    al
    lie
    d
    pu
    bl
    is
    he
    rs
    .
    T
    hi
    s
    ar
    tic
    le
    is
    in
    te
    nd
    ed
    so
    le
    ly
    fo
    r
    th
    e
    pe
    rs
    on
    al
    us
    e
    of
    th
    e
    in
    di
    vi
    du
    al
    us
    er
    an
    d
    is
    no
    t
    to
    be
    di
    ss
    em
    in
    at
    ed
    br
    oa
    dl
    y.

    130 MCLAUGHLIN AND KAN

    ality Assessment Inventory), cognitive/neuropsychological instru-
    ments (e.g., Wechsler scales, Halstead Reitan Neuropsychological
    Battery, Trailmaking Test), and projective techniques (e.g., Ror-
    schach Inkblot Test, Thematic Apperception Test) were the same
    for each forensic issue, as evaluators might use the same instru-
    ment(s) for different purposes. For instance, an evaluator might
    use the MMPI–2 to assess both response style and psychopathol-
    ogy as part of an MSO evaluation; in the survey, participants were
    asked to endorse the use of MMPI–2 for the forensic issues
    separately. Examples of FAIs/FRIs were specific for each type of
    FMHA, such as the TOMM and SIRS for assessing response
    style/malingering and the MacArthur Competency Assessment
    Tool for CST. Participants also estimated the percentage of eval-
    uations for which each type of instrument was actually used for
    each forensic issue.

    All participants were asked about their perceptions and practice
    in assessing response style/malingering because it is considered a
    part of all forensic evaluations (Frederick, 2012; Melton et al.,
    2007; Rogers, 2008). For the other forensic issues, participants
    only answered the respective questions if they addressed the issue
    in the past 6 months. Therefore, participants could answer ques-
    tions pertaining to one, two, or three additional forensic issues.

    Results

    Participants varied widely in the portion of their practice de-
    voted to forensic evaluations, from 10% to 100% (M � 67.39%,
    SD � 32.80). Response style/malingering was the most common
    forensic issue addressed, with respondents reporting that they
    evaluated the construct an average of 39.61 times (SD � 107.16)
    in the past year, followed by CST (M � 34.12, SD � 46.62), risk

    for nonsexual violence (M � 20.28, SD � 38.11), and MSO (M �
    14.12, SD � 25.61). Approximately 35% reported conducting
    other types of FMHA in the past year. However, seven participants
    reported that they did not assess response style/malingering in the
    past year, despite addressing other forensic issues, and a majority
    (82.4%) indicated that they addressed other forensic issues more
    frequently than they did response style/malingering.

    Regarding test usage, one participant did not answer any ques-
    tions regarding test usage for any of the four forensic issues and
    was therefore not included in further analyses. All other partici-
    pants (n � 101) reported using assessment tools to some extent.
    Only one of 82 (1.2%) evaluators who assessed CST and four of
    72 (5.6%) who assessed nonsexual violence risk indicated never
    using any assessment tool, versus 12 of 68 (17.6%) who assessed
    MSO. Mean frequencies (i.e., percentage of cases) of test usage
    differed by type of assessment tool and type of FMHA (see Table
    2). FAIs/FRIs were the most frequently used in evaluations of
    response style/malingering, CST, and risk for nonsexual violence,
    and multiscale inventories were the most frequently used in MSO
    assessments. Evaluators used projective techniques the least often
    across all four forensic issues.

    To compare our results with previous surveys, we also cate-
    gorized the frequencies of usage based on the categories in
    Borum and Grisso (1995). For evaluation of response style/
    malingering, more participants reported “almost always” (81–
    100%) using FAIs/FRIs than other instrument types (see Table
    2). The pattern was similar for CST and nonsexual violence risk
    evaluations, and more participants reported they “almost al-
    ways” used multiscale inventories in MSO assessments than
    other instrument types.

    Table 2
    Frequencies of Test Usage for Four Forensic Issues

    Instrument type
    Mean (SD) percentage

    of cases

    Categories of test usage frequency (%)a

    Never
    (0%)

    Rarely
    (1–10%)

    Sometimes
    (11–40%)

    Frequently
    (41–80%)

    Almost always
    (81–100%)

    Response style/malingering
    Multiscale inventories (n � 101) 52.21 (39.13) 9.9 17.8 14.9 25.7 31.7
    FRI/FAI (n � 100) 66.44 (35.70) 2.0 15.0 10.0 25.0 48.0
    C/N (n � 100) 21.41 (27.48) 30.0 28.0 22.0 15.0 5.0
    Projective (n � 100) 4.86 (15.97) 81.0 8.0 7.0 3.0 1.0

    CST
    Multiscale inventories (n � 82) 26.80 (33.20) 29.3 25.6 13.4 19.5 12.2
    FRI/FAI (n � 82) 42.83 (39.96) 11.0 25.6 20.7 13.4 29.3
    C/N (n � 82) 26.62 (29.69) 12.2 30.5 36.6 9.8 11.0
    Projective (n � 82) 2.52 (11.00) 87.8 7.3 2.4 2.4 0.0

    MSO
    Multiscale inventories (n � 68) 38.40 (39.04) 23.5 20.6 13.2 23.5 19.1
    FRI/FAI (n � 68) 20.79 (34.80) 57.4 14.7 7.4 5.9 14.7
    C/N (n � 68) 20.88 (26.77) 27.9 27.9 26.5 11.8 5.9
    Projective (n � 68) 4.79 (15.72) 80.9 10.3 4.4 2.9 1.5

    Nonsexual violence risk
    Multiscale inventories (n � 72) 49.13 (40.23) 16.7 15.3 16.7 22.2 29.2
    FRI/FAI (n � 72) 78.68 (32.98) 5.6 4.2 8.3 18.1 63.9
    C/N (n � 70) 26.91 (31.76) 24.3 28.6 21.4 15.7 10.0
    Projective (n � 70) 6.10 (16.25) 78.6 8.6 7.1 5.7 0.0

    Note. Percentage of cases refers to the proportion of cases in which participants reported using the instrument type for the forensic issue. CST �
    competence to stand trial; MSO � mental state at time of alleged offense; FAI � forensic assessment instrument; FRI � forensically relevant instrument;
    C/N � cognitive/neuropsychological instruments.
    a Values in these columns represent the percentage of respondents whose reported test usage frequency falls within that frequency category.

    T
    hi
    s
    do
    cu
    m
    en
    t
    is
    co
    py
    ri
    gh
    te
    d
    by
    th
    e
    A
    m
    er
    ic
    an
    Ps
    yc
    ho
    lo
    gi
    ca
    l
    A
    ss
    oc
    ia
    tio
    n
    or
    on
    e
    of
    its
    al
    lie
    d
    pu
    bl
    is
    he
    rs
    .
    T
    hi
    s
    ar
    tic
    le
    is
    in
    te
    nd
    ed
    so
    le
    ly
    fo
    r
    th
    e
    pe
    rs
    on
    al
    us
    e
    of
    th
    e
    in
    di
    vi
    du
    al
    us
    er
    an
    d
    is
    no
    t
    to
    be
    di
    ss
    em
    in
    at
    ed
    br
    oa
    dl
    y.

    131TEST USAGE IN FMHA

    Finally, we used nonparametric tests to examine whether and
    how test usage differed based on completion of postdoctoral train-
    ing in forensic psychology, state requirements for FMHA training,
    and recency of degree attainment, to accommodate the nonnormal
    distributions of the test usage data (see Table 3). We focused on
    the use of multiscale inventories, FAIs/FRIs, and cognitive/neuro-
    psychological tools given that only a few evaluators reported using
    projective techniques. To minimize the risk of Type I error, we set
    a more stringent statistical significance level at p � .001 (using
    Bonferroni correction [Warner, 2008], based on dividing the stan-
    dard statistical significance level of p � .05 by 36 comparisons).

    There were no significant differences in test usage based on
    evaluators’ postdoctoral training in forensic psychology or states’
    requirements for forensic training. The only significant finding
    involved years since degree attainment, which was negatively
    correlated with the use FAIs/FRIs in nonsexual violence risk
    assessments (rs � �.38, p � .001, one-tailed).

    Discussion

    In this study, our primary goal was to describe evaluators’ test
    usage of four common types of FMHA— response style/malin-
    gering, CST, MSO, and nonsexual violence risk. In general, the
    reported use of each instrument type averaged about 50% or less,
    regardless of the forensic issue. Two notable exceptions are the use
    of FAIs/FRIs in evaluations of response style/malingering and
    nonsexual violence risk; evaluators reported they used FAIs/FRIs
    in an average of 66% and 79% of their cases, respectively. This is
    encouraging, given the accumulating empirical support for these
    structured assessments for evaluating response style and violence
    risk (e.g., Conroy & Murrie, 2007; Rogers, 2008; Skeem & Mo-
    nahan, 2011). Furthermore, the occasional use of other instrument
    types, such as multiscale inventories and cognitive/neuropsycho-

    logical measures in CST evaluations, is consistent with the lack of
    consensus on the usefulness of such clinical assessment instru-
    ments in FMHAs (Giorgi-Guarnieri et al., 2002; Melton et al.,
    2007). On one hand, clinical assessment instruments do not spe-
    cifically address the psycholegal issue at hand, whether it be CST,
    MSO, or risk of nonsexual violence. On the other, they do offer
    structured approaches to assess clinical constructs, such as psy-
    chopathology and intelligence, which can be relevant to FMHAs.
    Their use might be particularly appropriate when evaluators are
    required to offer a rationale for observed functional impairments
    (e.g., a defendant cannot consult with defense counsel because of
    severe intellectual disability; Mossman et al., 2007) or when they
    need to clarify diagnoses to provide appropriate treatment recom-
    mendations.

    More disconcerting, however, is the result that a majority of
    participants indicated that they assessed other forensic issues more
    frequently than response style/malingering, despite the fact that
    various experts (Frederick, 2012; Melton et al., 2007; Rogers,
    2008) consider response style an integral part of all forensic
    evaluations. Our respondents simply might have misunderstood
    the instructions to count each forensic issue separately if they
    address multiple ones in the same evaluation. Or they might have
    worked in settings in which different evaluators are responsible for
    different aspects of forensic evaluations, although this is unlikely
    to be true for most of the respondents. A less innocuous reason is
    that evaluators are indeed omitting assessment of response style in
    forensic evaluations, which is arguably inconsistent with best
    practices.

    Another finding worth additional consideration is the use of
    FAIs/FRIs in nonsexual violence risk assessment. On the one
    hand, it is clearly the most used instrument type for the issue, with
    respondents indicating its use in approximately 79% of their cases
    on average, which is consistent with practice guidelines (e.g.,
    Skeem & Monahan, 2011). However, interpreted another way,
    evaluators reported they did not use FAIs/FRIs in about 20% of
    their cases. This might be reasonable and appropriate given that
    evaluators must consider the instruments’ relevance for the partic-
    ular evaluee in that particular context when determining whether
    and which tests to use (APA, 2002, 2013b; Heilbrun, 1992; Varela
    & Conroy, 2012). One potential reason for not using assessment
    tools is the mismatch between the evaluee and the instrument’s
    normative sample, which is common when working with diverse
    evaluees (Weiss & Rosenfeld, 2010). As Singh, Grann, and Fazel
    (2011) pointed out, normative samples for risk assessment tools
    tend to be predominantly Caucasian. Correspondingly, they found
    some evidence that predictive validity tended to be higher when
    study samples were predominantly Caucasian in their meta-
    analysis of violence risk assessment tools. In addition, many
    studies on risk assessment instruments failed to provide a thorough
    demographic breakdown of participants and neglected to include
    the racial makeup of their participants (Gonzalez, 2013), which
    makes it difficult for evaluators to determine whether the instru-
    ment is appropriate for use. Problems with using risk assessment
    tools also arise when there are limited records available for eval-
    uees. The PCL–R (Hare, 2003), a common FRI in risk assess-
    ments, relies heavily on records, to the extent that its use is allowed
    without interviewing the evaluee but is discouraged if sufficient
    records are unavailable. Without PCL–R results, use of other
    common risk instruments is restricted, as the Historical-Clinical-

    Table 3
    Relationship Between Test Usage and Evaluator Characteristic

    Postdoctoral
    training

    State-
    required
    training

    Years since
    degree

    attainment

    Test usage z p z p rs p

    Response style/malingering
    Multiscale inventories �1.20 .12 �1.24 .11 .007 .47
    FAI/FRI �0.18 .43 �1.28 .10 �.14 .09
    C/N �0.57 .29 �1.15 .13 .10 .17

    CST
    Multiscale inventories �0.75 .23 �0.32 .38 .09 .21
    FAI/FRI �1.24 .11 �1.58 .06 .15 .09
    C/N �1.77 .04 �1.82 .03 .05 .34

    MSO
    Multiscale inventories �0.71 .24 �1.44 .08 .10 .20
    FAI/FRI �0.83 .21 �0.04 .49 .28 .01
    C/N �0.43 .34 �1.03 .15 .29 .01

    Nonsexual violence risk
    Multiscale inventories �1.04 .15 �0.73 .23 .11 .18
    FAI/FRI �1.84 .03 �2.61 .004 �.38� .001
    C/N �0.32 .38 �0.06 .48 .03 .40

    Note. CST � competence to stand trial; MSO � mental state at time of
    alleged offense; FAI � forensic assessment instrument; FRI � forensically
    relevant instrument; C/N � cognitive/neuropsychological instruments.
    � Significant at p � .001, one-tailed.

    T
    hi
    s
    do
    cu
    m
    en
    t
    is
    co
    py
    ri
    gh
    te
    d
    by
    th
    e
    A
    m
    er
    ic
    an
    Ps
    yc
    ho
    lo
    gi
    ca
    l
    A
    ss
    oc
    ia
    tio
    n
    or
    on
    e
    of
    its
    al
    lie
    d
    pu
    bl
    is
    he
    rs
    .
    T
    hi
    s
    ar
    tic
    le
    is
    in
    te
    nd
    ed
    so
    le
    ly
    fo
    r
    th
    e
    pe
    rs
    on
    al
    us
    e
    of
    th
    e
    in
    di
    vi
    du
    al
    us
    er
    an
    d
    is
    no
    t
    to
    be
    di
    ss
    em
    in
    at
    ed
    br
    oa
    dl
    y.

    132 MCLAUGHLIN AND KAN

    Risk Management—20 (Webster, Douglas, Eaves, & Hart, 1997)
    and the Violence Risk Appraisal Guide (Quinsey, Harris, Rice, &
    Cormier, 1998, 2006) incorporate the PCL–R total score. Of
    course, there are less defensible reasons for not using risk assess-
    ment tools, including evaluators’ unfamiliarity with, or lack of
    competence in, using them (Tolman & Mullendore, 2003). The
    time and effort required to gather and review information neces-
    sary to use these instruments might deter their use among some
    evaluators. Systemic constraints, such as time limits in assessing
    risk during temporary holds for involuntary hospitalizations and
    lack of assessment measures provided by institutions, likely limit
    test usage, regardless of evaluators’ intentions or desires to engage
    in best practices.

    Perhaps most unexpected was the overall lack of significant
    differences on test usage based on evaluator characteristics. Con-
    trary to our hypotheses, evaluators who completed postdoctoral
    training in forensic psychology or practiced in states with required
    FMHA training did not report significantly higher rates of test
    usage. This might be due to the study’s sample, as most reported
    completing some forensic training. The primary methods of re-
    cruitment involved e-mailing psychologists board certified by
    ABFP and members of AP-LS; therefore, respondents (regardless
    of training requisites) might be equally likely to attend the orga-
    nizations’ presentations on FMHAs. The only significant differ-
    ence involved the use of FAIs/FRIs in nonsexual violence risk
    assessments. Consistent with our hypothesis, evaluators who had
    received their degree more recently were more likely to use fo-
    rensic instruments. This might be attributed to their being more
    aware of the recent advances in test development for FMHAs
    (Heilbrun et al., 2008; Viljoen et al., 2010).

    How do our results compare with previous practice surveys?
    Like the psychologists in Borum and Grisso’s (1995) study, almost
    all of our respondents reported using assessment tools to some
    extent. The pattern of test usage across different forensic issues
    (i.e., multiscale inventories more frequently used in MSO evalu-
    ations and FAIs/FRIs more frequently used for the other three
    issues) is consistent with Lally’s (2003) results on which tests were
    “recommended” for these types of FMHA. In addition, our finding
    that a majority of respondents “almost always” use FAIs/FRIs in
    nonsexual violence risk assessments is congruent with Viljoen et
    al. (2010).

    Taken together, we conclude that assessment tools are routinely
    used in these four common types of FMHA. Use of some specific
    instrument types might be lower than expected or desired, but such
    findings should not be considered recommended or acceptable
    practice. Rather, we interpret them as indicators of areas for
    improvement, either in further promotion and education of existing
    measures or in development of better tools. We agree with other
    scholars’ (e.g., Borum & Grisso, 1995; Conroy & Murrie, 2007;
    Viljoen et al., 2010) assertion that evaluators should provide sound
    rationale for their decisions not to use assessment tools in FMHAs,
    especially for issues for which empirical evidence strongly support
    such usage (e.g., risk assessment). There are clear preferences for
    the use of certain instrument types among all four issues, partic-
    ularly for the use of FAIs/FRIs in nonsexual violence risk assess-
    ments. However, the use of forensic instruments in nonsexual
    violence risk assessments is influenced by recency of evaluators’
    degree. These results, along with evidence that evaluators might
    not assess response style/malingering in all FMHAs, highlight the

    importance of developing and maintaining professional compe-
    tence (e.g., APA, 2002; Varela & Conroy, 2012).

    Professional competence in forensic psychology includes spe-
    cialized knowledge and skills, which are primarily developed at
    the postdoctoral level through a variety of avenues (Forensic
    Specialty Council, 2007; Otto & Heilbrun, 2002; Packer, 2008).
    Formal postdoctoral fellowships, typically 1 to 2 years long, offer
    intensive training and supervised experiences in forensic psychol-
    ogy (DeMatteo et al., 2009). The AP-LS/APA Division 41 Website
    and the Association of Psychology Postdoctoral and Internship
    Centers provide directories of postdoctoral fellowships with a
    forensic component. Completion of fellowships often provides the
    postdoctoral students with supervised experiences necessary for
    licensure in some states (DeMatteo et al., 2009), and it can lead to
    early eligibility for board certification in forensic psychology
    (American Academy of Forensic Psychology, n.d.; Packer, 2008).
    However, the likelihood of obtaining a fellowship is very low:
    Malesky and Proctor (2012) identified only 32 positions (of which
    29 were funded) among 16 formal postdoctoral fellowships in
    forensic psychology for 317 applicants in the year 2008–2009.

    Beyond formal fellowships, practitioners can gain postdoctoral
    training in forensic psychology through state certification pro-
    grams or continuing education (CE) workshops (DeMatteo et al.,
    2009), with supplemental supervised experience (Packer, 2008).
    For example, Virginia requires evaluators to complete a 5-day
    program at the Institute of Law, Psychiatry, and Public Policy to be
    eligible to conduct CST and MSO evaluations with adults (Insti-
    tute of Law, Psychiatry, and Public Policy, 2012). CE workshops
    on forensic psychology are widely available and can be several
    hours to several days long. AP-LS typically offers CE workshops
    before its annual conferences, and the American Academy of
    Forensic Psychology provides several intensive programs through-
    out the year. Some states also require periodic recertification or a
    minimum of CE hours in forensic psychology each year, which can
    help evaluators stay informed about advancements in the field
    (Heilbrun & Brooks, 2010).

    Finally, regardless of training, experience, or recency of degree
    attainment, evaluators need to maintain competence and stay
    abreast of relevant literature. Evaluators can learn about the latest
    research, practice recommendations, and legal developments by
    attending conferences and CE workshops, joining listservs, and
    reading texts or journals dedicated to forensic psychology such as
    the series in “Best Practices in Forensic Mental Health Assess-
    ment,” Law and Human Behavior, and Psychology, Public Policy,
    and Law. Staying informed allows evaluators to maintain an em-
    pirically and legally informed basis to justify their work, an inte-
    gral part of competent forensic practice (e.g., Conroy & Murrie,
    2007; Nicholson & Norwood, 2000).

    This study has several limitations. First, we recruited partici-
    pants mostly by e-mailing psychologists board certified by ABFP
    and members of AP-LS. As Archer and colleagues (2006) pointed
    out, these evaluators might be systematically different from those
    who do not hold memberships in professional organizations and
    thus might not represent the “average” forensic evaluator. Simi-
    larly, there might be important, but unknown, differences between
    those who agreed to participate in online research and those who
    did not. Future studies should also recruit participants in person
    (i.e., at conferences, workshops, work sites) and through other
    means (e.g., state registry of licensed psychologists) to address

    T
    hi
    s
    do
    cu
    m
    en
    t
    is
    co
    py
    ri
    gh
    te
    d
    by
    th
    e
    A
    m
    er
    ic
    an
    Ps
    yc
    ho
    lo
    gi
    ca
    l
    A
    ss
    oc
    ia
    tio
    n
    or
    on
    e
    of
    its
    al
    lie
    d
    pu
    bl
    is
    he
    rs
    .
    T
    hi
    s
    ar
    tic
    le
    is
    in
    te
    nd
    ed
    so
    le
    ly
    fo
    r
    th
    e
    pe
    rs
    on
    al
    us
    e
    of
    th
    e
    in
    di
    vi
    du
    al
    us
    er
    an
    d
    is
    no
    t
    to
    be
    di
    ss
    em
    in
    at
    ed
    br
    oa
    dl
    y.

    133TEST USAGE IN FMHA

    these concerns. Second, we focused on the use of different instru-
    ment types for different forensic issues rather than test usage
    overall or specific instruments. We chose to do this to shorten the
    survey and encourage participation, but we cannot identify which
    instrument(s) evaluators tend to use for different purposes. Future
    studies should address such questions to further inform practice
    trends.

    Third, we did not ask participants to provide rationale for their
    use (or nonuse) of assessment tools. Evaluators should only use
    instruments in cases for which there is sufficient justification
    (Nicholson & Norwood, 2000), and understanding their decision-
    making process can more clearly inform standards of care. Al-
    though evaluators will likely benefit from the use of instruments in
    most instances, there are appropriate reasons why evaluators do
    not use assessment tools in every case. Just as we are concerned
    about lower-than-expected test usage being erroneously consid-
    ered as recommended or acceptable practice, high utilization rates
    stemming from improper use of instruments should not form the
    basis for standards of care either. Thus, frequency of test use is
    only one of many pieces of information necessary for understand-
    ing competent practice. Competent evaluators should be able to
    provide explicit, thoughtful, and empirically based rationale for
    their decisions, from using assessment tools to forming their con-
    clusions, and future research should aim to explore such reasoning
    using both quantitative and qualitative methods. This can help
    better identify areas of training or development necessary to ad-
    vance the field of FMHAs.

    References

    American Academy of Forensic Psychology. (n.d.). Early career scholar-
    ship program. Retrieved from http://www.aafp.ws/Early-Career-
    Scholarship-Program.asp

    American Psychological Association. (2002). Ethical principles of psy-
    chologists and code of conduct. American Psychologist, 57, 1060–1073.
    doi:10.1037/0003-066X.57.12.1060

    American Psychological Association. (2010). Guidelines for child custody
    evaluations in family law proceedings. American Psychologist, 65, 863–
    867. doi:10.1037/a0021250

    American Psychological Association. (2013a). Guidelines for psychologi-
    cal evaluations in child protection matters. American Psychologist, 68,
    20–31. doi:10.1037/a0029891

    American Psychological Association. (2013b). Specialty guidelines for
    forensic psychology. American Psychologist, 68, 7–19. doi:10.1037/
    a0029889

    Archer, R. P., Buffington-Vollum, J. K., Stredny, R. V., & Handel, R. W.
    (2006). A survey of psychological test use patterns among forensic
    psychologists. Journal of Personality Assessment, 87, 84 –94. doi:
    10.1207/s15327752jpa8701_07

    Boone, K. B. (2009). The need for continuous and comprehensive sam-
    pling of effort/response bias during neuropsychological examinations.
    The Clinical Neuropsychologist, 23, 729 –741. doi:10.1080/
    13854040802427803

    Borum, R., & Grisso, T. (1995). Psychological test use in criminal forensic
    evaluations. Professional Psychology: Research and Practice, 26, 465–
    473.

    Campbell, M. A., French, S., & Gendreau, P. (2009). The prediction of
    violence in adult offenders: A meta-analytic comparison of instruments
    and methods of assessment. Criminal Justice and Behavior, 36, 567–
    590. doi:10.1177/0093854809333610

    Conroy, M. A., & Murrie, D. C. (2007). Forensic assessment of violence
    risk: A guide for risk assessment and risk management. Hoboken, NJ:
    Wiley. doi:10.1002/9781118269671

    D’Amato, C. P., & Denney, R. L. (2008). The diagnostic utility of the
    Rarely Missed Index of the Wechsler Memory Scale—Third Edition in
    detecting response style bias in an adult male incarceration setting.
    Archives of Clinical Neuropsychology, 23, 553–561. doi:10.1016/j.acn
    .2008.05.007

    Daubert v. Merrell Dow Pharmaceuticals, Inc., 113 S. Ct. 2786 (1993).
    DeMatteo, D., Marczyk, G., Krauss, D. A., & Burl, J. (2009). Educational

    and training models in forensic psychology. Training and Education in
    Professional Psychology, 3, 184–191. doi:10.1037/a0014582

    Forensic Specialty Council. (2007). Education and training guidelines for
    forensic psychology. Retrieved from http://www.apadivisions.org/
    division-41/education/guidelines

    Frederick, R. I. (2012). Malingering/cooperation/effort. In D. Faust (Ed.),
    Coping with psychiatric and psychological testimony (6th ed., pp. 229–
    247). New York, NY: Oxford University Press.

    Frye v. United States, 293 F. 1013 (D. C. Cir. 1923).
    Giorgi-Guarnieri, D., Janofsky, J., Keram, E., Lawsky, S., Merideth, P.,

    Mossman, D. A., . . . Zonana, H. (2002). AAPL practice guideline for
    forensic psychiatric evaluation of defendants raising the insanity de-
    fense. Journal of the American Academy of Psychiatry and the Law, 30,
    3–40.

    Gold, L. H., Anfang, S. A., Drukteinis, A. M., Metzner, J. L., Price, M.,
    Wall, B. W., . . . Zonana, H. V. (2008). AAPL practice guideline for the
    forensic evaluation of psychiatric disability. Journal of the American
    Academy of Psychiatry and the Law, 36, S3–S50.

    Goldberg, L. R. (1992). The development of markers for the Big-Five
    factor structure. Psychological Assessment, 4, 26–42.

    Gonzalez, E. (2013). Violence risk assessment and cultural diversity: A
    study space analysis. Unpublished manuscript, Department of Psychol-
    ogy and Philosophy, Sam Houston State University, Huntsville, TX.

    Hare, R. D. (2003). The Hare Psychopathy Checklist—Revised (2nd ed.).
    Toronto, ON: Multi-Health Systems.

    Heilbronner, R. L., Sweet, J. J., Morgan, J. E., Larrabee, G. J., & Millis,
    S. R. (2009). American Academy of Clinical Neuropsychology consen-
    sus conference statement on the neuropsychological assessment of ef-
    fort, response bias, and malingering. The Clinical Neuropsychologist,
    23, 1093–1129. doi:10.1080/13854040903155063

    Heilbrun, K. (1992). The role of psychological testing in forensic assess-
    ment. Law and Human Behavior, 16, 257–272. doi:10.1007/
    BF01044769

    Heilbrun, K. (2009). Evaluation for risk of violence in adults. New York,
    NY: Oxford University Press.

    Heilbrun, K., & Brooks, S. (2010). Forensic psychology and forensic
    science: A proposed agenda for the next decade. Psychology, Public
    Policy, and Law, 16, 219–253. doi:10.1037/a0019138

    Heilbrun, K., & Collins, S. (1995). Evaluations of trial competency and
    mental state at time of offense: Report characteristics. Professional
    Psychology: Research and Practice, 26, 61–67.

    Heilbrun, K., DeMatteo, D., Marczyk, G., & Goldstein, A. M. (2008).
    Standards of practice and care in forensic mental health assessment:
    Legal, professional, and principles-based considerations. Psychology,
    Public Policy, and Law, 14, 1–26. doi:10.1037/1076-8971.14.1.1

    Heilbrun, K., Grisso, T., & Goldstein, A. (2009). Foundations of forensic
    mental health assessment (Best practices in forensic mental health
    assessment). New York, NY: Oxford University Press.

    Heilbrun, K., Marczyk, G. R., DeMatteo, D., Zillmer, E. A., Harris, J., &
    Jennings, T. (2003). Principles of forensic mental health assessment:
    Implications for neuropsychological assessment in forensic contexts.
    Assessment, 10, 329–343. doi:10.1177/1073191103258591

    Heilbrun, K., Yasuhara, K., & Shah, S. (2010). Violence risk assessment

    T
    hi
    s
    do
    cu
    m
    en
    t
    is
    co
    py
    ri
    gh
    te
    d
    by
    th
    e
    A
    m
    er
    ic
    an
    Ps
    yc
    ho
    lo
    gi
    ca
    l
    A
    ss
    oc
    ia
    tio
    n
    or
    on
    e
    of
    its
    al
    lie
    d
    pu
    bl
    is
    he
    rs
    .
    T
    hi
    s
    ar
    tic
    le
    is
    in
    te
    nd
    ed
    so
    le
    ly
    fo
    r
    th
    e
    pe
    rs
    on
    al
    us
    e
    of
    th
    e
    in
    di
    vi
    du
    al
    us
    er
    an
    d
    is
    no
    t
    to
    be
    di
    ss
    em
    in
    at
    ed
    br
    oa
    dl
    y.

    134 MCLAUGHLIN AND KAN

    http://www.aafp.ws/Early-Career-Scholarship-Program.asp

    http://www.aafp.ws/Early-Career-Scholarship-Program.asp

    http://dx.doi.org/10.1037/0003-066X.57.12.1060

    http://dx.doi.org/10.1037/a0021250

    http://dx.doi.org/10.1037/a0029891

    http://dx.doi.org/10.1037/a0029889

    http://dx.doi.org/10.1037/a0029889

    http://dx.doi.org/10.1207/s15327752jpa8701_07

    http://dx.doi.org/10.1207/s15327752jpa8701_07

    http://dx.doi.org/10.1080/13854040802427803

    http://dx.doi.org/10.1080/13854040802427803

    http://dx.doi.org/10.1177/0093854809333610

    http://dx.doi.org/10.1002/9781118269671

    http://dx.doi.org/10.1016/j.acn.2008.05.007

    http://dx.doi.org/10.1016/j.acn.2008.05.007

    http://dx.doi.org/10.1037/a0014582

    http://www.apadivisions.org/division-41/education/guidelines

    http://www.apadivisions.org/division-41/education/guidelines

    http://dx.doi.org/10.1080/13854040903155063

    http://dx.doi.org/10.1007/BF01044769

    http://dx.doi.org/10.1007/BF01044769

    http://dx.doi.org/10.1037/a0019138

    http://dx.doi.org/10.1037/1076-8971.14.1.1

    http://dx.doi.org/10.1177/1073191103258591

    tools: Overview and critical analysis. In R. K. Otto & K. S. Douglas
    (Eds.), Handbook of violence risk assessment (pp. 1–17). New York,
    NY: Taylor & Francis.

    Institute of Law, Psychiatry, and Public Policy. (2012). Basic forensic
    evaluations: Principles and practice. Retrieved from http://www.ilppp
    .virginia.edu/OREM/AdultPrograms/Course/42

    Lally, S. J. (2003). What tests are acceptable for use in forensic evalua-
    tions? A survey of experts. Professional Psychology: Research and
    Practice, 34, 491–498. doi:10.1037/0735-7028.34.5.491

    Malesky, L. A., & Proctor, S. L. (2012). Training experiences essential for
    obtaining a forensic psychology postdoctoral fellowship. Journal of
    Forensic Psychology Practice, 12, 163–172. doi:10.1080/15228932
    .2012.650146

    McLaughlin, J. L. (2013). A look at how forensic evaluator personality
    traits relate to the use of tests in forensic assessment (Unpublished
    master’s thesis). Sam Houston State University, Huntsville, TX.

    Melton, G. B., Petrila, J., Poythress, N. G., & Slobogin, C. (2007).
    Psychological evaluations for the courts: A handbook for mental health
    professionals and lawyers (3rd ed.). New York, NY: Guilford Press.

    Mossman, D., Noffsinger, S. G., Ash, P., Frierson, R. L., Gerbasi, J.,
    Hackett, M., . . . Zonana, H. V. (2007). AAPL practice guideline for the
    forensic psychiatric evaluation of competence to stand trial. Journal of
    the American Academy of Psychiatry and the Law, 35, 3–72.

    Nicholson, R. A., & Norwood, S. (2000). The quality of forensic psycho-
    logical assessments, reports, and testimony: Acknowledging the gap
    between promise and practice. Law and Human Behavior, 24, 9–44.

    Nussbaum, D., Hancock, M., Turner, I., Arrowood, J., & Melodick, S.
    (2008). Fitness/competency to stand trial: A conceptual overview, re-
    view of existing instruments, and cross-validation of the Nussbaum
    Fitness Questionnaire. Brief Treatment and Crisis Intervention, 8, 43–
    72. doi:10.1093/brief-treatment/mhm026

    Otto, R. K., & Heilbrun, K. (2002). The practice of forensic psychology: A
    look toward the future in light of the past. American Psychologist, 57,
    5–18. doi:10.1037/0003-066X.57.1.5

    Otto, R. K., & Weiner, I. B. (Eds.). (2013). Handbook of psychology:
    Forensic psychology (2nd ed., Vol. 11). Hoboken, NJ: Wiley.

    Packer, I. K. (2008). Specialized practice in forensic psychology: Oppor-
    tunities and obstacles. Professional Psychology: Research and Practice,
    39, 245–249. doi:10.1037/0735-7028.39.2.245

    Packer, I. K. (2009). Evaluation of criminal responsibility. New York, NY:
    Oxford University Press.

    Quinsey, V. L., Harris, G. T., Rice, M. E., & Cormier, C. A. (1998). Violent
    offenders: Apprising and managing risk (2nd ed.). Washington, DC:
    American Psychological Association. doi:10.1037/10304-000

    Quinsey, V. L., Harris, G. T., Rice, M. E., & Cormier, C. A. (2006). Violent
    offenders: Appraising and managing risk (3rd ed.). Washington, DC:
    American Psychological Association. doi:10.1037/11367-000

    Rogers, R. (Ed.). (2008). Clinical assessment of malingering and decep-
    tion. New York, NY: Guilford Press.

    Rogers, R., Vitacco, M. J., & Kurus, S. J. (2010). Assessment of malin-
    gering with repeat forensic evaluations: Patient variability and possible
    misclassification on the SIRS and other feigning measures. Journal of
    the American Academy of Psychiatry and Law, 38, 109–114.

    Simon, M. J. (2007). Performance of mentally retarded forensic patients on
    the Test of Memory Malingering. Journal of Clinical Psychology, 63,
    339–344. doi:10.1002/jclp.20351

    Singh, J. P., Grann, M., & Fazel, S. (2011). A comparative study of
    violence risk assessment tools: A systematic review and metaregression
    analysis of 68 studies involving 25,980 participants. Clinical Psychology
    Review, 31, 499–513. doi:10.1016/j.cpr.2010.11.009

    Skeem, J. L., & Monahan, J. (2011). Current directions in violence risk
    assessment. Current Directions in Psychological Science, 20, 38–42.
    doi:10.1177/0963721410397271

    Slobogin, C., Rai, A., & Reisner, R. (2008). Law and the mental health
    system: Civil and criminal aspects (5th ed.). St. Paul, MN: Thompson/
    West.

    Tex. Gen. Laws. § 62–1.01. (2005). Retrieved from http://www.statutes
    .legis.state.tx.us/docs/CR/htm/CR.62.htm

    Tolman, A. O., & Mullendore, K. B. (2003). Risk evaluations for the
    courts: Is service quality a function of specialization? Professional
    Psychology: Research and Practice, 34, 225–232. doi:10.1037/0735-
    7028.34.3.225

    Varela, J. G., & Conroy, M. A. (2012). Professional competencies in
    forensic psychology. Professional Psychology: Research and Practice,
    43, 410–421. doi:10.1037/a0026776

    Viljoen, J. L., McLachlan, K., & Vincent, G. M. (2010). Assessing violence
    risk and psychopathy in juvenile and adult offenders: A survey of clinical
    practices. Assessment, 17, 377–395. doi:10.1177/1073191109359587

    Warner, R. M. (2008). Applied statistics: From bivariate through multi-
    variate techniques. Thousand Oaks, CA: Sage.

    Warren, J. I., Murrie, D. C., Chauhan, P., Dietz, P. E., & Morris, J. (2004).
    Opinion formulation in evaluating sanity at the time of the offense: An
    examination of 5,175 pre-trial evaluations. Behavioral Sciences & the
    Law, 22, 171–186. doi:10.1002/bsl.559

    Warren, J. I., Murrie, D. C., Stejskal, W., Colwell, L. H., Morris, J.,
    Chauhan, P., & Dietz, P. (2006). Opinion of formation in evaluating the
    adjudicative competence and restorability criminal defendants: A review
    of 8,000 evaluations. Behavioral Sciences & the Law, 24, 113–132.
    doi:10.1002/bsl.699

    Webster, C. D., Douglas, K., Eaves, D., & Hart, D. (1997). HCR-20:
    Assessing risk for violence (Version 2). Vancouver, BC: Mental Health,
    Law, & Policy Institute, Simon Fraser University.

    Weiner, I. B., & Hess, A. K. (Eds.). (2006). The handbook of forensic
    psychology (3rd ed.). Hoboken, NJ: Wiley.

    Weiss, R., & Rosenfeld, B. (2010). Cross-cultural validity in malingering
    assessment: The Dot Counting Test in a rural Indian sample. Interna-
    tional Journal of Forensic Mental Health, 9, 300–307. doi:10.1080/
    14999013.2010.526680

    Zapf, P. A., & Roesch, R. (2009). Evaluation of competence to stand trial.
    New York, NY: Oxford University Press.

    Received July 17, 2013
    Revision received January 13, 2014

    Accepted January 24, 2014 �

    T
    hi
    s
    do
    cu
    m
    en
    t
    is
    co
    py
    ri
    gh
    te
    d
    by
    th
    e
    A
    m
    er
    ic
    an
    Ps
    yc
    ho
    lo
    gi
    ca
    l
    A
    ss
    oc
    ia
    tio
    n
    or
    on
    e
    of
    its
    al
    lie
    d
    pu
    bl
    is
    he
    rs
    .
    T
    hi
    s
    ar
    tic
    le
    is
    in
    te
    nd
    ed
    so
    le
    ly
    fo
    r
    th
    e
    pe
    rs
    on
    al
    us
    e
    of
    th
    e
    in
    di
    vi
    du
    al
    us
    er
    an
    d
    is
    no
    t
    to
    be
    di
    ss
    em
    in
    at
    ed
    br
    oa
    dl
    y.

    135TEST USAGE IN FMHA

    http://www.ilppp.virginia.edu/OREM/AdultPrograms/Course/42

    http://www.ilppp.virginia.edu/OREM/AdultPrograms/Course/42

    http://dx.doi.org/10.1037/0735-7028.34.5.491

    http://dx.doi.org/10.1080/15228932.2012.650146

    http://dx.doi.org/10.1080/15228932.2012.650146

    http://dx.doi.org/10.1093/brief-treatment/mhm026

    http://dx.doi.org/10.1037/0003-066X.57.1.5

    http://dx.doi.org/10.1037/0735-7028.39.2.245

    http://dx.doi.org/10.1037/10304-000

    http://dx.doi.org/10.1037/11367-000

    http://dx.doi.org/10.1002/jclp.20351

    http://dx.doi.org/10.1016/j.cpr.2010.11.009

    http://dx.doi.org/10.1177/0963721410397271

    http://www.statutes.legis.state.tx.us/docs/CR/htm/CR.62.htm

    http://www.statutes.legis.state.tx.us/docs/CR/htm/CR.62.htm

    http://dx.doi.org/10.1037/0735-7028.34.3.225

    http://dx.doi.org/10.1037/0735-7028.34.3.225

    http://dx.doi.org/10.1037/a0026776

    http://dx.doi.org/10.1177/1073191109359587

    http://dx.doi.org/10.1002/bsl.559

    http://dx.doi.org/10.1002/bsl.699

    http://dx.doi.org/10.1080/14999013.2010.526680

    http://dx.doi.org/10.1080/14999013.2010.526680

      Test Usage in Four Common Types of Forensic Mental Health Assessment
      Method
      Participants
      Procedure
      Materials
      Results
      Discussion
      References

    Calculate your order
    Pages (275 words)
    Standard price: $0.00
    Client Reviews
    4.9
    Sitejabber
    4.6
    Trustpilot
    4.8
    Our Guarantees
    100% Confidentiality
    Information about customers is confidential and never disclosed to third parties.
    Original Writing
    We complete all papers from scratch. You can get a plagiarism report.
    Timely Delivery
    No missed deadlines – 97% of assignments are completed in time.
    Money Back
    If you're confident that a writer didn't follow your order details, ask for a refund.

    Calculate the price of your order

    You will get a personal manager and a discount.
    We'll send you the first draft for approval by at
    Total price:
    $0.00
    Power up Your Academic Success with the
    Team of Professionals. We’ve Got Your Back.
    Power up Your Study Success with Experts We’ve Got Your Back.

    Order your essay today and save 30% with the discount code ESSAYHELP