ACQ Vol 13 no 2 2011

components can be captured, quantified, and then targeted directly through intervention. The Living with Aphasia: Framework for Outcome Measure- ment (A-FROM; Kagan et al., 2008) is a conceptual framework that builds on the ICF. The four key domains (Severity of disorder; Participation in life situations; Communication and language environment; and Personal identity, attitudes and feelings) are represented as intersecting circles, with the point of overlap constituting “life with aphasia” or quality of life (Kagan & Simmons-Mackie, 2007; Kagan et al., 2008). While the conceptual framework was developed for use with clients with aphasia, it has potential for use with any client group or disorder. Routinely used assessment tools can be mapped on to the domains of the ICF or A-FROM, to ensure that measurements are holistic and capture function at each level (Kagan & Simmons-Mackie, 2007; Kagan et al., 2008; McLeod & Threats, 2008). Psychometric properties of assessment tasks While the ICF and A-FROM provide overarching conceptual frameworks to guide assessment, an evidence based practitioner must still consider the validity, reliability, and psychometric make-up of the individual assessment tools or methods selected. This can be a daunting and time- consuming task in clinical practice; however, it is a critical component of reliable and valid assessment practice. Evaluation of psychometric properties is particularly important when assessment is being used to serve screening or diagnostic purposes. Screening tools aim to provide a quick and efficient means of identifying the presence or absence of a disorder while more comprehensive assessment or diagnostic batteries seek to profile and classify impairments and provide indices of severity. It is critical that clinicians consider features such as the extent to which the test measures what it is designed to measure ( validity ), whether the test provides representative sampling of the domain of behaviours ( content validity ), whether it has strong theoretical and empirical foundations ( construct validity ), whether its scores are reproducible and consistent ( reliability ), and whether it has sufficient sensitivity and specificity to detect the behaviours in question (Tate, 2010; Turkstra et al., 2005). Sensitivity values reflect the percentage of people with a disorder correctly identified by a given test or diagnostic procedure according to a reference standard (Dollaghan, 2007). Specificity values reflect the percentage of people without the disorder that are correctly identified as such (Dollaghan, 2007). The small number of systematic reviews that do exist in the literature have highlighted that many of the tests and measures used by speech pathologists have strong content and face validity (i.e., they are thoughtfully and carefully constructed); however, the construct validity is often weaker (Turkstra et al., 2005). Furthermore, many of the screening tools that are available, such as those for aphasia, provide insufficient reliability, validity and sensitivity/specificity data to make a true assessment of their clinical utility (Koul, 2007). These are again issues that need to be addressed by the field and considered in practice. It has been acknowledged that psychometric appraisals can be difficult and time-consuming for clinicians to complete in practice, yet there are useful guides available in the literature. For example, Dollaghan (2007) provides a practical and useful framework for the critical appraisal of diagnostic evidence (CADE). It allows the evaluation of screening tools and standardised batteries designed specifically for detection of a disorder, differential diagnosis

and classification. It allows clinicians to make reasoned judgements about the validity, reliability, sensitivity, and overall utility of the tool in question, supporting the quest for evidence based assessment. While the CADE framework allows evaluation of a tool’s ability to diagnose or classify a particular disorder, it does not identify how effective or useful the tool is in directing goal-setting or treatment. As a result, in practice, clinicians also require a process for determining which assessment tools and measures can be used to direct goal-setting and enhance treatment outcomes (Dollaghan, 2007). Such targeted research is relatively absent in the evidence based practice literature and clinicians are again encouraged to return to their theoretical frameworks to ensure coherence between their overarching goals and selection of assessment measures. Ecological validity It is widely recognised that performance on standardised language batteries such as the Western Aphasia Battery – Revised (WAB-R; Kertesz, 2006) and the Clinical Evaluation of Language Fundamentals (4th ed., Australian); (CELF-4 Australian; Semel, Wiig, & Secord, 2006) does not reflect real-life communication skills (Apel, 1999; Turkstra et al., 2005). Chaytor and Schmitter-Edgecombe (2003) state that ecological validity “refers to the degree to which test performance corresponds to real world performance” (p. 182). An important distinction should be made between the content and construct validity of a test and its ecological validity. In other words, a standardised test may have strong psychometric properties with little real world relevance. It is promising that an increasing number of functional communication measures are being developed in the field. However, surveys of speech pathology services suggest that impairment-driven batteries remain the most commonly used assessments in clinical practice (Verna et al., 2009). Verna et al. (2009) found that 92.8% of their 70 respondents routinely used impairment-based language assessments, while only 21.4% included measures of functional communication and 2.9% of clinicians completed discourse analysis. Expert consensus supports a shift in practice, viewing standardised assessments as “only one component of an evaluative process that includes multiple sources of information” (Turkstra et al., 2005, p. 220). As a profession we need to continue developing and increasing the use of functional, dynamic assessment tasks to supplement the data obtained from standardised tests (Turkstra et al., 2005). Considering client values and perspectives Our final, but perhaps most important, point of discussion requires reflection on the role that client values and perspectives play in evidence based assessment. Kagan and Simmons-Mackie (2007) suggest that the selection of assessment tools should be guided by the “real-life outcome goals” (p. 309) that are relevant to each individual client. This approach stands in stark contrast to the impairment-driven or traditional assessment. The desired end point is likely to be different for each client and is expected to change and evolve over time (Kagan & Simmons-Mackie, 2007). The uniqueness of each person’s situation highlights the need for a tailored assessment approach that considers the desired end point from a functional perspective, with life participation in mind (Kagan & Simmons-Mackie, 2007). Kovarsky (2008) presents an interesting discussion on the use of “personal experience narratives” when

86

ACQ Volume 13, Number 2 2011

ACQ uiring Knowledge in Speech, Language and Hearing

Made with