Introduction

Management of dysphagia for patients admitted to acute hospitals is a core component of quality care, as dysphagia often has a severe impact on patient health outcomes, including aspiration pneumonia, malnutrition and dehydration [13]. Allied health professionals, particularly speech language pathologists (SLP), are commonly an integral part of this assessment management process, and clinical bedside swallowing assessments (CBSA) are the foundation of their practice. CBSA provide information on the patient’s swallowing (dis)ability, inform decisions regarding further instrumental assessment, and are used to develop a management plan and monitor progress [46]. Specifically, SLPs aim to minimise risks to health by evaluating and ameliorating the risk of aspiration and maximise safe consumption of food and water. Strategies include diet modification, positioning, swallowing techniques or mealtime strategies. SLP intervention may also include therapy that aims to restore safe swallowing by improving airway protection mechanisms and oropharyngeal function. SLPs use the CBSA to inform this process by identifying the likelihood of aspiration, characterising the nature of the swallowing difficulty and making a determination regarding severity and risk [5, 6]. This informs decision-making regarding referral for further investigation, intervention planning and monitoring of progress.

However, there is considerable variation with regard to the components SLPs’ self-report as routinely included in this assessment process [711]. The few standardised protocols available are either screening tools [12, 13] or have been validated only on patients’ post-stroke [14]. In addition, the majority of research on clinical dysphagia assessment has narrowly focused on the utility of various combinations of examination items within the CBSA for reliably predicting if a patient will be identified as aspirating on instrumental assessments [15]. The broader aspects of SLPs’ CBSA practices have received limited attention and have focussed on standardising and validating assessment practices [14] or identifying what components SLPs self-report as including in their CBSA. Interestingly, SLPs themselves report that less than half of the components of bedside assessments that are considered important are used ‘usually/always’ [711]. These findings have led to a call for increased consistency in clinical practice via standardisation of the clinical examination to manage this variability. The assumption is that this will improve the quality of assessment practice and therefore patient outcomes, as well as yield reliable measures for intervention research.

The Mann Assessment of Swallowing Ability (MASA) [14] is the only current diagnostic bedside assessment tool that was originally designed to address the broader focus of CBSA by SLPs, having been standardised and validated with patients in a rehabilitation setting after their first stroke. However, SLPs also omit assessment components even when using a standardised and validated protocol like the MASA [16]. In a retrospective investigation into the utility of the MASA for diagnosing the presence/absence of aspiration in a mixed non-stroke population in an acute hospital, 79 of the 168 cases selected for review had up to five components omitted [16]. This is particularly striking given the finding that SLPs’ clinical judgement, as represented by ordinal ratings on completion of the MASA, had good sensitivity when compared to instrumental findings, even though they were frequently made on ‘incomplete’ bedside assessments. Gonzalez-Fernandez et al. [16] suggest that these patterns of component omission were non-random and were likely to be due to SLPs’ decision-making regarding a component’s perceived relevance for their diagnostic reasoning.

This pattern of omitting assessment components aligns with SLPs’ reported patterns of assessment practice. Furthermore, the suggestion that these patterns are non-random is supported by the considerable similarity across self-report studies regarding the components that 90 % of responding SLPs across Ireland, the UK, the USA and Australia reported that they do ‘usually/always’ use (see Table 1) [711]. Across these studies, it has been speculated that components may have been less frequently used because the research evidence is not strong or has ruled them out as irrelevant (e.g., gag reflex, cervical auscultation), the survey components were ambiguous (e.g., whether speech and language screening was conducted meant informally or via a standardised assessment) or SLPs may not have had access to required training (e.g., indirect laryngoscopy). When interviewed about their self-reported use of assessment components, respondents stated that they were influenced by their knowledge of the current evidence base, accepted practice in their workplace, the aetiology of the dysphagia and relevant medical conditions, and individual SLP characteristics, such as education and experience [11]. These respondents also noted that they would omit components if they thought the information derived would not add information that would change their management of the patient. This supports the suggestion that patterns of usage of clinical assessment components were logically based on a clinical decision-making process. The SLP would start broadly with the more commonly used components and probe with less commonly used components until a decision was reached regarding the diagnosis and intervention plan or whether further instrumental assessment was required [8]. This process mirrors those used in other disciplines, such as medicine, where clinical examinations and history taking are guided by an overall evidence-based framework to gather information to make a diagnostic judgement or decision for referral for further instrumental assessment; for example, in the clinical assessment of pneumonia [17]. Therefore, good CBSA practice by SLPs is characterised by an efficient decision-making process that determines which components are relevant for a specific patient presentation and therefore should be used. This hypothesis appears to be borne out by the overall similarity in the assessment components that SLPs self-report using during CBSA across studies conducted in Ireland, the USA, the UK and Australia (see Table 1) and may account for the variability in using some components with all patients.

Table 1 Clinical bedside assessment components that SLPs self-report using

In summary, self-report and retrospective research has found that SLPs omit assessment components that are considered by experts to be important for quality CBSA. This pattern may be intentional and may be related to factors important for high-quality clinical practice. In particular, it suggests engagement in a logical and efficient decision-making process that determines which history and clinical examination components are relevant given the patient’s presentation and should be used during an assessment. However, the actual patterns of assessment practice by SLPs using a standardised CBSA protocol designed by clinicians to be relevant for use with patients in an acute care hospital remains to be evaluated. This study aimed to prospectively identify these assessment patterns and sought to answer the following research question: what components of an SLP-developed standardised CBSA protocol are consistently used in practice by SLPs in an acute care hospital with patients who have mixed aetiologies?

Methods

Research Environment

This research was conducted at a 600-bed acute hospital in a major Australian city. Admitted patients who are suspected of having dysphagia are routinely referred to a SLP, who is part of a multi-disciplinary team and who conducts the CBSA. The area health service ethics board was consulted prior to commencement of data collection and advised that the proposed data collection constituted a quality improvement activity and did not require ethics review. The research team adhered to the Australian National Health and Medical Research Council guidelines [18] and the Declaration of Helsinki for ethical conduct [19].

Bedside Assessment Protocol

SLPs identified that there were variations in how CBSAs were being conducted within their group and developed a consensus on the components to be included in the trial protocol that closely matched typical practice. The selected components represented two aspects of the assessment:

  1. 1.

    Procedural and established best practice [20] elements; e.g., history taking, summary of assessment findings, recommendations and plans [6].

  2. 2.

    Direct assessment of dysphagia with components derived from:

  3. a.

    The MASA: validated components and assessment processes (see description below) were identified as potentially appropriate for use with a heterogeneous acute care patient group and matched current CBSE practices [14].

  4. b.

    Cranial nerve examination to determine the precise nature of impairments and underlying neurological correlates [6, 21].

A standard protocol combining these assessment items in an order that reflected the current practice of the SLPs at that hospital and which was suitable for inclusion in medical files was developed based on consensus. This protocol, operationally termed the Speech Pathology Initial Dysphagia Assessment (SPIDA), was trialed and modified twice until the current format was established. The scoring process and the 3–5 point rating scales for the MASA components remained as per the technical manual. Additional cranial nerve assessment components were rated as ‘normal/abnormal’. Provision was made for comments against all aspects of the assessment process. See “Appendix 1” for the pro forma SPIDA. All SLPs participated in a 2-h training session on using the protocol and were instructed to use all assessment components with all patients.

Sampling

The SLPs were instructed to use the SPIDA for all initial swallowing assessments, regardless of diagnosis, for the 6-month data collection period. All SLPs working in the five tertiary hospitals in the state are evaluated against the same state-wide acute dysphagia competency framework [20]. The SLPs in this study were credentialed for independent practice in acute dysphagia assessment and management using this framework and had satisfactorily demonstrated competence endorsed under supervision. The SLPs recorded their assessment during and immediately after seeing the patient, photocopied it, placed the original in the patient’s medical records and filed the copy in a secure collection box on their return to the SLP offices.

Data Entry

Data were entered into Microsoft Access® database (Microsoft, 2007). SLP students on research internships (N = 6) double-entered all data on each SPIDA under the supervision of the authors. Fidelity of data entry was further checked by a professional staff member with expertise in Access®.

Analysis

Descriptive analyses were conducted using SPSS® (IBM, 2013) to investigate patterns of component use and rating behaviours.

Results

Return Rate

An estimated return rate was calculated retrospectively and conservatively, based on the number of patients admitted to the hospital whose files were centrally coded as having been referred for a dysphagia assessment and who had no previous assessments during their admission. The maximum number of inpatients who could have been included in this data collection was 516. The total number of patient SPIDA forms submitted for analysis was 308, a minimum estimated return rate of 60 %. SLPs stated that forms were not returned due to practical constraints (e.g., forgot to bring the SPIDA to the bedside or photocopy it) or because they chose an alternative assessment process tailored to the patient’s specific presentation (e.g., if the patient was intubated or non-compliant with standard assessment processes due to cognitive deficits).

Population

The population was representative of patients usually referred for dysphagia assessments at this service. They had mixed diagnoses and were elderly, with 75 % of the sample aged 73 years or above and a median age of 82 years. The majority of the sample stayed longer than average for this hospital (75 % admitted six or more days), with a median stay of 12 days and 40 % being considered long-stay (>12 days). Sixty-two patients (20.1 %) were admitted with a first-time stroke.

Patterns of Assessment Component Use and Ratings

Procedural and Practice Components

The majority of these components (17/26) were completed by the SLPs more than 90 % of the time (see Table 2). Six of the nine elements completed less than 90 % of the time relied on information being available from case notes and/or colleagues; three related to SLP-directed practices (SLP-associated diagnoses, long- and short-term goals).

Table 2 Completion rates for each procedural and practice component

Dysphagia Examination Components

Seven sets of data were removed from the total data set of 308 as the assessment had not proceeded to a direct assessment of swallowing function. Tables 3 and 4 identify the rating patterns for all components. Table 3 indicates that the SLPs were highly variable in completing the cranial nerve components, with five out of ten items having been rated less than 90 % of the time: masseter palpation, reflexive spontaneous swallow, facial quadrants, pharyngeal sensation and tongue protrusion symmetry. The MASA components were more consistently used, with three out of 24 items having been rated fewer than 90 % of the time: tongue strength, gag reflex and soft palate (see Table 4). However, some rating categories that were available for the MASA components were rarely used, with 20 of the 24 components having rating categories that attracted fewer than ten selections (see italic values, Table 4). In addition, different combinations of items were omitted for different patients, and as a result only 18.6 % (56) of patients had a judgement recorded for all cranial nerve components and only 55.8 % (168) had all MASA components rated. Overall, there were only 7/301 forms returned where all components of the dysphagia examination were completed.

Table 3 Rating categories for cranial nerve components (N = 301)
Table 4 Rating categories used by SLPs for each MASA component (N = 301)

Comments

All components had comments in addition to ratings, with the percentage ranging from 2.3 % through to 47.1 % of items having been commented on. Eight of the ten cranial nerve ratings had 25 % or more commented on, with seven of the 24 MASA-derived components having more than 25 % comments.

Discussion

This research prospectively evaluated the CBSA components that SLPs select as important for inclusion in a standardised clinical bedside assessment, and their patterns of use for these components when assessing patients referred for initial dysphagia assessment in an acute hospital over a 6-month period. The results indicate that SLPs included components that captured the process of assessment from prior to the bedside clinical examination (e.g., social history), observations made during the direct clinical examination, which included trials of food and fluids at the bedside, as well as components that capture the diagnostic and treatment planning elements of the process. It is worth noting that these assessment components are in agreement with the CBSA components identified by SLPs across multiple countries as being used 90 % or more of the time and therefore reflect current international best practice.

Overall, the SLPs were highly likely to use most of the components, but did not necessarily use all of them, or all of the available rating categories, with every patient. It was not unusual for comments to be made in relation to many of the items. The documented variability in item use was unexpected given that the SLP team were involved in the selection of components to be included in the SPIDA, were instructed to use all components, and were aware of the importance of adhering to a standardised and validated protocol as a result of the training for this research as well as their own pre-professional education.

Assessment Components Included by SLPs

Following iterative consultation with, and review by, the participating SLPs, the SPIDA included items that map the entirety of the bedside assessment, including components related to both process and outcomes. This confirms that SLPs see their assessment as an evolving process, more than simply identifying the presence/absence or risk of aspiration at the bedside in accordance with best practice [5, 6]. Specifically, SLPs are gathering a wide range of data from a range of sources, including case notes, the patient and their caregivers, and colleagues, as well as a clinical bedside assessment that comprises oral motor assessment elements as well as oral intake trials. The SLPs in this study used this data for multiple purposes, including to identify aspiration risk, to determine dysphagia severity and to develop recommendations for short-term care as well as to assess the prognosis and the need for further bedside and/or instrumental assessment. This range of assessment purposes is consistent with recommended practice that CBSAs should include medical and patient history, assessment of cognition, communication, relevant physiology, anatomy and functioning, including cranial nerve examination, and assessment of ability to consume food and liquids [6].

Variability in Assessment Use

Components Assessed

The SLPs in this study used 46 of the 61 (75 %) components of the SPIDA 90 % or more of the time, suggesting that they had included components that were relevant to their practice and that they were conscientious in their professional practice and use of the tool. There are a number of plausible explanations for the less frequent use of the remaining 15 items. The SLPs may not have had access to the information required to complete some components, such as pre-admission diet, or some tests may not have been conducted for some patients at the time of the CBSA; for example, chest X rays, C reactive protein (CRP) and white cell count (WCC). It is also possible that there may not have been enough information gathered at this initial bedside assessment to complete a diagnosis and set goals. Some components from the MASA may have been neglected because their use is not supported by research evidence. For example, the SLPs interviewed by Vogels and colleagues [11] indicated that the evidence base influenced the components they included in a CBSA and cited the elicitation of the gag reflex as an example. In line with this, the gag reflex was assessed in 71.4 % of the assessments in the current study and in 57.7 % of assessments in the research presented by Gonzalez-Fernandez and colleagues [16]. Cranial nerve components may have been used less frequently due to lack of relevance to the specific aetiology of the dysphagia (i.e., stroke or not stroke-related), or could not be completed (e.g., a hypersensitive gag or a resistant patient will make completion of the pharyngeal sensation component difficult). In addition, the relevant information may have already been observed during elicitation of the MASA components of the assessment and therefore excluded from further probing (e.g., reflexive spontaneous swallow, tongue protrusion symmetry). Some components may have been omitted because the patient was unable to cooperate for particular items due to alertness, cooperation or ability to follow instructions. This may also account for why so few patients with severely impaired alertness (7.3 %) and cooperation (4.7 %) were included in this sample, as the assessment would not have proceeded to a full clinical examination. In turn, this was also reflected in the overall return rate of 60 % of SPIDA forms.

It is possible that not including an ‘unable to assess’ option for the majority of the clinical assessment components (24 of 34 components) may have also contributed to SPIDA forms not being completed/submitted for analysis. The pattern of use of the comments option supports this notion, in that some of the least frequently recorded items were associated with 25 % or more use of the comments section, although this pattern was not consistent across all items. Comments were made against all assessment components on 2.4–46.8 % of assessment occasions, with 16 of the 34 clinical components attracting comments on 25 % or more occasions. The frequent use of the comments option indicates that SLPs found it necessary to add to, or qualify, the judgements they were making on a regular basis, suggesting that the ratings of these components did not always best reflect the information gathering process they were undertaking.

Assessments Completed

SLPS did not complete all assessment components in any of the 308 SPIDA protocols submitted for analysis. The completion rate for all MASA components (44 %) was very similar to previously reported rates of 47 % [16]. Overall, for those forms where a physical examination was conducted (N = 301) the clinical components of the SPIDA were only fully complete for seven assessments. The observed variability of assessment item use by graduate SLPs using a protocol they themselves developed in a research study, where they were trained to complete all assessment components, lends weight to the hypothesis by Martino et al. [8] that this may not be random. In line with this hypothesis, the pattern of item use in this study may reflect a hierarchical decision-making process where some components were frequently used but then others were selected on the basis of their relevance for a specific patient. This observation is also in agreement with comments by SLPs, who stated that they used assessment components if they felt they would usefully contribute further information to the diagnostic process [11].

This merits further consideration, given that Gonzalez-Fernandez et al. [16] found that although SLPs did not provide a rating on all components for 47 % of their cases when using the MASA, their overall judgement had good sensitivity when compared to the findings of a videofluoroscopic swallowing study. This fits the approach to clinical assessment used by other health professionals, which have similar outcomes. For example, general medical practitioners make accurate judgements as to whether a patient with an acute cough may or may not have pneumonia based on their case history and clinical examination findings [17]. Likewise, the way in which SLPs interact with CBSA protocols is likely to be influenced by the clinical reasoning processes they employ and their clinical judgement based on previous professional experience. As such, their overall clinical assessment, informed by their observations of the patient’s skill across a variety of assessment components combined with their clinical expertise, may be more accurate than calculating ratings from a strictly administered standardised protocol.

It is well established that clinical reasoning involves the exercise of professional judgement and arises from a combination of different types of knowledge that are applied to diagnostic decision-making in a variety of ways that changes with experience. This knowledge base includes a network of propositional and clinical knowledge that is used by novices to explain the causes of signs and symptoms. With experience, more expert clinicians organise this underlying network of knowledge and experience into overarching illness scripts or patterns of clinical symptoms that they quickly recognise [22]. The clinician uses these to inform the process of data collection and diagnostic decision-making to rule the clinical hypothesis(es) in or out [2327]. Consequently, decisions are made, frequently intuitively, regarding what information should be gathered to test a restricted number of hypotheses. The current study suggests that SLPs exercise similar clinical reasoning processes to determine which components to assess during a CBSA rather than adhering strictly to a prescribed or standardised assessment process. Understanding the SLPs’ expertise and supporting the diagnostic and decision-making process may better support quality assessment compared to constraining the process to a standardised set of data-gathering steps.

Categories Rated Within the MASA Components

SLPs rarely used some rating categories that are available within the MASA components, with 20 of the 24 components having rating categories that attracted fewer than ten selections (see italic values, Table 4). This suggests that these categories are not adding meaningful information to the respective measure, and that therefore it may be argued that these categories could be discarded for these items [28]. There are a number of potential explanations for this finding, all of which require further investigation. For example, the MASA was validated on patients who had experienced their first stroke, whereas the present study was conducted on first assessments for a heterogeneous acute care population. Therefore, it is possible that the rating categories are not all relevant for mapping a continuum of severity for all members of a heterogeneous patient population.

Some of the observed rating behaviours also may arise because SLPs, or indeed any human rater, are not objective measurement instruments and therefore their ratings arise from an interaction between themselves and the tool [29]. For example, research on how rehabilitation teams interact with standardised assessment items has clearly identified that rating decisions are negotiated to better represent patient progress and this is influenced by the clinical reasoning processes employed [30, 31]. Therefore, it is possible that the SLPs may have consistently used categories that best represented their clinical judgement and neglected others. For example, it may be that the three MASA categories consistently used to rate chest status [chest infection (2), fine basal crepitations (6) and chest clear (10)] best represented this group of SLPs’ implicit continuum of severity ratings and therefore they did not use the interim categories [coarse basal crepitation, chest physiotherapy (4) and sputum upper airway, other condition (8)]. It is also possible that the SLPs may have been making a judgement that the information in rarely used categories did not contribute meaningful information to their assessment process. For example, the SLPs only consistently used two of the five available categories for the palate component (no spread or elevation and NAD) which may suggest they treated this as a present/absent judgement rather than a continuum of severity.

Limitations and Future Directions

Conducting this research as a quality improvement process ensured that data were collected in the real world of clinical practice rather than a controlled validation trial. However, this also meant that return rates were modest and may have been influenced by non-random factors and that information was not collected that may have explained omitted ratings and non-return of forms. For example, the latter may have been due to assessments being discontinued and therefore forms not returned on patients with more severe or complex presentations and poor alertness.

The competence and level of expertise of the SLP team was assumed based on their credentialing status rather than formally assessed prior to inclusion in this study. Furthermore, the experience level for each SLP was not recorded for each returned form which precluded investigating this as a factor influencing assessment choices.

This study only captured the initial bedside assessment; which, depending on the patient’s presentation, may be only the start of information gathering to inform decision-making. This could account for some components that were utilised on less than 90 % of occasions; for example, those related to decision-making such as SLP-associated diagnoses and long- and short-term goals. Conversely, component usage was on average higher than in self-reported studies; for example, there was relatively high compliance with examining the gag reflex (71.4 % in this prospective study compared to 56.4 % in the retrospective study by Gonzalez-Fernandez et al. [16] and less than 20 % in more self-report studies [9, 11]). SLPs’ compliance may have been higher due to their awareness of the need to adhere to a prospective research protocol and standardised assessment process. Actual practice may be more discriminating given that SLPs have commented that they carry out examination components based on research evidence and their judgement as to whether it will contribute useful information to the assessment picture [11].

The SPIDA was designed by the clinicians to support their clinical practice and therefore had limitations as an assessment tool that was not apparent prior to completion of the data collection. A number of likely SLP responses were not able to be recorded including: unable to assess; already noted; assessment discontinued; noting that further assessment was required; or that an item was omitted as it would not contribute relevant information.

Finally, this study protocol focussed on the assessment information collected during the initial CBSA which is only one component of quality care. It also identified that SLPs use the data for multiple purposes including identification of aspiration risk, dysphagia severity, prognosis and identification if further data collection (instrumental and/or clinical) is required to inform diagnosis and management decisions. These decisions would have been monitored and reviewed based on patient outcomes. Further research is needed to determine how effectively the CBSA contributes to this and whether the variability of CBSA practices by SLPs positively influences patient outcomes.

Notwithstanding these limitations, the findings of this study suggest a number of useful lines of inquiry to address concerns regarding the variability of CBSA practices. In particular, it may be that the processes required for high-quality CBSA conflict with establishing a baseline for measuring change through strict adherence to a standardised assessment tool. This is not to say that the CBSA should not be subject to rigour, however, many clinical assessments are conducted by medical and allied health professionals and are accepted as an integral part of high-quality health care. These clinical assessments rely on high-quality clinical examination processes and are informed by the exercise of professional reasoning to interpret findings to guide and inform professional judgement and action. Investigating the clinical reasoning SLPs use during CBSA will identify whether it would be more fruitful to develop strategies to support this process rather than further refinement and/or development of standardised and psychometrically validated assessment tools.

The development and refinement of CBSA components that effectively support the process of consistent and valid clinical judgement may benefit from further investigation of actual assessment processes in the real world of clinical practice with SLPs of various levels of experience and expertise. Lines of inquiry could include identifying assessment content and processes, including further information gathering, and how aspects of dysphagia are being evaluated and the degree of impairment judged. It is also necessary to develop clarity about the decisions being made, and therefore, the purpose of the CBSA; for example, identifying dysphagia type, severity, predicting aspiration risk, determining the impact and suitability of a range of intervention and management strategies, contributing to differential diagnosis or other aspects, or all of these combined. Further investigation of how well component description generalises across SLPs and identification of more continua of severity for assessment components that match clinical judgement may assist with development of baselines against which to measure change. Finally, understanding how the that data gathered during the initial CBSA relates to other data gathered about the patient and contributes to diagnosis and care will be critical to ensure that quality clinical examinations are conducted.

Summary

In summary, the observations of the current study suggest that SLPs will prioritise and exercise their clinical reasoning during a bedside assessment of dysphagia over following a strict item-based assessment protocol. Evidence that SLPs contribute positively to dysphagic patient outcomes [3] indicates that this clinical reasoning and decision-making process may be a sounder approach than following a strict item-based protocol or checklist. This would be in accordance with research on medical clinical reasoning that has found that diagnostic reasoning is not a linear process and is influenced by experience [2224, 27]. Finally, there were indications that the SLPs’ assessment of swallowing was not limited to determining the presence/absence of dysphagia and aspiration risk or measuring its severity. SLPs also included components to describe the type of dysphagia (e.g., oral and/or pharyngeal) and to recommend management, including diet/fluid modification, medication management, mealtime strategies and a care plan for the patient. This confirms that the SLPs use the CBSA for multiple purposes and the way in which they engage in clinical reasoning to do this is likely to be complex. The utility of standardised item-based assessments and the nature of the clinical reasoning employed during CBSA, the quality of the data yielded and their impact on patient care need further investigation in this context.

Furthermore, the concern consistently expressed in the research about SLPs’ inconsistent use of recommended CBSA components may be misplaced. It is possible that exerting control over the clinical reasoning and decision-making process by standardising and validating assessment protocols may not improve the quality of the diagnostic process. The findings of this research suggest that a standardised assessment did not match the clinical reasoning engaged by SLPs during a CBSA, as they select relevant components to assess and judgements to record. Although it is not yet clear what this clinical reasoning process is, it is well accepted and indeed recommended, that the patient and context should determine the assessment process [6] and the CBSA does support better outcomes for patients [3]. Standardised item-based assessments may constrain the clinical reasoning process and the multiple purposes of the assessment further confuse the issue. Evaluation of what and why SLPs do what they do and how CBSA assessment data contribute to the assessment and management of patients in acute care is warranted. This will enable identification of ways to support high-quality assessment processes that yield data that will improve patient outcomes, establish baselines for evaluating effectiveness of interventions and ensure quality pre- and post-professional training and practice. It is likely that awareness of one’s own clinical reasoning processes will increase diagnostic specificity and sensitivity, and in turn improve clinical management and outcomes for patients referred for dysphagia assessments in acute care hospitals.