Abstract
The assessment of pharyngeal residues during fiberoptic endoscopic evaluation of swallowing (FEES) is based on visual-perceptual scales that involve clinical subjectivity. Training might be helpful to increase agreement among clinicians. This paper aims to assess the efficacy of training for the assessment of pharyngeal residue in FEES frames and videos through the Yale Pharyngeal Residue Severity Rating Scale (YPRSRS). Twenty-nine clinicians (Phoniatricians, Otorhinolaryngologists, Speech and Language Pathologists) and 47 students in Speech and Language Pathology participated in this study. Fourteen clinicians were randomly allocated to the training group, whilst the remaining 15 served as a control group; all the students participated in the training. Participants scored 30 pairs of videos and frames using the YPRSRS twice, before and after the training for the training groups and at least two weeks apart for the control group. Construct validity, defined as the agreement between each rater and the experts’ scores, and inter-rater reliability were compared among the groups and between the first and the second assessments to verify the efficacy of the training. Construct validity significantly improved at the second assessment in the training group for the pyriform sinuses videos (baseline 0.71 ± 0.04, post-training 0.82 ± 0.05, p = .049) and in the students’ group for the valleculae (baseline 0.64 ± 0.02, post-training 0.84 ± 0.02, p < .001) and pyriform sinuses videos (baseline 0.55 ± 0.03, post-training 0.77 ± 0.02, p < .05). No significant differences were found in the inter-rater reliability in any group. In conclusion, the training seems to improve participants’ agreement with experts in scoring the YPRSRS in FEES videos.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
The fiberoptic endoscopic evaluation of swallowing (FEES) is a standard procedure to study swallowing. In 1988, the first FEES report was published by two speech and language pathologists (SLPs) and an Otorhinolaryngologist [1]. Several professionals share the FEES procedure in different countries. In Anglo-American countries, it is predominantly performed by SLPs [2, 3]; in European countries, FEES is mainly performed by Phoniatricians and Otorhinolaryngologists as a medical procedure [3]. Together with the videofluoroscopic study of swallowing (VFSS), FEES is widely considered the gold standard instrumental exam [4] 1. The analyses of FEES recordings allow clinicians to identify signs of impairments in swallow safety (penetration and aspiration) and swallow efficacy (pharyngeal residues) [5].
Various measurement scales are currently available to analyze FEES videos [6]. The assessment of FEES recordings is usually based on visuoperceptual measures; this implies that evaluations of FEES videos are fundamentally subjective. Visuoperceptual scales created for interpreting various signs of dysphagia in FEES are often accompanied by specific training to learn how to use them [7,8,9,10,11]. Neubauer et al. developed the Yale Pharyngeal Severity Rating Scale (YPRSRS) [9], an image-based scale to assess the amount of residue in the valleculae and pyriform sinuses. The severity levels on a 5-point scale are defined “none, trace, mild, moderate, severe” for the valleculae and the pyriform sinuses. An operational description, an anchor image, and a percentage of residue are provided for each point. In a systematic review comparing pharyngeal residue severity rating scales for FEES [12], the YPRSRS was judged to be the most reliable and valid scale. In the original study, 20 raters (Otorhinolaryngology residents, Otorhinolaryngologists, and SLPs) attended a training that included written definitions, visual depictions, explanations, and clarification of doubts [9]. Similarly, a short training on the use of the scale on FEES frames was provided to raters in the validations of the YPRSRS in German [10], Turkish [11], and Italian [13]. The German training [10] lasted 8 min, while the Italian training lasted 4 min [13].
Training in analyzing FEES is a relevant topic, recently reviewed in a scoping sudy [14]. Several post-basic trainings on FEES were designed for exam interpretation and execution by SLPs, Phoniatricians, Otorhinolaryngologists, or other professions [15, 16]. Other training programs targeted different professionals, such as medical student residents [17], neurologists physicians [18], and nurses [15]. The training duration ranged from a minimum of 30 min [18] to over 10 h [19]. A few studies used video-recorded lectures [15] or online lessons [17], and some programs have included self-paced exercises [15, 16, 19], and self-assessments on skills learned [15, 16].
In the validation studies on YPRSRS, the effects of the training was evaluated only on clinicians [9,10,11, 13], and no comparison was made between ratings of clinical and inexperienced raters such as students. Furthermore, previous papers did not include how to identify and assess pharyngeal residues by using YPRSRS on FEES videos [9,10,11, 13]. Recently, the YPRSRS has been shown to be reliably used to evaluate pharyngeal residues also in FEES videos, although with average lower reliability coefficients compared to FEES frames [20]. Therefore, training on how to apply the YPRSRS on FEES video is required in order to improve its scoring in these circumstances.
The purpose of the present study was to: (i) develop a training aimed at acquiring skills for the interpretation and evaluation of pharyngeal residue in FEES videos and frames using the YPRSRS scale for medical doctors, SLPs, and students who attend the bachelor’s degree in Speech and Language Pathology; and (ii) verify the training efficacy in improving construct validity and inter-rater reliability. The hypothesis was that training could support participants in developing specific skills in assessing pharyngeal residues in FEES.
Methods
This project was carried out following the Declaration of Helsinki of the World Medical Association (WHO). Consent of the Ethics Committee of the University of Milan was obtained on 17/11/2020 (number 102/20). Frames and videos used in this work were selected from pre-existing archived material. A randomized controlled trial was conducted among clinicians, whereas a prospective observational pre-post study was performed in students.
The Yale Pharyngeal Residues Severity Rating Scale
The YPRSRS is an ordinal scale that rates the amount of pharyngeal residues in the valleculae and pyriform sinuses [9]. The definitions of severity are distributed on a 5-point scale (1 = none, 2 = trace, 3 = mild, 4 = moderate, 5 = severe). Each level of the scale corresponds to an operational description, an anchor image, and a percentage of residue. A separate score is provided for the valleculae and the pyriform sinus.
Selection of the Materials
Seventy pairs of videos and frames with different consistencies were selected from FEES video-recordings collected for previous studies. All FEES examinations were kept anonymous. The FEES were conducted using a XION EF-N flexible fiberscope (XION GmbH, Berlin, Germany) attached to an EndoSTROBE camera (XION GmbH, Berlin, Germany) and recorded as an AVI format. A standard FEES protocol was used including the sequential administration of boluses of thin liquids (5-10-20 ml of blue-dyed water x 3 trials for each volume; International Dysphagia Diet Standardisation Initiative – IDDSI 0; < 50 mPa·s at 50s-1 and 300s-1), pureed food (5-10-20 ml of pudding x 3 trials for each volume; IDDSI 4; 2583.3 ± 10.41 mPa·s at 50s-1 and 697.87 ± 7.84 mPa·s at 300s-1), and regular food (half biscuit x 2 trials; IDDSI 7) [21]. For the present study, only the 5 ml trials were selected for thin liquids and pureed food. For all consistencies, a frame was selected after swallowing the last bolus. As a first step, the 70 pairs of videos and frames were independently assessed by two experienced raters (> 10 years of experience in FEES analysis), a Phoniatrician, and a SLP. Only the pairs of videos and frames that were assigned the same YPRSRS score were selected. Thirty pairs of videos and frames (15 for valleculae, 15 for pyriform sinuses) were selected for for validity and reliability analysis, and an additional 6 pairs (3 for valleculae, 3 for pyriform sinuses) were chosen for the training by consensus. All YPRSRS scores and consistencies of the recorded swallows were included in the analysis.
Raters
A sample size of 29 clinicians was determined based on the previous studies on the YPRSRS. Inclusion criteria were professional activity as Phoniatricians, Otorhinolaryngologists, SLPs, or resident Otorhinolaryngologists with a minimum clinical experience of 1 year in the dysphagia. In addition, a convenience sample of students who attended the 2nd year of the bachelor’s degree in Speech and Language Pathology was recruited. To participate in the study, all students had to have already attended the classes on dysphagia assessment and treatment.
Training
Clinicians were randomly allocated to the training or the control group (1:1) based on a random number sequence. Randomization was stratified for the profession (medical doctor vs. SLP) and years of experience (< 5 years vs. 5 ≥ years), according to a previous study [10]. All students received the training.
The training design was based on the characteristics of the trainings for the interpretation of FEES previously described in the literature [15,16,17,18,19, 9, 10]. It consisted of 3 steps, for a total of 160 min. As a first step, the participants viewed a pre-recorded 75-minute video-audio lesson composed of theoretical modules describing dysphagia signs observed during the FEES procedure, pathophysiological mechanisms and consequences of pharyngeal residues, an introduction to the YPRSRS, the clinical application of the scale, and case studies. After the online lesson, each participant independently practiced assigning a YPRSRS score to 6 pairs of frames and videos of FEES (3 for valleculae, 3 for pyriform sinuses), not included in the pre-post training assessment. The results of individual practice and any questions or uncertainties were discussed among participants during a one-hour synchronous meeting moderated by an expert tutor. Debriefing meetings consisted of small groups (4–5 participants) were formed to encourage communication and discussion.
The whole training was delivered using Microsoft Teams (Microsoft Corporation, Redmond, WA).
Data Collection
Both the training and the control group scored the 30 pairs of videos and frames using the YPRSRS twice: (i) before and after for the training for the training group and the students, at least two weeks between assessments; (ii) at least two weeks between assessments for the control group. The FEES material was submitted for evaluation by the participants through the Google form platform; the order of the videos and frames was randomized for both the first and second assessments. The material was sent together with the scale and the anchor images. All data were treated in a pseudo-anonymized form; each participant was assigned an alphanumeric code. At the end of both assessments, participants completed a self-evaluation questionnaire created ad hoc to investigate participants’ perceived self-efficacy in interpreting FEES with the YPRSRS scale. The self-evaluation questionnaire consisted of 9 items, 2 of which were about anatomical identification of valleculae and pyriform sinuses and 7 items about scoring, based on a 5-point scale (1 = never, 2 = rarely, 3 = sometimes, 4 = often, 5 = always).
Statistical Analysis
For the analysis, the IBM SPSS v26.0® software for Windows (SPSS Inc, Chicago, Illinois) and the R software v.4.2.0 [22] were used. The clinicians’ and the students’ ratings were analyzed separately.
The baseline characteristics of clinicians were analyzed to make a comparison between the control and training. The Kolmogorov-Smirnov test was used to assess the normality of continuous variables and, as none of them were normally distributed, the Mann-Whitney U test was performed to compare group distributions; frequencies were compared through the chi-square test.
Construct validity and inter-rater reliability were used as a measure for the efficacy of the training. Conversely, intra-rater reliability was not considered a suitable outcome measure to assess the efficacy of the training due to the manipulation introduced by the training itself. Construct validity was defined as the agreement between each rater and the expert score employed as the “gold standard” [9, 10]. Inter-rater reliability was defined as the degree of agreement among raters scoring the same object on the same assessment [23].
Construct validity was calculated through weighted Cohen’s Kappa (quadratic weighting), separately for each rater. The distribution of raters Cohen’s Kappa was compared between the first and the second assessment using the paired t-test. Concerning control, training, and student groups, Cohen’s Kappa distribution was compared separately for the first, and the second assessment, using the one-way analysis of variance (ANOVA) with Tukey HSD adjustment to correct the significance level for post hoc pairwise comparisons.
As for inter-rater reliability, the level of agreement among each group of raters was calculated with Fleiss Kappa (quadratic weighting) both for the first and the second assessment; for each group, indices were subsequently compared using paired sample t-tests based on the linearization method for correlated agreement coefficients [24]. ANOVA with Tukey’s HSD method was also employed to check for differences in Fleiss Kappa values among the three groups separately for the first, and the second assessment.
The benchmark of Landis and Koch was used to evaluate the levels of agreement for the Kappa [25]: 0.00-0.20 slight agreement, 0.21–0.40 fair agreement, 0.41–0.60 moderate agreement, 0.61–0.80 substantial agreement, 0.81-1.00 almost perfect agreement. For the Fleiss Kappa, the following benchmark was adopted: < 0.40 poor, 0.40–0.75 intermediate to good, > 0.75 excellent [26].
To compare the results of the self-assessment questionnaire between the first and the second evaluation within each group, the non-parametric Wilcoxon signed rank test was used. Comparisons among groups for the first and second assessments were performed using Kruskal-Wallis test; pairwise comparisons were adjusted using Bonferroni correction.
Results
Raters’ Characteristics
Twenty-nine clinicians were initially recruited for the study. However, 4 clinicians, 2 from the training group and 2 from the control group, dropped out before completing the second assessment. Thus, data from 25 raters were ultimately analyzed. Table 1 shows descriptive statistics on clinicians’ characteristics, together with a comparison between training and control groups. No significant differences were observed between the two groups, except for the higher number of FEES the raters have participated in/performed among the control group (p = .02).
In additions to clinicians, 47 students (22.36 ± 3.54 years, 100% female) also participated in the present study as raters. Overall, 59.57% of the students had never observed a FEES, while 63.83% had performed a university internship with patients with dysphagia.
Training Results
Results
Effect of the Training on Construct Validity
Results of the comparison of construct validity values between the first and the second assessment within each group are reported in Table 2. Concerning FEES frames, an almost perfect agreement in the control, training, and students’ groups was observed; no statistically significant differences were found between the first and second evaluations for any of the three groups. As for FEES videos, construct validity values significantly improved between baseline (substantial agreement) and post-training (almost perfect agreement) in the training group for the pyriform sinus videos. In students, the construct validity associated significantly improved in vallecule and pyriform sinus videos: agreement improved from substantial to almost perfect in valleculae videos, and from moderate to almost perfect in the pyriform sinus videos (Table 2).
Pairwise comparisons of construct validity values among groups are reported in Table 3. At the baseline, the ANOVA post-hoc analysis showed that students exhibited significantly lower values of construct validity compared to the control group for the valleculae (both frames and video) and compared to both groups of clinicians for the pyriform sinus videos (Table 3). At the second assessment, students had lower values of construct validity compared to the control group only for the assessments of the valleculae frames (Table 4). No significant differences in construct validity scores were found between the training and control groups.
Effect of the Training on Inter-Rater Reliability
Table 4 shows inter-rater reliability results. No significant differences were observed between the first and the second assessments. Moreover, no significant differences in the inter-rater reliability values were found among groups at any time point (Table 5).
Self-Assessment
Self-assessment results on the perceived self-efficacy in interpreting FEES with the YPRSRS scale are reported in Tables 6 and 7. The Wilcoxon signed-rank test found no significant difference between the first and the second assessments among clinicians, regardless of their training status. Students’ group values significantly improved at the second assessment (Table 6), although the perceived self-efficacy remained lower than clinicians’ (Table 7).
Discussion
This study examined the training efficacy in improving the raters’ performance by assessing pharyngeal residue in FEES videos and frames using the YPRSRS and self-assessment changes from baseline and second assessment. The results obtained showed an improvement in agreement between participants, particularly students, and the experts in interpreting FEES videos. This improvement cannot solely be attributed to task repetition.
Our training spanned approximately 160 min, encompassing video lessons, independent practice, and a debriefing meeting. Unlike previous trainings for the use of the YPRSRS [9,10,11, 13], our current study introduced both practical exercises and a debriefing meeting to complement the theoretical lessons. Notably, the presented training lasted nearly 3 h, a substantial increase compared to the previous studies (8 min in the German study [10] and 4 min in the Italian study [13]). Other training programs on visuoperceptual scales typically have a similar or longer duration compared to the current one. In Kaneoka et al.‘s study [8], which aimed to demonstrate the reliability and validity of The Boston Residue and Clearance Scale (BRACS), four speech-language pathologists (SLPs) participated in a 3-hour session led by an expert clinician and co-creator of BRACS. The Visual Analysis of Swallowing Efficiency and Safety (VASES) training [7] comprised five parts covering VASES rules, practice with five FEES videos, a pre-recorded 60-minute session, additional practice with another five FEES videos, and a live 60-minute session. The median completion time for this training was 6 h.
Furthermore, previous studies that validated the YPRSRS in different languages [9,10,11, 13] analyzed the effectiveness of the training in frames but not in videos. Videos were included in this study because they better reflect the dysphagia assessment in the actual clinical practice. In addition, the video application is preferred in the instrumental evaluation of swallowing to ensure the quality of diagnosis and management of dysphagia [27]. The training presented in this study has advantages and disadvantages. The online mode allowed participants to take classes from home at the times most convenient for them. Online and face-to-face training are comparable in terms of effectiveness [17]. However, the asynchronous mode did not allow direct and timely exchange between instructors and participants. This disadvantage was partly offset by the debriefing.
Concerning clinicians, construct validity for pyriform sinus videos significantly improved after training, while no improvement emerged from the construct validity analysis of the control group in the second assessment. No group showed significant differences in inter-rater reliability values between the baseline and second assessments. Despite this, in the “training” group, it is possible to consider a trend of improvement in the valleculae videos (kappa values ranged from an intermediate to good agreement to an excellent agreement). The limited sample size, based on previously published studies [10], could have led to a lack of statistical power to detect significant differences. The training seems to have improved the rating precision of clinicians in assessing videos; however, similarly to previous studies [9,10,11, 13], such a result was not observed in raters’ performance during frames baseline evaluation.
Differently from previous studies, no difference was found between training and control groups for inter-rater reliability scores associated with frames assessment: Neubauer et al. [9] reported a significantly higher value for inter-rater reliability of trained raters in frames for both locations and Gerschke et al. [10] found significant differences between trained and untrained raters only for valleculae frames ratings. To date, the YPRSRS is widely employed, although it was newly devoloped or recently validated when the cited studies were conducted [9, 10]. Thus, it is possible to assume that participants were still not familiar with the scale in previous studies [10, 11, 13], while the higher level of clinical experience with the scale from participants in the present study could have influenced the efficacy of the training for frames.
In the students’ group, values for construct validity in FEES videos were significantly higher after training; the training reduced the gap between students’ and clinicians’ accuracy at the baseline as students achieved values of construct validity similar to clinicians. For the first time, students were selected to participate in training on the YPRSRS. The present results seem to confirm that the training, especially on less experienced raters, improves the accuracy of pharyngeal residues assessment in videos.
Self-questionnaire analyses of efficacy showed that students felt more confident using the scale after the training. This improvement in self-efficacy aligns with the progress noted in the second-assessment analysis, particularly in the pairwise comparisons with clinicians. No significant differences were observed between the baseline and second assessments in the clinicians’ groups. Notably, students, in general, exhibited lower confidence in applying the scale compared to clinicians in pairwise comparisons. An interesting finding emerges when comparing the training group to both the control and student groups during the second assessment regarding item 9. Item 9 involves the necessity of reviewing the video multiple times before assigning a score. The training likely enhanced clinicians’ thoroughness; consequently, they may have found it beneficial to watch the videos several times before finalizing their scores. Self-assessment questionnaires were not present in previous studies [9,10,11]. Moreover, none of the studies mentioned in a recent review on training to analyze functional parameters with FEES [14] included a measure specifically addressing self-confidence [18, 7,8,9,10, 28]. Investigating this aspect assumes relevance because, especially in inexperienced trainees such as students, good self-assessment skills are considered a key component in developing clinical skills [29].
There are some limitations in this study. All the students participated in the training. Thus, having a control group composed of untrained students was impossible. Future studies should also include a control group of students. Because of drop-outs, the number of raters who completed the second assessment was less than the expected sample size. Moreover, this study differs from previous studies on the YPRSRS as the “best-of-the-best” criterion was not used in choosing frames and videos. Although it may better reflect what clinicians experience in clinical practice, having included videos gathered from everyday clinical practice in a University Hospital may have made the evaluation more complex.
Conclusion
The results showed that the training can improve clinicians’ and students’ agreement with experts in assessing pharyngeal residues with the YPRSRS in the FEES videos. Promoting evidence-based training courses would allow students and clinicians with different backgrounds to have a standard method of interpreting FEES and sharing information in clinical practice and research.
Data Availability
The data that support the findings of this study are available from the corresponding author upon reasonable request.
References
Langmore SE, Schatz K, Olsen N. Fiberoptic endoscopic examination of swallowing safety: a new procedure. Dysphagia. 1988;2(4):216–9. https://doi.org/10.1007/BF02414429.
Schindler A, Baijens LWJ, Geneid A, Pizzorni N. Phoniatricians and otorhinolaryngologists approaching oropharyngeal dysphagia: an update on FEES. Eur Arch Otorhinolayngol. 2022;279(6):2727–42. https://doi.org/10.1007/s00405-021-07161-1.
Pisegna JM, Langmore SE. Parameters of Instrumental swallowing evaluations: describing a diagnostic dilemma. Dysphagia. 2016;31(3):462–72. https://doi.org/10.1007/s00455-016-9700-3.
Giraldo-Cadavid LF, Leal-Leaño LR, Leon-Basantes GA, Bastidas AR, Garcia R, Ovalle S, Abondano-Garavito JE. Accuracy of endoscopic and videofluoroscopic evaluations of swallowing for oropharyngeal dysphagia. Laryngoscope. 2017;127(9):2002–10. https://doi.org/10.1002/lary.26419.
Yoon JA, Kim SH, Jang MH, Kim SD, Shin YB. Correlations between Aspiration and pharyngeal Residue Scale scores for fiberoptic endoscopic evaluation and Videofluoroscopy. Yonsei Med J. 2019;60(12):1181–6. https://doi.org/10.3349/ymj.2019.60.12.1181.
Swan K, Cordier R, Brown T, Speyer R. Psychometric properties of visuoperceptual measures of videofluoroscopic and fibre-endoscopic evaluation of swallowing: a systematic review. Dysphagia. 2019;34:2–33. https://doi.org/10.1007/s00455-018-9918-3.
Curtis JA, Borders JC, Perry SE, Dakin AE, Seikaly ZN, Troche MS. Visual analysis of swallowing efficiency and safety (VASES): a standardized Approach to Rating Pharyngeal Residue, Penetration, and Aspiration during FEES. Dysphagia. 2022;37:417–35. https://doi.org/10.1007/s00455-021-10293-5.
Kaneoka AS, Langmore SE, Krisciunas GP, Field K, Scheel R, Mcnally E, Walsh MJ, O’Dea MB, Cabral H. The Boston residue and clearance scale: preliminary reliability and validity testing. Folia Phoniatr Logop. 2013;65:312–7. https://doi.org/10.1159/000365006.
Neubauer PD, Rademaker AW, Leder SB. The Yale Pharyngeal Residue Severity rating scale: an anatomically defined and image-based Tool. Dysphagia. 2015;n 30:521–8.
Gerschke M, Schottker-Koniger T, Forster A, Netzebandt JF, Beushausen UM. Validation of the German version of the Yale Pharyngeal Residue Severity rating scale. Dysphagia n. 2019;34:308–14.
Atar Y, Atar S, Ilgin C, Anarat MEA, Uygan U, Uyar Y. Validity and reliability of the Turkish translation of the Yale Pharyngeal Residue Severity rating scale. Dysphagia. 2022;37(3):655–63. https://doi.org/10.1007/s00455-021-10316-1.
Neubauer PD, Hersey DP, Leder SB. Pharyngeal Residue Severity Rating scales based on fiberoptic endoscopic evaluation 361 of swallowing. Syst Rev Dysphagia. 2016;31(3):352–9. https://doi.org/10.1007/s00455-015-9682-6.
Nordio S, Maistrello L, Favaretto N, Dellai A, Montino S, Agostinelli A, Ramacciotti G, Gheller F, Berta G, Koch I, Zanoletti E, Battel I, Cazzador D. Validity and reliability of the Italian translation of the Yale Pharyngeal Residue Severity rating scale. Acta Otorhinolaryngol Ital. 2023;43(1):12–9. https://doi.org/10.14639/0392-100X-N2289.
Costa BOI, Machado LS, Augusto MM, Magalhães DDD, Alves TC, Pernambuco L. Training to analyze functional parameters with fiberoptic endoscopic evaluation of swallowing: a scoping review. Dysphagia. 2023. https://doi.org/10.1007/s00455-023-10614-w.
Yoshida M, Kagaya H, Kamakura Y, Miura Y, Saitoh E, Okawa Y, Sanada H. Safety and the effectiveness of a new education program for nurses to assess swallowing function using fiberoptic endoscopic evaluation of swallowing (FEES). Japan J Nurs Science: JJNS. 2020;17(2):e12313. https://doi.org/10.1111/jjns.12313.
Benadom EM, Potter NL. The use of simulation in training graduate students to perform transnasal endoscopy. Dysphagia. 2011;26(4):352–60. https://doi.org/10.1007/s00455-010-9316-y.
Brady SL, Rao N, Gibbons PJ, Williams L, Hakel M, Pape T. Face-to-face versus online training for the interpretation of findings in the fiberoptic endoscopic exam of the swallow procedure. Adv Med Educ Pract. 2018;9:433–41. https://doi.org/10.2147/AMEP.S142947.
Warnecke T, Teismann I, Oelenberg S, Hamacher C, Ringelstein EB, Schäbitz WR, Dziewas R. Towards a basic endoscopic evaluation of swallowing in acute stroke - identification of salient findings by the inexperienced examiner. BMC Med Educ. 2009;9:13. https://doi.org/10.1186/1472-6920-9-13.
Pilz W, Vanbelle S, Kremer B, van Hooren MR, van Becelaere T, Roodenburg N, Baijens LW. Observers’ agreement on measurements in fiberoptic endoscopic evaluation of swallowing. Dysphagia. 2016;31(2):180–7. https://doi.org/10.1007/s00455-015-9673-7.
Rocca S, Pizzorni N, Valenza N, Negri L, Schindler A. Reliability and Construct Validity of the Yale Pharyngeal Residue Severity rating scale: performance on videos and effect of Bolus consistency. Diagnostics (Basel Switzerland). 2022;12(8):1897. https://doi.org/10.3390/diagnostics12081897.
Cichero JAY, Lam PTL, Chen J, Dantas RO, Duivestein J, Hanson B, Kayashita J, Pillay M, Riquelme LF, Steele CM, Vanderwegen J. Release of updated International Dysphagia Diet Standardisation Initiative Framework (IDDSI 2.0). J Texture Stud. 2020;51(1):195–6. https://doi.org/10.1111/jtxs.12481.
Team R, Core. (2013). R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, 373 Austria. http://www.R-project.org/.
Mokkink LB, Terwee CB, Patrick DL, Alonso J, Stratford PW, Knol DL, Bouter LM, de Vet HCW. The COSMIN checklist for assessing the methodological quality of studies on measurement properties of health status measurement instruments: an international Delphi study. Qual Life Res. 2010;19:539–49.
Gwet KL. Testing the difference of correlated agreement coefficients for statistical significance. Educ Psychol Meas. 2016;76:609–37.
Landis JR, Koch GG. (1977) «The measurement of observer agreement for categorical data» Biometrics, vol. 33, pp. 159–174.
Fleiss JL. Statistical methods for rates and proportions. Jhon Wiley. New York; 1981. pp. 38–46.
Hey C, Pluschinski P, Pajunk R, Almnahameed LG, Sader R, Stover T, Zaretsky Y. (2015) Penetration–aspiration: is their detection in FEES. Reliable without video recording? Dysphagia. 30(4):418–22. https://doi.org/10.1007/s00455-015-9616-3.
Borders JC, O’Dea MB, McNally E, Norberg E, Kitila M, Walsh M, Liu R, Pisegna JM. Inter- and intra-rater reliability of laryngeal sensation testing with the Touch Method during Flexible endoscopic evaluations of swallowing. Ann Otol Rhinol Laryngol. 2020;129(6):565–71. https://doi.org/10.1177/0003489419901145.
Pennbrant S, Nilsson MS, Öhlén J, Rudman A. Mastering the professional role as a newly graduated registered nurse. Nurse Educ Today. 2013;33(7):739–45. https://doi.org/10.1016/j.nedt.2012.11.021.
Funding
The authors did not receive support from any organization for the submitted work.
Open access funding provided by Università degli Studi di Milano within the CRUI-CARE Agreement.
Author information
Authors and Affiliations
Contributions
Conceptualization, Nicole Pizzorni and Antonio Schindler; methodology Nicole Pizzorni and Antonio Schindler; formal analysis Sara Rocca and Luca Negri; investigation Sara Rocca and Nadia Valenza; data curation Sara Rocca and Nicole Pizzorni; writing – original draft preparation Sara Rocca; Writing - review and editing Sara Rocca, Luca Negri, Nicole Pizzorni and Antonio Schindler; resources, Antonio Schindler; Supervision, Nicole Pizzorni and Antonio Schindler. All authors approved the final versione of the manuscript.
Corresponding author
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Rocca, S., Negri, L., Valenza, N. et al. Pharyngeal Residues Scoring through the Yale Pharyngeal Residue Severity Rating Scale (YPRSRS): Efficacy of Training. Dysphagia (2024). https://doi.org/10.1007/s00455-024-10725-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s00455-024-10725-y