Introduction

Improvements in efficiency and workflow in today’s operating room (OR) will significantly enhance surgery in the OR of the future [1]. Harnessing technology and redesigning work processes help deliver higher quality at lower cost [2]. Both junior and senior personnel constantly face new challenges as new technologies such as robot-assisted surgery [3] and adaptive workflow systems are introduced [4].

This trend toward increasingly integrated technological support systems and the rise of streamlined processes in the OR have led to the demand for better-trained personnel. Consequently, the traditional apprenticeship model in healthcare no longer meets the needs of learners in perioperative nursing education [5]. Moreover, given the growing shortage of nurses in the workforce, perioperative leaders are facing a significant challenge in recruiting new staff who are committed to perioperative nursing [6]. The undersupply of nursing staff is exacerbated by a rise in average age [7]. Furthermore, at a time when there is a critical need for nurses, nursing education is beset by a severe shortage in nursing faculty [8]. New strategies to ensure structured, forward-looking training for novices and to support the lifelong learning of experienced personnel are therefore essential to cope with the predictable need for extended scrub nurse education for tomorrow’s OR. Stahl et al. stated that the introduction of new technology into the OR environment can create significant stress despite achieving desirable goals, such as reducing patient flow time, necessitating risk groups among personnel to receive additional support to manage technology transitions [9]. Simulation clearly offers enormous potential for safely developing expertise in procedural skills [10]. The ability to enable personnel to practice and learn in a safe, controlled environment makes simulation a valuable tool for initial training and continued competency verification [11]. Although the need to integrate information technology into undergraduate nursing programs has been recognized, there are few reports of teaching and learning with these new technologies [12].

Currently, interactive training and simulation systems as preparation for the OR almost exclusively target the surgeon’s role. Commonly applied to a wide variety of surgical tasks [13] such as robotic laparoscopy [14, 15], they are rarely used for OR functions not performed by the surgeon. Sturm et al. analyzed a variety of simulation-based surgical training approaches with skill transfer to the operative setting [16]. Palter and Grantcharov [17] reviewed the simulators available to surgical educators in 2010 and discussed the potential uses of simulation technology in surgery for the assessment of technical and non-technical skills. Although there are several simulation approaches in nursing education, they are rarely used for perioperative nursing and are practically non-existent for instrument nursing in particular. Most existing approaches focus on the simultaneous training of all parties involved in an intervention. Aggarwal et al. [18] presented a setup for a simulated operating theater project to train and assess the technical and non-technical skills of the entire surgical team. Fort and Fitzgerald [19] described skill learning in a simulated OR with perioperative evidence-based practice modules, such as handling surgical attire (e.g., gloves and hospital gowns) and scrubbing. Sexton et al. [20] conducted team-oriented medical simulations (TOMS) in a full-scale OR in Basel (Switzerland) with complete OR teams, including the scrub nurse.

These simulation scenarios mostly use human-like manikins [21] and rely on human observation to control the progress of the scenario and its evaluation. This requires a great deal of organization and manpower, prepared premises, and expensive equipment. For example, a top-of-the-line, high-fidelity human patient simulator is a large investment [22] and cannot be used for complex surgical interventions. These simulations are definitely justified, especially for surgical crisis management [23] and the development of non-technical skills. However, they are too intricate for the exclusive training of the skills of scrub nurses, especially for beginners lacking technical skills such as knowledge of surgical instrument names and the intervention progress. We had already gained extensive experience with the working environment around the instrument table during our previous work on intra-operative surgical instrument usage detection [24] and scrub nurse viewing [25], which led to the fundamental concept for the creation of a training system. The goal was to focus on the technical skills needed by beginners and personnel changing clinical departments while also supporting the range of scrub practitioners’ non-technical skill requirements as systematically identified by Mitchell et al. [26, 27].

This article explains the implementation of this concept by describing the architecture and features of a newly developed system called the Nosco Trainer. We evaluate the suitability of this novel system for professional education and the development of technical and non-technical skills.

Methods

System design: structural system setup

The Nosco Trainer combines multiple computer systems and monitors to simulate the scrub nurse workplace. Figure 1 provides an overview of the setup and the sensor equipment involved. The heart of the setup comprises a Microsoft PixelSense (formerly called Microsoft Surface) interactive surface-computing platform, henceforth called the instrument table system. This device is based on a Samsung SUR40, which combines a horizontal 40” screen with an integrated computer system, giving it a table-like appearance. Its proportions offer sufficient space for the life-size simulation of an entire typical surgical instrument table. The device can recognize 52 simultaneous multitouch contact points, allowing concurrent interaction on the table with multiple fingers and hands.

Fig. 1
figure 1

Overview of the system components

The instrument table system has a local area network (LAN) connection to another computer called the operating room simulation system. In the setup used for the study, this system was connected to two monitors. The surgeon screen displayed the scene including the surgeon from the scrub nurse’s point of view. Its touchscreen enabled simulated interaction with the surgeon. The second monitor, a conventional screen, displayed the corresponding view of the endoscope. Figure 2 shows both monitors and the complete system in the demonstration OR at ICCAS.

Fig. 2
figure 2

Demonstration of the trainer system in the demonstration OR at ICCAS. Here, the endoscope screen has been replaced by a connection to the OR monitor matrix

Fig. 3
figure 3

Logical model view of the nosco.sub XML schema for the annotation of a surgical intervention

System design: dynamic system behavior

The underlying software framework is based on a client–server architecture divided into two subsystems. The instrument table system runs on the Nosco Table Client Software written in C#. The choice of programing language was driven by the need to access the hardware features of the PixelSense system, which rely on the Microsoft Surface 2.0 software development kit (SDK) [28]. The client was implemented to be as lean as possible. It handles display and interaction with the instruments and the additional instrument tray. The server can remotely control every aspect displayed on the client. Moreover, it is constantly notified of every change happening on the client, including instrument movement and rotation. The side tray was implemented as a scrollable list displaying slightly downscaled pictures of the instruments on top of the typical underlay mat of a surgical tray. Scrolling is performed by placing a finger on the tray list and dragging it up or down. Instruments can be added and removed by selecting an instrument and pressing one of the arrow buttons next to the tray on the table showing the corresponding direction. Figure 5 shows a screenshot of an instrument table with tray. At any time during the simulation, smooth and simultaneous interaction with the content of the table is possible. Instruments can be moved on the table and also rotated with two fingers. As an additional feature, selecting an instrument causes a small square button to appear for several seconds on the top left corner of its translucent gray selection area. If this button is pressed, a label bearing the name of the instrument briefly appears next to the instrument.

The OR simulation system runs on the Nosco Trainer Server Software written in Java 1.7. Its main function is to handle every aspect of the simulation and all the connected subsystems. The simulation is based on the synchronous playback of different videos and sounds on the surgeon screen and the endoscope screen. A language called nosco.sub was developed in Extensible Markup Language (XML) to annotate the videos with information relevant to the simulation, as shown in Fig. 3. An upcoming instrument request by the surgeon in the simulation is represented by an element of InstrumentRequestType. The From and To elements represent the beginning and end points in milliseconds for the display of the request for the surgical instrument on the surgeon screen.

Fig. 4
figure 4

Excerpt of underlying information video metadata in nosco.sub format

During this time, a large button appears under the video on the surgeon screen showing the name of the requested instrument. At the end of the time period, all videos freeze until the participant has passed the correct instrument. The participant can pass an instrument to the virtual surgeon by holding it on the instrument table screen with one hand and simultaneously pressing the button on the surgeon screen. If the wrong instrument is passed or if no or multiple instruments have been selected, the participant is notified by a message window appearing on the surgeon screen. If the correct instrument is passed, the instrument disappears from the instrument table screen and the simulation and all videos continue. The definition of the InstrumentReturnType is analogous to InstrumentRequestType. By pressing the button shown to accept the instrument that the surgeon wants to pass back, it returns to the instrument table screen and the simulation continues. ActivityType bundles other possible activities simulated in the trainer. Finally, a custom interview component was integrated for streamlined trainee interviews.

Fig. 5
figure 5

Left annotated instrument table state at the beginning of the simulation. Right list of instruments visible on the table, sorted alphabetically and without the instruments contained in the additional tray on the left

Evaluation study

Design of the evaluation study

A multi-center study to evaluate the Nosco Trainer System was conducted under laboratory conditions. Participants were recruited for the study by ENT physicians, who also supervised the medical aspects of the simulation. This study was embedded in a more extensive study which also evaluated aspects of the current level of standardization in the operating room, such as the similarity of surgical instrument descriptions used by scrub nurses [29]. To include a diverse target audience and ensure a significant group of participants, the study was conducted at the ENT departments of three different hospitals: Acqua Klinik in Leipzig, Germany, Inselspital University Hospital of Bern in Switzerland, and Leipzig University Hospital in Germany. The study was conducted in three blocks, one for each hospital, between fall 2014 and spring 2015. Each clinic provided a secluded room where the demonstrator was set up throughout the study. Five participants from each clinic were interviewed, totaling fifteen study participants. A surgical intervention known as functional endoscopic sinus surgery (FESS) was selected as the basis for the study. As described by Stammberger and Posawetz [30], the general aim of FESS is to clear diseased ethmoid clefts and compartments under the guidance of a rigid endoscope and to reestablish ventilation and drainage of the diseased larger sinus via their physiological routes. The intervention included the use of a surgical navigation system. FESS was chosen because of its limited range of instruments (which fit on a single instrument table) and because it is commonly practiced in ENT clinics (to ensure all participants’ familiarity with the type of intervention). To give the participants an impression of the simulation environment, a 3-min video recording of the intervention was selected for the simulation. The excerpt was taken shortly after surgical incision when the processus uncinatus is removed, indicating the current stage of the intervention to the participants. The video was annotated with the information to be simulated during the demonstration. Figure 4 shows an excerpt from the associated file created in nosco.sub format.

In total, fourteen instrument requests and responses were contained in the demonstration as well as six activities included solely for information purposes (e.g., cleaning instruments and refilling fluids). These were intended to give the participants an outlook on possible future extensions of the simulation. Figure 5 shows an annotated screenshot of the instrument table at the beginning of the simulation. All necessary instruments were already laid out on the table so that no interaction with the tray was required during the simulation. The participants had already gained hands-on experience of the functionality of the table before the start of the simulation by performing a table setup from scratch, which was not included in this study. The selection of instruments provided on the table was reduced and was not intended to represent any ideal or preferred table state. It included optional spare instruments for potential extension to any intervention, as is the case in actual surgery.

To evaluate the system, a questionnaire was devised to interview the participants before and after performing the demo simulation. All questions were reviewed with experts in the field, both from a medical angle and in terms of user experience. All interviews started with a general set of questions to determine the composition of the participants including factors such as age, language skills, work experience, and experience of training personnel. Subsequent to the demo simulation, the participants were asked twenty-two questions to determine their detailed ratings of various technical and non-technical aspects of the demonstrator system and to ascertain the most important things that beginners could learn with the trainer. For all rating questions, an identical rating scale was chosen ranging from 1 to 5 (1—very good, 2—good, 3—neither good nor bad, 4—poor, 5—bad). All participants took part in the interview voluntarily, were assured anonymity, and were told that they were free not to respond to questions if they wished. They were informed that their performance and speed while working with the system would not be measured, and that the purpose of the demonstration was solely to give them a detailed insight before reviewing and commenting on the demonstration system. For all rating questions, the average rating per clinic was calculated, along with the overall average rating. All questions were translated from German into English, as were all the participants’ answers.

Results of the evaluation study

Two-thirds of the 15 participants were female; one-third was male. Three of the participants were aged between 18 and 30 years, six between 31 and 40, five between 41 and 50, and one between 51 and 60. Thirteen of the 15 participants were native German speakers, and all participants were fluent in German. All participants were smartphone users. All participants had undergone traditional nursing education with continuing education as an OR specialist with the exception of three participants in the Swiss hospital, who had received equivalent training called qualified operating room technician Footnote 1 or technical operation assistant.Footnote 2 The 15 participants had on average 15.7 \(\pm \) 9.4 years experience as nurses, including 12.8 \(\pm \) 9.8 years as scrub nurses. Eight of the 15 participants had been trained in the same OR in which they were working at the time of the interview. Seven participants had never trained anyone inside the OR before, two participants had trained fewer than 10 persons, and six participants had trained more than 10 persons during their career. All participants had performed surgical instrumentation for the type of intervention presented in the demonstration of the simulator and were familiar with the type of intervention. Table 1 presents the results for the questions for which participants were asked to provide detailed ratings of the basic features of the simulation system (Q1–Q9).

Table 1 Closed-ended questions regarding the basic features of the simulator system

Table 2 presents the results for the block of questions in which participants were asked about their detailed ratings of the effectiveness of the simulator system for training various skills (Q10–Q21).

Table 2 Closed-ended questions regarding the effectiveness of the simulator system

Table 3 presents the answers to the question about the most important things that beginners should learn with the trainer (Q22).

Table 3 Categorized results for the open-ended question “In your opinion, what is the most important thing that beginners could learn with the Nosco Trainer? (multiple answers permitted)” (Q22), the results show the number of mentions and the percentage rounded to one decimal place

Discussion

We created a system to simulate a surgical intervention in an OR from the point of view of a scrub nurse by developing client–server software and a connected set of touchscreen monitors and an interactive table. A study was conducted to evaluate both the demonstrator system and its related design decisions.

The system was very positively received with an average of 1.8 rating points for the realism acquired (Q1). All participants said they could imagine training a full-length operation with the system (Ø 1.7 for Q2), as long as the intervention was not monotonous. They also agreed that the system was properly suited to prepare them for a new type of operation (Ø 1.7 for Q10) as well as for education and training (Ø 1.7 for Q11). These results are especially promising considering the participants’ long mean experience (12.8 years) as scrub nurses. Because all participants were familiar with the type of intervention, any distraction caused by the OR situation presented was minimized. The design decision to use a Microsoft Surface 2.0 to simulate the instrument table proved valuable. The instruments were clearly distinguishable despite lacking a third dimension (Ø 1.9 for Q3). Moreover, most participants considered the reaction speed for the movement of instruments on the table to be sufficient (Ø 1.9 for Q4). The design decision to simulate the additional tray on the side by using a scrollable list displaying slightly downscaled pictures of the instruments on top of the typical underlay mat of a surgical tray was considered good (Ø 2.0 for Q5) but prompted a great deal of discussion. In general, it was felt that the organization of the tray as a list could be improved, especially because the organized and always identically prepared surgical trays of real ORs minimize the time spent searching for additional instruments. Therefore, the question of whether the system trains the ability to change the instrument table configuration during the operation (Ø 2.0 for Q17) received a similar rating and feedback. The same applies to the question of whether the system trains the ability to keep the instrument table in order (Ø 2.0 for Q14). Most points of criticism could be resolved by organizing the list differently, e.g., with groups of instrument categories, or by showing the additional tray as a whole, either displayed on the same table after pushing a button or by moving it to a second interactive table. Also, the automatic return of instruments to their original position on the table when returned back from the surgeon needs to be reconsidered.

Showing the scrub nurse’s line of sight to the surgeon was well received (Ø 1.5 for Q6), as was the implemented mechanism of passing and receiving instruments by pushing a button on the surgeon screen (Ø 1.7 for Q7). When developing the system, we compared different possible passing techniques from a man–machine interaction perspective. We chose a two-handed approach (the first on the table touching the chosen instrument and the second pushing the button on the surgeon screen) because it minimized unwanted mistakes such as selecting multiple instruments while still giving the feeling of “handing over something”. However, the simulator does not yet cover the skill of correctly passing instruments, which is considered important by the experts (Q21) and a central part of traditional hands-on training. This could be addressed in a future update by replacing the single “pass instrument” button with a selection of photos showing hands holding the instrument, from which the trainee has to choose the right passing technique for the situation.

The audio playback including the surgeon’s voice was considered primarily good (Ø 1.7 for Q8), although the sound quality was often rightfully criticized. Replacing the surgeon with a (possibly virtual) actor with recorded announcements for instrument requests could enhance comprehensibility. This approach could also solve an existing central weakness of the system compared to traditional hands-on training: the fact that unclear verbal statements cannot yet be queried by the trainee. Even so, the participants could easily follow the course of the intervention (Ø 1.9 for Q9), although various details led to distractions. For instance, some considered the arrangement of the screens too widely spread or unusual because it did not resemble the familiar environment in their own OR. Also, some participants took longer than others to become acquainted with the user interface. Nevertheless, the question about whether the system could improve the ability to keep track of the different fields of information during an intervention was rated highly (Ø 2.0 for Q13), as was the system’s ability to train people to follow the operation (Ø 2.0 for Q15). The fact that the course of the intervention and the use of instruments also depend heavily on the individual surgeon was mentioned often and reflected in the question about whether the system trains the ability to recognize the next instrument required (Ø 2.5 for Q16). However, this may not be a serious issue because the majority of participants felt the trainer was suitable for training the designated use of instruments (Ø 1.9 for Q21).

The question about whether the system is able to train the ability to keep calm during real interventions was often answered negatively (Ø 2.5 for Q18). The associated question about whether the system has positive implications for the scrub nurse’s performance during lengthy operations especially revealed the limitations of a simulator (Ø 2.9 for Q19). It became apparent that a simulation would never be able to cover all aspects of the daily routine of a scrub nurse (Ø 2.4 for Q12). Because the evaluation was limited to a small part of a FESS intervention and a group of scrub nurses specialized in ENT surgery, different surgical domains could also require many additional features. It should be borne in mind, though, that the target group consisted of novices who first need to learn basic techniques. With a rating of 1.5 for Q20, the system’s ability to teach the names and shapes of instruments was affirmed almost unanimously. Familiarization with instruments’ names was the most-cited aspect that beginners ought to learn with the trainer, closely followed by the instrument table setup (Q22).

Conclusions

Interactive training through simulation is a new approach to prepare novices for the challenges at the instrument table in the OR. It can also improve the lifelong learning of perioperative personnel and provide insights into the fast-growing sector of new technologies and operation methods.

The proposed system is currently unique in its kind, and, for the first time, applies concepts for surgeon training to operating room personnel. Judging by the thoroughly positive feedback and encouragement from the study participants, the new demonstrator was well liked by both surgeons and the OR personnel interviewed. Even criticized parts of the system were met with several proposals for improvement. The results are especially noteworthy because of the very experienced participant group from three ENT departments with a mean value of 12.8 years experience as scrub nurses and with over 50 % of the participants having experience in training personnel. Compared to traditional training, the demonstrator enables lifelike or even critical situations to be simulated without the need for real patients and contributes significantly to patient safety. It can be used to train both technical and certain non-technical skills. Furthermore, the approach circumvents the need for human supervision in every training phase of the learner. Instead, it can adapt to the user’s individual self-learning pace. Furthermore, the high cost of real surgical instruments is eliminated. With the hardware costing roughly €10,000, the system is affordable for training centers and professional schools. The simulation approach presented is not dependent on a specific type of surgical intervention or medical discipline. With an expandable simulation language and an open communication interface, the system is prepared for the future integration of supplementary simulation aspects such as other activities, more monitors, or additional hardware.

A follow-up project will integrate the most promising aspects into an expanded system and then evaluate it with groups of novice scrub nurses in professional schools.