The emergence of laparoscopic surgery and growing economic pressure in the medical field demand the establishment of new, more efficient educational training standards for residents to acquire surgical skills [13]. The first contact with laparoscopic surgery that students or residents have is the role of the “camera-man” in the operating room. Current training curricula make novices start using a camera in laparoscopic surgery mostly without any previous education or experience.

Virtual reality (VR) simulators are currently being evaluated for inclusion in surgical training curricula [46], mainly because skills learned on a simulator can be transferred to the operating room (OR) [711]. While most of this research focuses on the training of surgical operative skills in a simulation-based environment, only one study investigated the effect of simulator-based training on camera performance [12], showing a benefit of VR camera training versus no training in a porcine model. Adequate use of a 30° angled laparoscopic camera incorporates a whole set of skills such as anatomical orientation, horizon alignment, and scope orientation [13]. The absence of these skills may be associated with a decrease in general surgical performance, an increase in operating time [1, 14] and surgeon frustration. In addition, studies [12, 13] showing a transfer of manipulative skills learned on a simulator to the OR used a “nontraining” group as controls. It is obvious that a trained group performs better than a nontrained group. Therefore, the real value of simulator-based training and its transfer to the OR, especially in camera navigation, is not yet proven.

The present prospective randomized controlled study aimed to determine whether focused VR simulator-based laparoscopic camera training of novices could improve camera performance in an actual clinical situation in the same manner as does traditional training in the OR.

Material and methods

Thirty-two consecutive medical students on a surgical clerkship at the Department of Visceral and Transplantation Surgery of the University Hospital of Zurich, Switzerland, were enrolled in the study. Any previous active experience in laparoscopic camera handling in the OR and/or use of a VR simulator checked by a questionnaire was an exclusion criterion (n = 2 students). The study flowchart is shown in Fig. 1. The study was approved by the IRB of the University Hospital of Zurich and was registered on clinicaltrial.gov (identifier NCT01092013).

Fig. 1
figure 1

Study flowchart

All eligible participants (n = 30 students) completed the validated visuospatial “Stumpf-Fay cube perspective” test [15] and were given an identical 60-min introduction on the technical functionality and the correct handling of an angled laparoscopic camera to create an equal level of knowledge.

Pretraining test

Students were then sent to the OR to perform a baseline camera skills assessment test (pretraining test) involving a standardized set of navigational tasks they would have to accomplish at the beginning of an actual operation. All patients were placed in a supine position. Participants were positioned on the patient’s right side and were given the 30° angled laparoscope introduced in the trocar. They had to center and hold for 5 s the following positions/organs and had to maintain the correct horizontal alignment during camera movement: (1) left abdominal wall, (2) ascending colon, (3) right lobe of the liver, (4) sigmoid colon, (5) cecum, (6) pelvis, (7) trocar entry site in the upper left quadrant (simulated by a finger pressing externally), and (8) descending colon. Maximum duration of this test was set at 5 min. No additional trocars had to be placed and no manipulation of tissue occurred. No patient had had previous abdominal surgery. This assessment was videotaped. Performance of 14 camera tests done by laparoscopic experts (>100 laparoscopic operations each) were also taped to check for the ability of the test to distinguish between experts and novices.

Randomization and training

The students were randomized by sealed, opaque envelopes into two groups to obtain camera training for the next 3 weeks in either the skills lab (SL group) or the OR (OR group). Subjects in the SL group trained twice a week for 1 h for 3 weeks (total of six training hours) using two Xitact IHP™ instrument haptic ports as interfaces for the laparoscopic instruments and a third unidirectional electromechanical interface, the Xitact ITP™ instrument tracking port, for the camera navigation (Mentice AB, Gothenburg, Sweden) with the LAP Mentor™ (Simbionix USA, Cleveland, OH) software. As a second simulator, a haptic ProMIS™ surgical hybrid simulator (Haptica Ltd., Dublin, Ireland) was used. They had to follow a standardized protocol, performing 40 min of camera navigation-specific tasks on the different simulators (25 min on the basic task modules of the Lap Mentor camera manipulation and 15 min on the laparoscope orientation Core modules on the ProMIS) and 20 min of training on noncamera or camera-specific simulator exercises of free choice. After every two sessions, progress was monitored by an expert and corrections for improvement were made as necessary. Total time spent in the skills lab and actual time using the different individual exercise modules were documented by the participant immediately after each training session using standard forms.

Participants randomized to the OR group assisted at six laparoscopic surgeries (including hemicolectomy, rectum resection, gastric bypass, and cholecystectomy), navigating the camera at the surgeon’s direction. They were trained the traditional way by immediate hands-on practice in the OR. Total time spent in the OR and actual time navigating the camera were documented by the participant immediately after each operation using standard forms.

Post-training test

After both groups were trained for 3 weeks, the same “camera assessment” test in the OR (post-training test) and the validated visuospatial test were repeated.

Analysis of videotapes

All videotapes (of participants and of experts) were independently reviewed by five blinded observers, all experienced surgical attending physicians. For each of the eight steps (see above), performance of (a) horizon alignment and (b) centered, steady organ visualization was graded using a predefined rating scale (4 points: achieved goal >75% of time, 3 points: >50%, 2 points: >25%, or 1 point: ≤25%). The range of score was 8 (minimum) to 32 (maximum). Correct orientation of the angled scope (30° optic) for step 7 was also assessed (yes or no). Time to complete the test was noted, and time spent for lens cleaning or technical problems was subtracted.

Hypothesis

We hypothesized that both groups would show improvement in the same camera skills as assessed by the videotaped camera test, but that the OR group would be significantly less efficient regarding the proportion of actual camera navigating time over the overall time spent in the respective training facility (OR vs. skills lab).

Statistical analysis

A sample size of 12 participants in each group will have an 80% power to detect a difference in means of 100 min (assuming a mean of 300 ± 15 min camera training for the SL group and a mean of 400 ± 115 min for the OR group) using a two-group Satterthwaite t-test with a 0.050 two-sided significance level.

Statistical analysis was performed using standard software SPSS v16 for Windows (SPSS, Inc., Chicago, IL). To compare continuous variables between the two groups, the Mann–Whitney U test was used. Categorical variables were compared using the χ2 test or, when appropriate, Fisher’s exact test. A P value less than 0.05 was considered to indicate statistical significance. An interrater reliability analysis calculating the single-measure intraclass correlation was performed to determine consistency among raters of the videotapes.

Results

Thirty medical students fulfilled the inclusion criteria and were randomized. Six trainees dropped out during the study because they failed to acquire enough training time (n = 2) or to attend the two necessary camera assessment tests (n = 4). Therefore, 24 trainees completed the study.

Baseline demographics and pretraining test results

No difference in baseline characteristics of the 24 participants was noted (Table 1). The pretraining camera assessment tests did not reveal any group difference (Table 1). The 14 experts performed significantly better in the camera assessment test than the 24 participants, demonstrating that the test can distinguish between novices and experts (construct validity): organ visualization score (30.9 ± 1.2 vs. 23.7 ± 4.2, P < 0.001), horizon alignment score (29.2 ± 1.5 vs. 21.4 ± 4.1, P < 0.001), time to completion (69 ± 12 s vs. 171 ± 65 s, P < 0.001), and percentage of correct camera rotation (93% vs. 100%, P < 0.001). Interrater reliability, single-measure intraclass correlation among the five independent experts grading all camera tests were 0.68 for organ visualization and 0.66 for horizon alignment.

Table 1 Comparison of baseline characteristics and results of the pretraining camera test between the SL group and the OR group

Training was performed according to the study protocols and was efficient

The analysis of the training protocols showed that participants in the SL group adhered to the standardized 40-min protocol. They spent their remaining 20 min of free choice mainly performing manipulative task exercises equally on both simulators. Participants in both groups spent equal time actually training on camera navigation. However, participants in the OR group spent significantly more overall time in the OR to do so (Table 2).

Table 2 Comparison of the training time and results of the post-training camera test between the SL group and the OR group

Post-training test results

The comparison of the results of the post-training test between the SL group and the OR group is given in Table 2, the results of the pretraining test compared to the post-training test for each group is given in Table 3, and the difference in progress between the SL group and the OR group is listed in Table 4. Both groups showed significant progress in the organ visualization score and a significant decrease in mean time to complete the test. Improvements in horizon alignment and scope rotation handling were not significant. There was no significant intergroup progress difference for any of the parameters.

Table 3 Difference in the pretraining as compared to the post-training test results
Table 4 Difference in progress of the camera navigation test before and after training of the SL group compared to the OR group

Discussion

To our knowledge this is the first prospective randomized study indicating that simulator-based camera navigation training can be transferred to the OR using the traditional hands-on training in the OR as control rather than a “nontraining” control group as used in other simulator studies [1618]. Our study showed that traditional teaching in the OR still achieved its goal of improved and correct camera navigation but it was not as time efficient as VR simulator-based training. Camera handling generally represents the first step in today’s laparoscopic training for residents and is considered the basis for the acquisition of other laparoscopic skills. So far, current training standards make novices start using a camera in laparoscopic surgery mostly without any previous education or experience in the matter [19, 20]. As shown in this study, simulator-based camera navigation training is effective and highly time efficient and therefore should be integrated early into the training curriculum.

Surgical residency programs worldwide have been trying to integrate simulator-based training into their standard education [4, 5, 20]. However, the integration of such skill labs in standard surgical training curriculum is a demanding task [21]. Residents’ time restraints, the need for additional infrastructure including human resources (teachers), and the ambiguity about what such a curriculum should look like in detail [4, 22] are just a few of many important factors. It has been reported that unsupervised use of a surgical skills lab leads to uneager participation and the student fails to accumulate significant training time [5, 22]. Thus, training should be mandatory, with a dedicated schedule, and be at least partially supervised; this has been shown to increase the attendance rate substantially [22]. In this study we used a simple and short curriculum of 1 h twice a week for 3 weeks. The first 40 min of every hour included standardized exercises on simulators followed by 20 min of free exercises. After every two sessions, progress was monitored by a teacher and corrections were made as necessary. This combination of structured, controlled training with free time to discover the simulators (so-called “fun” part) has proven to highly motivate trainees as 10 of the 12 participants of the SL group returned to the simulator lab after the end of the study and 8 of the 12 in the OR group asked for further training on simulators. Trainees with little or no laparoscopic experience seem to profit most from simulator-based training [3], which indicates the importance of early incorporation into a training curriculum [6, 20]. This is supported by our findings of novices improving their camera navigation skills after completing a simple and short simulator-based training curriculum.

The main difficulty in learning laparoscopic camera handling and surgery is psychomotor and perceptual in nature [23]. Those spatial visualization abilities may vary significantly within the general population [24] and among residents [25]. There are more than ten different simulators available for camera navigational training functionality [26], while only a small number of devices have been properly validated [10]. In this study, we used two validated computer-based simulators with haptic and nonhaptic peripheral instrument ports. Although this study was not designed to determine if a single device or the combination of them was responsible for developing relevant camera navigational skills, participants’ feedback suggests that a combination of different training tools is highly appreciated and increased compliance as well as motivation.

Although in this study accuracy of organ visualization, a measure of orientation in the abdomen, improved significantly after training in both groups, correct horizon alignment was equal before and after training in the SL group, and improved, but not significantly, in the OR group. One could hypothesize that in simulated environments, participants can orientate on existing straight lines indicating the correct horizon level that do not exist in a real abdomen, and thus trainees were in an unfamiliar situation during the camera test. It has been stated that experienced laparoscopic surgeons might feel unfamiliar and possibly are at a disadvantage and learn less when training in a conceptual visual environment because of the absence of anatomical landmarks [13, 27]. As a consequence, one could postulate for more camera training modules using VR simulator software in an anatomical, not abstract, environment. However, training in the SL group was significantly more time efficient. Therefore, residents should be sent more often to simulator-based training, but not exclusively to simulations, as hands-on training is as effective and important. It is well known that an increase in operating time due to training in the OR is very expensive. Although our study was not designed to measure the effect of simulator training on overall operating time, students in the SL group achieved skills equivalent to those trained in the OR and were able to complete the camera test considerably faster, thus potentially reducing the operating time by handling the camera more efficiently. However, this needs to be proven in specially designed trials.

There are some limitations to this study. Participants in this study were students and not surgical residents and the results might not reflect the results of doctors with a particular interest in surgery. Recent studies found a correlation of visual-spatial abilities test scores and manual skills performed in surgery and on a VR simulator among novices [28], but failed to show such an association with experienced surgeons [2932]. However, most simulator studies choose students as trainees mainly because of their availability, their lack of any experience in surgery, and residents’ time [12, 33, 34] restraints. A second limitation was that the exact camera training time was not predefined in both groups and hands-on training in the OR was not standardized. However, “learning by doing” most likely represents the current practice of teaching in many hospitals all over the world. Standardized, controlled, simulator-based training will become the new gold standard because it is effective and more efficient, as shown in this study, and provides objective feedback.

In summary, this is the first prospective randomized controlled study indicating that simulator-based training of camera navigation can be transferred to the OR using the traditional hands-on training in the OR as control. Correct camera navigation is the first step in laparoscopic surgery and training using simulators should be integrated in all training curricula.