Intuitive Surgical’s da Vinci Surgical System (Sunnyvale, CA) was granted FDA approval in July 2000. Since then, numerous papers have been published focusing on patient outcomes data, cost comparisons to open and laparoscopic techniques, and emerging robotic technologies [13], but little literature has focused on the mechanical failures of the da Vinci system. Several series have described the rates of unrecoverable faults of the robotic arms or the console requiring conversion to open or conventional laparoscopy or outright case cancellation [48]. We have instead focused on the portion of the hardware that experiences some of the most intense mechanical stress: the surgical instruments. A survey of the literature for instrument failure descriptions and rates yielded single-center analyses [911] and reports based on data reported voluntarily by surgeons or institutes [12, 13]. In addition, two papers [14, 15] reported on a single instance of instrument failure.

Mues et al. [9], Kim et al. [10], and Nayyar and Gupta [11] reported all instances of either a single type of instrument failure [9] or all types of instrument failures [10, 11] experienced at their respective institutions over a period of time. The results of the three studies are summarized in Table 1. Andonian et al. [12] and Kaushik et al. [13] looked at failures that occurred at many different institutions.

Table 1 Summary of single-institution instrument failure rates

Rather than contacting surgeons and asking them to fill out a survey, Andonian et al. [12] reported on data available through the MAUDE database. The MAUDE (Manufacturer and User Facility Device Experience) database [16] is run by the United States Food and Drug Administration. Surgeons and institutions can voluntarily and anonymously report adverse events (defined as “potential and actual product use errors and product quality problems” [12]) and the product manufacturer can provide a response. In many cases, the response includes specific descriptions of the damage caused to the device by the failure as well as a suspected cause. Andonian et al. summarized the reports that were posted as of August 27, 2007, and found a total of 168 reports for the da Vinci Surgical System. Of the 168 reports, 43 (or 26 %) were reports of instrument failure. For each instrument failure, the type of instrument is listed but the failure mechanism is not reported.

Andonian et al. [12] acknowledge the limitations of their study in the discussion section of their paper. The limitations include the fact that reporting failures via the MAUDE database is voluntary and therefore it is likely that not all failures are reported. Some reports may have also been missed because they were filed with an incorrect brand or company name. Even knowing that their survey most likely did not capture all da Vinci system failures, Andonian et al. attempted to estimate an overall failure rate of the system. Their estimate is based on an estimate of the number of robotic urologic procedures that had been performed as of the time of their survey. Overall, they estimated a total system failure rate of 0.38 %. Using the same data, the instrument failure rate is 0.13 % (or 13 instrument failures out of every 10,000 procedures). This estimate is far lower than is reported elsewhere and almost certainly underreports the total number of failures.

Patients and clinicians deciding on the appropriate surgical approach, and medical institutions deciding on outcomes and cost, need to be informed of the realities of applying robotic technology to medicine. In an effort to provide this information, we have conducted a survey of the MAUDE database, and tallied up-to-date robotic instrument failures to derive an understanding of the baseline problem and suggest future improvements that could be made to robotic instrumentation.

Materials and methods

Using the MAUDE database, we performed a survey of failure events that occurred between January 2009 and December 2010, listed "Intuitive Surgical" as the manufacturer, and were reported as of January 25, 2011. We did not include non-instrument failures or instances where the failure was caused by an avoidable user error. Some of the reports appeared to be duplicates, which we excluded. If it was unclear, we erred on the side of overreporting. In comparison to the data presented by Andonian et al. [12], where only 43 instrument failures were reported in more than seven years, our survey found 528 reports of instrument failure in two years. The increased number of reports is most likely due to the larger number of robotic procedures being completed each year; it may also reflect a higher rate of voluntary reporting.

For each report, we attempted to identify the root failure whenever possible. If the instrument was returned to Intuitive Surgical, we assumed their analysis of the failure was correct. If the instrument was not returned to Intuitive Surgical, we assumed the site correctly reported the mode of failure.

Institutions may choose from several reporting options when filing a failure report. The easiest option requires the institution to download a free application from the FDA's website. The application guides the user through filling out the standardized reporting form and then submits the report electronically.

Results

We considered five main categories of failures (cautery, shaft, wrist or tool tip, cable, and control housing) and divided the reports into those categories. Of the 528 reports, a total of 565 failures were reported, which are summarized in Fig. 1. Ten of the reports had no discernible failure mode, 45 exhibited two independent failure modes, and four exhibited three independent failure modes.

Fig. 1
figure 1

Failure modes observed in da Vinci instruments over a 2-year period, as reported via the FDA MAUDE database [16]. Of 565 failure modes observed, 285 were failures of the wrist or tool tip, 174 were cautery failures, 76 were shaft failures, 29 were cable failures, and 7 were control housing failures

Wrist or tool-tip failures

The most commonly reported failures involved the instrument’s wrist or tool tip. Of the 285 reported wrist or tool-tip failures, some involved failures of more than one wrist or tool-tip component. It is unknown whether wrist and tool-tip failures occur more frequently than other types of failures or if they are simply more frequently reported because they are most easily noticed. The latter may be the case, since the surgeon is looking at or is near the tool tips for most of a standard operation.

Wrist and tool-tip failures are summarized in Table 2. The most commonly reported failure mode was a jaw or tool-tip cracking or breaking off of the instrument. It was also quite common for a report to state that “the instrument broke” or that "a piece broke off," without providing additional information. When the instrument was returned to Intuitive Surgical, more information was available in the report, summarizing the damage. Those instruments not returned to Intuitive Surgical are included in the wrist and tool-tip failure category as that is the type of damage seen in the majority of the clarified reports.

Table 2 Wrist or tool-tip failure modes observed in da Vinci instruments

Cautery instrument failures

The second most commonly reported failure mode was a cautery failure. Of the 528 filed reports, 410 involved either a monopolar or bipolar cautery instrument. Within that group, 174 incidents of electrical failure occurred. When a cautery instrument had a reported failure, there was therefore a 42 % chance of the failure being an electrical failure. This failure rate may not hold across all cautery instrument failures, but it does indicate that both electrical- and nonelectrical-related failures are possible when using cautery instruments. Of the 174 incidents of electrical failure, 156 (or 90 %) were arcing incidents (observed arcing, thermal tissue damage, or postoperative evidence of arcing). Cautery failures are summarized in Table 3.

Table 3 Cautery failure modes observed in da Vinci instruments

Table 3 lists a total of 184 cautery failures, which occurred in a total of 174 reports. In nine of the reports, damaged conductor wires were observed, and evidence of arcing was also present but not due to the damaged conductor wires. In a tenth report, the user noticed damage to an instrument preoperatively but then used the instrument in a procedure and observed arcing intraoperatively.

Instrument shaft failures

Shaft failures were less common than we would have expected, given their prevalence (9 of 19 failures) in the Kim et al. study [10], occurring in only 76 of the reports. Of the reported shaft failures, most involved material breaking off the shaft, or splintering or cracking of the shaft. Several of the instruments also broke circumferentially, either at the connection to the wrist (between the proximal clevis and the main tube interface) or slightly above it. Scratches were also observed on several tool shafts in conjunction with other failures. Shaft failures are summarized in Table 4.

Table 4 Shaft failure modes observed in da Vinci instruments

Cable and control housing failures

The last two categories of failures were cable failures and control housing failures. A total of 29 cable failures were reported; 28 were reports of the cable breaking or fraying at the tool-tip end of the instrument, and one was a report of the cable slipping off a pulley at the tool-tip end of the instrument. Seven incidents of the control housing failing were also reported. Two reports said that the instrument could not be removed from the robot arm. Two reports said that the housing broke. One report said that the instrument would not register, even though it had uses remaining. One report said that the cleaning nozzle had been pushed into the housing. The last report said that the instrument jaws would not open and close while attached to the robot arm but that they would if the dials were spun manually.

Both categories of failure occurred less commonly than expected based on previous literature. Nayyar and Gupta [11] reported 11 cable failures out of 23 total failures, with the other 12 being control housing failures. Kim et al. [10] reported 6 control housing failures out of 19 total failures.

Monopolar curved scissors instrument failures

Of the 150 reports of an instrument jaw or tool-tip breaking or cracking, 108 involved the monopolar curved scissors instrument. In most of those failures, one of the scissor blades broke off, fell into the patient, and was recovered. In addition, of the 174 cautery instruments that experienced an electrical failure, 125 were monopolar curved scissors. 116 of the reports were arcing incidents and another six were incidents where smoking, melting, or a burning smell was noted intraoperatively.

Discussion

The MAUDE database represents an objective source of reported instrument failures, yet it inherently underrepresents the true denominator of instrument errors because proper tracking requires timely and accurate reporting from the surgeon, the OR staff, and the hospital, each having different incentives and challenges to actually making the report. If a hospital chooses to report, it may choose to report all failures, no failures, or only a subset of failures. In addition, because the reporting is anonymous, there is no way to consider the failures on an institution-by-institution basis, a point brought up by Andonian et al. [12]. Even the instrument manufacturer may not know the true numerator if hospitals fail to report instrument failures to the company, and the true denominator (either the number of instruments on the market or the total number of instrument uses) is known only by the company itself. Our data do show, however, that many more failures exist in the database now than had been cited by Andonian et al. They observed only 43 instrument failures over more than seven years, whereas our survey found 528 reports of instrument failure in two years. The increased number of reports may be due to the increased adoption and utilization of robotics, but it may also reflect a higher rate of voluntary reporting.

An additional, less obvious, source of reporting bias may be that instruments returned to Intuitive Surgical had more detailed descriptions of the failure mechanism and were more likely to exhibit evidence of multiple types of failure. Some inconsistency is also added because different people at Intuitive Surgical wrote the damage reports and may have focused on different components of the instrument. For example, of the 28 total reports of scratching on the shaft, 13 occurred in January 2009 alone, and only 5 occurred in all of 2010. Although it is possible that this particular failure was truly less prevalent, another explanation could be that personnel changes or reporting practice changes occurred, yielding disparate data.

Failures must be observed by the user in order to be reported. In this analysis, we listed many types of failures, some of which are easier to miss than others. For instance, an instrument that fails to register when attached to the robot arm is more likely to be noticed than an instrument that has a cable starting to fray. Likewise, a surgeon is more likely to notice that one of the jaws on her scissors has fallen off than she is to notice that a small piece of the wrist covering or shaft covering has fallen off.

Even knowing that their survey most likely did not capture all da Vinci system failures, Andonian et al. [12] attempted to estimate an overall failure rate of the system. Their estimate is based on an estimate of the number of robotic urologic procedures that had been performed as of the time of their survey. Overall, they estimated a total system failure rate of 0.38 %. Using the same data, the instrument failure rate is 0.13 % (or 13 instrument failures out of every 10,000 procedures). This estimate is far lower than is reported elsewhere and almost certainly underreports the total number of failures. Kim et al. [10] retrospectively identified 19 instrument failures that occurred over 1,797 robotic surgeries performed between July 2005 and December 2008 at their institution. Nayyar and Gupta [11] prospectively identified 23 instrument failures out of 340 robotic procedures performed between July 2006 and March 2009 at their institution. These two reports yielded a range of instrument failures from 1.1 to 6.8 % [10, 11]. The variation in failure rates between institutions may be due to several factors, including the length of the study and the number of surgeons and procedures included, all of which modify the portion of the learning curve over which the study was conducted. The types of failures reported (e.g., Mues et al. [9] included only cautery failures, whereas Kim et al. [10] and Nayyar and Gupta [11] reported no cautery failures) will also produce variation. Whether data were captured prospectively or retrospectively could also have influenced the results of these reports.

Later studies conducted by Andonian et al. [12] and by Kaushik et al. [13] looked at failures that occurred at many different institutions. Kaushik et al. [13] sent a web-based survey to a group of urologists and asked them to report on past failures they remembered having experienced with the da Vinci system. Of the 260 reports of failures he received, only 21 (or 8 %) were reports of instrument malfunctions. The limitation of that study, however, was that it was based on recall which inherently biases the results. In contrast, Kim et al. [10] found that instrument malfunctions made up 44 % of the overall types of robot failures experienced at their institution, and Nayyar and Gupta [11] reported that 62 % of their institution’s failures were instrument malfunctions. Failure rates calculated from recall-based studies are likely biased toward significant failures that resulted in patient injury or that required conversions, and they can be expected to underreport less serious failures such as instrument failures that added only a short delay.

When looking specifically at individual classes of failures, it is unknown whether wrist and tool-tip failures occur more frequently than other types of failures or if they are simply reported more frequently because they are most easily noticed. The latter may be the case since the surgeon is looking at, or near, the tool tips for most of a standard operation. Although we cannot conclude a failure rate for monopolar curved scissors, the data suggest that this style of instrument usually fails in two specific ways: either a jaw breaks off or an arcing event occurs. Redesigning the instrument’s tool tip to strengthen the jaws around the cable crimp could help reduce the number of broken jaws. Modifying the design of the disposable covers would also improve the instrument’s reliability. The disposable covers are very difficult (and require a fairly large amount of force) to install correctly. The database descriptions for this type of instrument indicate that the disposable covers are also somewhat fragile and can be damaged or torn if the covers are installed incorrectly before use or if the instrument tips collide during a procedure.

The reported number of observed arcing incidents with the monopolar curved scissors may be considered slightly inflated as several hospitals responded to observed arcing by replacing the disposable tips on the pair of scissors and then continuing to use them for the procedure. In some cases, this resulted in two or three reports of the same type of failure from a single procedure, before the instrument itself was pulled or a new batch of tips was used. Even if half of the 78 reports of observed arcing were eliminated, however, the monopolar curved scissors still account for 64 % of all cautery instrument failures. The large number of failures for the monopolar curved scissors is less surprising when taking into account the fact that it is perhaps the most commonly used of all da Vinci instruments. The study performed by Mues et al. [9] reported on failures of the tip cover accessories for the monopolar curved scissors instrument experienced at a single institution over a period of seven months. Their prospective study identified 12 such failures out of 454 robotic procedures, for a failure rate of 2.6 %. The Mues et al. study identifies the tip cover accessory as a likely point of failure but unfortunately does not report any other instrument failures that occurred during the same time period.

Mues et al. [9], Kim et al. [10], and Nayyar and Gupta [11] identified failures that fit into each of the five failure categories. The distribution of failures over the five categories does not at all match the distribution we saw in our survey. Mues, Kim, and Nayyar together reported more than twice as many failures of the control housing as we found in our survey, despite our survey including ten times the total number of failures, whereas we found 285 wrist or tool-tip failures compared to their two. Although the failure categories are represented in the three papers [911], not all failure modes within each category are reported.

The variability in previous reports substantiates the need to create standard reporting practices from all institutions. Industry improvements need to be driven by objective metrics, and when describing surgical techniques to patients, accurate data are paramount to ensure that the patient is properly informed about his/her procedure’s risks and benefits. A review of methods to engineer improved instrumentation would be valuable but was outside the scope of this analysis.

Conclusions

The data show that the number of robotic instrument failures reported to MAUDE has dramatically increased in parallel with the rapid expansion of robotic surgery. In reality, many instrument failures may go unreported, thus a true failure rate cannot be determined from these data. However, education of hospital administrators, operating room staff, surgeons, and patients should be incorporated into discussions regarding the introduction and utilization of robotic technology. Because our data show such a significant increase in reported failures from the previous reports, we believe that it is imperative to know the true numerators and denominators. This can be achieved only by standardized reporting of instrument failure by surgeons and hospitals. Furthermore, our data show nonuniform reporting in the level of detail about the instrument failures. This increase in nonuniformity is concerning, and more concerning is that underreporting is certainly a confounder in the true failure numbers. We, as surgical robot users, will be unable to apply customer pressure and demands for improved instrumentation without standard reporting processes in place to know the full scope of the problem. We recommend that institutions incorporate standard failure reporting policies so that the community of surgical robot companies and surgeons can improve on existing technologies for optimal patient safety and outcomes. In addition, as with training to manage other operating room equipment failures, training to familiarize the OR staff with what to do in the event of a robotic instrument failure should be standardized.