Introduction

Although it presents many advantages for the patient compared to open surgery [15], minimally invasive surgery (MIS) is very complex and challenging for surgeons. Guidelines, good practices, and checklists are key tools to preserve the patient’s safety [10], while simultaneously reducing costs [16]. These tools are generally developed by a community of expert surgeons in a surgical specialty who reach a consensus following discussion and compromise [3, 4]. In addition, global rating scales (GRS) have been developed to formalize and assess some aspects of surgical practice [2, 9].

In recent years, with the widespread availability of artificial intelligence (AI) and machine learning, Surgical Data Science [8], which aims to extract knowledge from surgical data, has made tremendous progress. For instance, in relation to training and evaluation of laparoscopy, it allows the recognition of surgical tasks and can be used to assess technical skills [12]. In both cases, image content and information on the presence/motion of surgical instruments were of paramount importance in the recognition.

Due to numerous challenges summarized in [8], one of which is lack of explicability, SDS has not yet penetrated laparoscopic operating rooms. How can we trust an AI “black box” for decision-making if the clinical relevance of the information used by the algorithms to make their predictions cannot be explained to or discussed with the experts in the field? This question of explainable AI (XAI) is key in various domains. In SDS, even though some papers do address this XAI question [5], to our knowledge no work has been performed to present the selection criterion of an algorithm to the non-formalized expertise/knowledge of surgeons.

In a previous study [1], we developed a prediction model, which was used to investigate two specific indicators of surgical practice: how well the surgeon exposes the surgical target in the scene (ESS) and the profile of practice (SPP). Based on this work (briefly summarized in section “Previous work”), we now wished to address the explicability question by proposing a proof-of-concept framework (3.2) through which to interpret the model’s prediction results. Similarly to [14, 17], we used classical methods of feature extraction for the Support Vector Machine (SVM) technique. Our interpretations were submitted to 6 digestive surgeons from two hospitals, to assess whether they agreed with the model’s choices (section “Results of the survey submitted to experienced surgeons”). We end our article with a discussion of our results and present some perspectives for future studies (section “Discussion”).

Previous work

In [1], we predicted two facets of surgical practice during laparoscopic sleeve gastrectomy. The first was the exposure of the surgical scene (ESS), defined as the accessibility and visibility of the surgical target during the intervention. This notion characterizes the management of the tight surgical space, as well as the restricted view available in MIS. For the surgeon, it is crucial to maintain a clear surgical space as well as good visibility of the surgical target. The second facet was the surgeon’s profile of practice (SPP), which reflected the specific approach used by a particular surgeon, his/her habits of practice. This parameter was important as, even though surgeons follow guidelines and good practices, each individual has a personal way of operating that could be characterized. Overall, we assumed that ESS and SPP would describe different aspects of surgical practice and that video-based information was sufficient to classify them.

For our analysis, we focused on the Dissection of fundus step as it is critical in this surgical technique. Within this surgical step, we particularly analyzed the “dissection activities,” as their aim is to improve the ESS quality for subsequent surgical actions. The ESS and SPP ground truth were binary variables obtained based on annotation by an expert surgeon involved in the study (for ESS) and knowledge of the practicing surgeon (for SPP). To predict the ESS and the SPP, we created an annotated dataset from laparoscopic images of dissection gestures: each object visible in the images was segmented, and features were extracted from these segmentations. The dataset is described in greater detail in section “Dataset.”

The model took as input the features extracted for all the images in the dataset and was trained once to predict either the ESS or the SPP. Prediction consisted in a normalization followed by a Linear Discriminant Analysis (LDA) for the dimension reduction step followed by application of a Support Vector Machine (SVM) for the prediction per se. We optimized our model in a Train-and-Test Cross-Validation (TnT-CV) environment to avoid an over-fitting bias. Prediction performances were on par with other studies in the field, as summarized in Table 2 (lines 1 and 2).

Interpreting and questioning the prediction of surgical practice

Having produced a model capable of predicting ESS and SPP based on video information, our goals with this study were:

  • to identify what input information had been used by the model to make its predictions,

  • to assess the relevance of this information, based on feedback from experienced digestive surgeons.

We start by describing the video data used as input of our model (section “Dataset”); then, we present the methodology used to identify the most important input features for the prediction and to build the survey administered to the experienced surgeons (3.2).

Dataset

An expert surgeon annotated the ESS and SPP as binary variables for each image of a dissection gesture (referred to as a “dissection sample”). The spatial state of each of these 10 objects visible in the video (see Fig. 1) was described by 13 features (see Table 1). The 130 resulting variables could then be analyzed by applying “feature-object” semantics.

Fig. 1
figure 1

Illustration of the 6 organs and 4 surgical instruments segmented in the surgical videos

Tableau 1 Spatial features extracted for each object segmented in the image

Our dataset was composed of an input matrix X of size \(m \times n\), and two potential output vectors \(y_{{\textit{ESS}}}\) characterizing the ESS, and \(y_{{\textit{SPP}}}\) characterizing the SPP, both of size \(m \times 1\). We had \(m=735\) samples, and given the 10 objects and 13 features, \(n=130\) variables. These data were used as such in the prediction model presented in section “Previous work.” We constituted 6 clusters allowing us to more closely investigate the links between the input data and output vectors. Figure 2 illustrates the construction and inclusion relationships used to establish these clusters.

Fig. 2
figure 2

Construction and inclusion relationship of clusters—the \(C_{\textit{ESS}}\) cluster with its two subclusters \(C^A_{\textit{ESS}}\) and \(C^B_{\textit{ESS}}\)—the \(C_{\textit{SPP}}\) cluster and its three subclusters \(C^A_{\textit{SPP}}\), \(C^B_{\textit{SPP}}\) and R (the remaining samples)

Three clusters aimed to predict the ESS:

  • Cluster \(C_{\textit{ESS}}\) associates the input data X and the output vector \(y_{\textit{ESS}}\). It answers the question “What characterizes the quality of exposure for both surgeons overall?”

  • Cluster \(C^A_{\textit{ESS}}\) associates \(X^A_{\textit{ESS}}\), the subset of input data X restricted to surgeries performed by surgeon 0, with \(y^A_{\textit{ESS}}\), the corresponding output subset of \(y_{\textit{ESS}}\). It answers the question “What characterizes the quality of exposure as surgeon 0 operates?”

  • Similarly, cluster \(C^B_{\textit{ESS}}\) associates \(X^B_{\textit{ESS}}\), the subset of X restricted to surgeries performed by surgeon 1, with \(y^B_{\textit{ESS}}\), the corresponding output subset of \(y_{\textit{ESS}}\). It answers the question “What characterizes the quality of exposure as surgeon 1 operates?”

Three clusters aimed to predict the SPP:

  • \(C_{\textit{SPP}}\) associates the input data X and the output vector \(y_{\textit{SPP}}\). It answers the question “What characterizes the surgeon’s profile of practice independently of the quality of exposure?”

  • \(C^A_{\textit{SPP}}\) associates \(X^A_{\textit{SPP}}\)—the subset of input data X restricted to the 3 surgeries with the highest mean \(y_{\textit{ESS}}\) score - with \(y^A_{\textit{SPP}}\), the corresponding output subset of \(y_{\textit{SPP}}\). It answers the question “What characterizes the surgeon’s profile of practice when the quality of exposure is good?”

  • \(C^B_{\textit{SPP}}\) associates \(X^B_{\textit{SPP}}\)—the subset of input data X restricted to the 3 surgeries with lowest mean \(y_{\textit{ESS}}\) score—with \(y^B_{\textit{SPP}}\), the corresponding output subset of \(y_{\textit{SPP}}\). It answers the question “What characterizes the surgeon’s profile of practice when the quality of exposure is non-optimal?”

Validation methodology

In order to construct the survey, we studied the input variables that had the most impact on the prediction results, and translated our findings in the form of statements. These statements were submitted to experienced surgeons alongside illustrative video segments. Our validation methodology consisted in 4 steps described in the 4 following sub-sections (see Fig. 3).

Fig. 3
figure 3

Methodological steps in model interpretation and assessment

Estimating the importance of input variables

Similarly to [7], we were interested in observing the impact of the input variables on performance in the prediction task: for each variable we applied a value permutation and observed its impact on the resulting prediction. The value permutation consisted in shuffling values over each input variable iteratively, to observe the induced prediction shift [13].

We implemented the permutation process within the TnT-CV environment in our prediction model to avoid introducing an overfitting bias. For each CV split configuration, we first trained the model on a train-fold and evaluated the baseline prediction performance of the model \(p_{ref,cv}\). Then, value permutation was performed for each input variable v, the prediction performance \(p_{v,cv}\) was determined, and the distance to \(p_{ref,cv}\) was computed. This distance corresponds to the amplitude of the shift in performance induced by the value permutation: variables that have a smaller distance from the reference are more important to the prediction than those further from the reference.

We defined the importance \(i_v\) of input variable v as this distance averaged over the \(n_{cv}\) split configurations of the TnT-CV environment:

$$\begin{aligned} i_v = \frac{1}{n_{cv}}\sum _{cv=1}^{n_{cv}}{p_{ref,cv}-p_{v,cv}} \end{aligned}$$
(1)

The vector of importance \(I_V\) is composed of the importance \(i_v\) of each variable (\(I_V\) is thus of size \(1 \times n\)). We also computed the mean importance \(i_{v,obj}\) of each object over its 13 features.

Experimental protocol We performed two experiments on the 6 clusters (see section “Dataset”):

  • Experiment 1: using the permutation process, we computed the mean importance of each object \(i_{v,obj}\). For each of the 6 clusters, we extracted the object with the highest mean importance \(i_{v,obj}\).

  • Experiment 2: having identified the object with the highest mean importance for the 6 clusters, we restricted our input variables to the 13 features of this object and computed corresponding \(I_V\) using the permutation process. Thus, for each of the 6 clusters, we extracted the feature with the highest importance \(i_v\).

In this way, we extracted the object and its corresponding most important feature predicting each cluster’s output.

Translating the importance results into statements

Based on the analysis of the importance results, a scientist with extensive knowledge of the clinical context translated each importance result into a meaningful sentence, or statement. Each importance result referred to a cluster and a set of most important variables. Considering these constraints, the scientist translated each importance result by answering the question related to the cluster with the most important variables of the importance result using the “feature-object” semantic (see section “Dataset”).

Each statement satisfied the following constraints:

  • Each importance result has a unique associated statement.

  • A statement refers to one cluster.

  • A statement refers to a set of input variables characterized by one or more features related to a single object.

  • A statement refers either to the ESS output, or to the SPP output.

Illustrating statements with samples and associated video clips

We wished to illustrate each statement with two video clips, one for each output value. In these two situations, we automatically pre-selected \(m_{{\textit{clip}}}\) samples and their related video clips. Then, among the \(m_{{\textit{clip}}}\) clips, we manually selected the one best illustrating the situation, where the output was either 0 or 1.

A translated statement S refers to an importance result, its cluster and its most important variables. These constraints define some \((X_S; y_S)\) data for statement S, with a reduced number of input variables: 13 variables for Experiment 1 where the importance result related to the most important object and all its features, and 1 variable for Experiment 2 where we focused on the most important feature of the most important object. The \(y_S\) vector was either a subset of \(y_{\textit{ESS}}\) or \(y_{\textit{SPP}}\) (restricted to the samples of \(X_S\)) depending on the cluster studied.

In the pre-selection, we determined which samples were the most decisive for our model regarding a statement S by training our predictive model on the \((X_S, y_S)\) data. After this training, the decision function in the SVM step of our model returned a decision value for each trained sample (i.e., a prediction probability). The \(n_{{\textit{clip}}}\) highest decision values were automatically selected for the “1” and the “0” labels of \(y_S\), allowing us to identify the most discriminant samples for statement S.

Once we had identified the \(m_{{\textit{clip}}}\) most decisive samples, it was straightforward to obtain their associated timestamp from the video and extract representative video clips. A clip lasts a given duration \(t_{{\textit{clip}}}\) and its beginning is shifted from the sample’s timestamp by \(t_{{\textit{shift}}}\). Finally, an operator considered each statement and its video clips. For each label of the output variable \(y_S\), he selected one sample among the \(m_{{\textit{clip}}}\) which best illustrated the clinical idea expressed by the statement S.

Experimental protocol We characterized the selection process by setting the parameters \(m_{{\textit{clip}}}=5\), \(t_{{\textit{clip}}}=10\,\mathrm{s}\) and \(t_{{\textit{shift}}}=-7\,\mathrm{s}\). Based on these parameters, for each statement, 5 sample video clips were pre-selected for each output label. Each sample was illustrated with a 10 s video clip that started 7 s before the sample timestamp. The operator who selected one sample out of the 5 proposed was a non-clinician specialist. In total, two video clips were presented with each statement.

Tableau 2 Prediction performances of the model for the 6 clusters—in each cluster, the performance of the binary classification was evaluated for each image sample

Design of the survey

A survey was designed with questions consisting in a statement associated with two video clips. Surgeons responded using a 5-level Likert scale of agreement: 0 = strongly disagree, 1 = rather disagree, 2 = no opinion, 3 = rather agree, and 4 = strongly agree.

Experimental protocol The survey was produced with questions in a randomized order and submitted to 6 experienced digestive surgeons (1 PGY5 resident, and 5 experts).

Results

Extraction of the most important variables

The prediction performance of our model informs us on the level of confidence we can accord to the estimation of the importance vector \(I_V\). Table 2 presents the prediction performance for the complete clusters \(C_{\textit{ESS}}\) and \(C_{\textit{SPP}}\) (lines 1 and 2) taken from [1], and for each sub-cluster defined in section “Dataset.” The number of input samples in each cluster is also provided.

Table 3 provides a summary of the results of Experiment 1, highlighting the most important objects identified for each cluster studied. For instance, the most important object when predicting the ESS in cluster \(C_{\textit{ESS}}\) was identified as the spleen (id #1).

Table 4 provides a summary of the results from Experiment 2, highlighting the most important feature of the objects identified as most important in Experiment 1. For instance, id #3 indicates that the most important feature of the spleen to predict the ESS was the texture. Result id #12 was not translated into a statement as this feature had exactly the same importance as the 11 other features. More detailed results are presented in suppl. material #1.

Results of the survey submitted to experienced surgeons

Statements (see Table 5) and their related video clips were extracted for each of the most important objects and features to build a survey comprising 11 questions. Figure 4 illustrates two questions from the questionnaire, and the full questionnaire is provided as a Power Point file in the Supplementary Material #2 with the embedded videos.

Tableau 3 Results of Experiment 1 with extraction of the most important objects for the prediction tasks performed with each of the 6 clusters (computed as \(\overline{i_{v,obj}}\), the highest mean importance over features)
Tableau 4 Results of Experiment 2 with extraction of the most important feature for the most important object for the prediction task applied to each of the 6 clusters (computed as \(i_v\), the highest importance)
Tableau 5 Clinical statements based on the extraction results that were submitted to surgeons in the survey
Fig. 4
figure 4

Two examples of statements with their two associated video clips, as submitted in the survey

Fig. 5
figure 5

Overall agreement with the statements for each surgeon

We gathered responses from 6 surgeons expressed as a level of agreement on a Likert scale and present them here in summary form. The complete responses are provided in suppl. material #3.

Figure 5 presents the overall level of agreement of each surgeon with the statements. We observe that, apart from surgeon #1 who expressed a low overall level of agreement, the five other surgeons tended to agree with the statements in our survey.

Figure 6 shows a more detailed breakdown of the responses to each statement with a Likert chart. 7/11 answers show a positive agreement trend, while 4/11 answers show a trend for disagreement. For each statement, at most one neutral opinion was recorded.

Fig. 6
figure 6

Likert chart representing the level of agreement of surgeons, centered on the neutral answer, and ordered by agreement level

As surgeons answered the survey and immediately after administering it, we collected their views on the clinical implications and the design of the survey. Here are some of their open comments:

  1. 1.

    Three of the 6 surgeons observed at least once during the survey that the video clips did not correctly illustrate their related statement. These observations were made for 5 different statements, not for a single one.

  2. 2.

    Three surgeons discussed the duration of the video clips, two out of three considered that they were too short.

  3. 3.

    In response to statement #3, four surgeons considered the texture of the spleen to be a tactile notion, whereas it represented a visual notion for one surgeon.

Discussion

In this study, we wished to investigate the explicability of a predictive model based on laparoscopic video data and to compare its interpretation with surgeons’ expertise. AI applications are generally evaluated based on a set of metrics quantifying their performance when completing a specific task, but this type of assessment might miss significant clinical concerns. By presenting our model to clinicians, we were able to strengthen its clinical relevance.

A similar XAI method for the prediction of hypoxemia during surgeries based on temporal hemodynamic data, ventilation data and patient history data was proposed in [7]. In their study, four anesthetists were asked to rank the same variables as their model, and the two rankings were compared. Thanks to an established literature in the field of anesthesia, a consensus already exists on the major risk factors to monitor, and the results of the comparison mainly helped to validate the predictive model.

A majority of studies relating to surgical practice work with practice indicators. Only a few recent studies have exploited laparoscopic videos with the aim of better understanding clinical practice: [6] proposed to automatically detect intra-operative adverse events based on the annotation of surgical activities and hidden semi-Markov models, but the image content was not exploited. [11] used convolutional neural networks (CNNs) to automatically assess the critical view of safety in laparoscopic cholecystectomy, but the use of a CNN makes it challenging to interpret the network’s predictions.

Here, we studied two more specific indicators of surgical practice: the quality of handling of exposure of the surgical target by the surgeon (ESS) and the profile of surgical practice (SPP), and proposed a method to link the choices made by a machine learning model to clinical expertise. Due to the lack of literature on these two notions, the comparative results provided by our survey raise questions on the interpretations of our model, as well as the implicit, and non-formalized knowledge of the surgeon relative to these phenomena. In the future, we hope that XAI will become a tool to validate AI models but also to formalize complex and often implicit knowledge, with potential applications in surgical training and surgical quality control.

Clusters and prediction

Our dataset \((X, y_{\textit{ESS}}, y_{\textit{SPP}})\) led us to build 6 clusters, each of which was described by a meaningful question. The “feature-object” semantics describing the input variables were used to define Experiments 1 and 2. Overall, this rich conceptual context allowed us to propose a substantial and meaningful analysis of this dataset with 12 distinct scenarios (6 clusters \(\times \) 2 Experiments).

The prediction results for clusters \(C_{\textit{ESS}}\) (\({\textit{Acc}}=0.64\)) and \(C_{\textit{SPP}}\) (\({\textit{Acc}}=0.6\)) described in Table 2 are on par with the literature. Results for clusters \(C^A_{\textit{ESS}}\) and \(C^B_{\textit{SPP}}\) have similar mean accuracies but higher standard deviations, which might be imputed to the smaller numbers of samples.

The prediction performance of our model on clusters \(C^A_{\textit{SPP}}\) and \(C^B_{\textit{SPP}}\) may be questionable. These two clusters are very small (150 samples as against the 735 samples in the complete dataset), and almost perfect prediction performances were obtained. This type of result is typical of overfitting, and the interpretations presented for these two clusters should be considered with care.

We assumed that our model’s predictive performances were acceptable and sufficient to build our method (importance vector computation, automatic selection of video clips). However, these performances are not optimal and may affect the resulting variable importance vectors and pre-selected video clips.

The variable of importance \(I_V\)

To select the most important objects and variables, we chose to look at the mean prediction differences with value permutation (Eq. 1). Using this approach, variable importance is easily interpretable as it represents how much impact a variable has on the predictive task along the CV splits.

However, examination of the detailed importance results (suppl. material #1) reveals relatively similar \(i_v\) values, and selecting a set of most important objects/features rather than a single one may give more realistic results. To complement our method, we could also evaluate the stability of the variables importance by including the standard deviation of the prediction shift.

Translating the importance results into clinical statements

Interpreting the variable importance \(i_v\) and translating it into a clinically meaningful statement is the key step in our methodology. We translated video-based variables to texts referring to the laparoscopic practice by placing ourselves at the interface between computer science and clinical practice. To our knowledge, this is a first attempt in the domain.

We chose to perform this translation manually and we avoided the human bias of the operator by defining sufficient constraints on the translation process:

  1. 1.

    The definition of the clusters, their related questions and the “feature-object” semantics provided a framework in which the translation gained clarity.

  2. 2.

    The operator had to answer the cluster’s question with an importance result: the most important object (Experiment 2) or its most important feature (Experiment 2). As discussed in section “The variable of importance \(I_V\),” other object/feature selection approaches could have been considered, but might have made the interpretation/translation process more complex.

  3. 3.

    Constraints were defined directly for the translations in section “Translating the importance results into statements.”

Despite these constraints, we faced challenges related to the “feature-object” semantics which required different translation approaches: barycenter_x and barycenter_y have intricate meanings and would be difficult to understand separately when characterizing the position of the object. Surgeons gave us contradictory feedback on the texture feature: two surgeons observed that it is a tactile sensation, whereas another considered it to be visually interpretable. Similarly, we processed instruments and organs as equivalents despite their clear clinical distinctions. Thus, depending on the context (computer science or medical), some concepts take on a very different meaning and should be considered with extreme caution.

To reduce the human bias, several experts should collaborate to translate statements. To avoid the human bias, completely automated translation could be used, but this is an open, complex question. Automated translation implies a Natural Language Processing approach with input data composed of labels (variable name) and values (importance value) to generate text content at the interface between computing and laparoscopy expertise.

Illustrating statements with samples and associated video clips

To complement the statements, we selected video clips from the initial video dataset. Surgeons appreciated this mixture of text and image as it encouraged them to investigate very specific aspects of their practice.

We performed an automated pre-selection of the most illustrative samples to remain as objective as possible. For each statement, we reduced the variable count (13 variables for Experiment 1 and 1 variable for Experiment 2). Despite the high importance of these variables, our model was designed and optimized for a matrix with \(n=130\) variables, and not with 13 or less. Thus, the automated selection of images requires external validation. Future studies will include a focus on the impact of the variable count on the model’s performance.

To validate the automated pre-selection, we added a manual selection step which introduced a human bias. Choosing the best clip among those automatically pre-selected was non-trivial, and when the survey was presented to surgeons, three of them observed at least once that the video clips did not correctly illustrate their associated statement. Statement #10 focused on the stomach, but one of its clips failed to show the stomach correctly. The two clips for statement #6 showed different, incomparable, surgical steps. This gap between the statement and its illustration shows the complexity of selecting a satisfactory clip.

We defined \(t_{{\textit{clip}}} = 10\,\mathrm{s}\) as the clip duration with regard to the mean period between two successive annotations of ESS (\(19.1\,\mathrm{s} \pm 21.9\,\mathrm{s}\)). One surgeon confirmed that this duration suited him/her, but two others found that the clips were too short to correctly analyze the surgical practice. We will need to investigate the clip duration in future studies.

The survey submitted to experienced surgeons

We proposed a very novel clinically-oriented validation methodology with the presentation of a survey to 6 surgeons: 4 expert surgeons and 1 PGY5 resident came from one hospital, the last expert worked at another hospital. This proof-of-concept study was limited to assessing the relevance of our XAI approach and the distribution of the levels of agreement of the surgeons surveyed reveals their interest in our approach. We aim to include more surgeons from different hospitals in a larger-scale study in the future.

Out of the 11 statements presented, surgeons indicated positive agreement for 7. This positive agreement trend throughout our survey indicates we have validated our proof-of-concept study. Through this survey, 5/6 surgeons agreed with the interpretations of our predictive model and confirmed the overall clinical relevance of 7/11 statements. Only the combined responses for 1 of the 6 surgeons revealed an overall disagreement with our survey (7/11 disagreements). In section “Clusters and prediction” we already mentioned the potentially overfitted prediction result due to the small number of samples in clusters \(C^A_{\textit{SPP}}\) and \(C^B_{\textit{SPP}}\). Among the 4 statements with which surgeons disagreed, 3 were related to these two clusters. This result may confirm the weakness of the model’s prediction for these two clusters.

In section “Illustrating statements with samples and associated video clips,” we discussed the lack of complete concordance between some statements and their video clips. When faced with this situation, surgeons were tempted to assess the ability of the video clips to correctly illustrate their statement rather than assessing the relevance of the statement itself. We asked them to keep the initial goal of the survey in mind, but we should have anticipated this ambiguity. In the future, we will work with sociologists and/or linguists to build a systematic survey methodology correctly answering questions related to the presence/absence or the appropriate duration of the clips.

Conclusion

In this paper, we propose a proof-of-concept method to validate the clinical relevance of a model predicting two indicators of the quality of surgical practice. This validation was based on a survey administered to 6 surgeons that took the form of statements illustrated by short videos. To our knowledge, no previous study proposed similar clinical interpretations, or validated such an approach with a survey presented to surgeons. Overall, we developed a novel approach with a high potential for the analysis and understanding of surgical practice.

If we could entirely automate our approach, it would allow us to automatically provide the surgeon with a video associated with relevant subtitles. Such an algorithmic approach would make it possible to produce some clinical applications for surgical training, post-training review, or postoperative feedback.