Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.
Advertisement
Scientific Reports volume 12, Article number: 14682 (2022)
Metrics details
An innovative algorithm to automatically assess blood perfusion quality of the intestinal sector in laparoscopic colorectal surgery is proposed. Traditionally, the uniformity of the brightness in indocyanine green-based fluorescence consists only in a qualitative, empirical evaluation, which heavily relies on the surgeon’s subjective assessment. As such, this leads to assessments that are strongly experience-dependent. To overcome this limitation, the proposed algorithm assesses the level and uniformity of indocyanine green used during laparoscopic surgery. The algorithm adopts a Feed Forward Neural Network receiving as input a feature vector based on the histogram of the green band of the input image. It is used to (i) acquire information related to perfusion during laparoscopic colorectal surgery, and (ii) support the surgeon in assessing objectively the outcome of the procedure. In particular, the algorithm provides an output that classifies the perfusion as adequate or inadequate. The algorithm was validated on videos captured during surgical procedures carried out at the University Hospital Federico II in Naples, Italy. The obtained results show a classification accuracy equal to (99.9%), with a repeatability of (1.9%). Finally, the real-time operation of the proposed algorithm was tested by analyzing the video streaming captured directly from an endoscope available in the OR.
Indocyanine green (ICG) is a molecule developed in the 1950s at Kodak’s R &D laboratories1, applied in the field of infrared photography. This molecule is the first substance discovered capable of emitting fluorescence in the near infrared (NIR) spectrum, as it becomes fluorescent when illuminated with infrared light. This substance has negligible toxicity and is quickly disposed of by the body without side effects, except for rare allergic reactions easy to prevent2. In 1959, the Food and Drug Administration (FDA) approved its use in clinical settings3, and since then, it has been widely used for diagnostic investigations for pathology affecting heart, eyes, liver, and lungs. This substance is injected into the patient’s vein before surgery or near the tumor mass to be removed the day before surgery. The molecule binds to plasma proteins present in the blood, giving its fluorescent properties to the blood, liver, and biliary circulation4.
More recently, ICG has been largely used in the surgical field thanks to the introduction of fluorescence detectors, namely optical systems for excitation and detection of the emitted fluorescence. A relevant topic of research is the adoption of ICG to estimate the perfusion quality in laparoscopic surgery5,6,7,8,9,10. This is essential to assess whether the intestine is adequately perfused; this, in fact, serves as an indication of the outcome of the procedure11,12,13. In fact, a perfusion deficiency at the point where an anastomosis is performed increases the risk of anastomotic dehiscence, which consists in a failure to heal the sutures with the consequent appearance of fistulas and tissue perfusion14. Therefore, assessing the quality of the perfusion by means of ICG allows the surgeon to promptly intervene while the surgical procedure is ongoing. The most used technique to verify the perfusion of an intestinal segment is to inject ICG into the patient body. This element makes the blood fluorescent with a green tinge if lightened with infrared light. The evaluation of the intensity and the uniformity of this fluorescence allows to assert if the parts are adequately perfused. This technique was successfully used by Boni14 to provide information related to perfusion during colorectal surgery, and assist the surgeon in adopting the best strategy in the phases of colorectal anastomosis (stitching of two conical stumps), often necessary in colorectal interventions. Moreover, other applications of ICG relate to the dynamic discrimination of primary colorectal cancer using systemic indocyanine green with NIR endoscopy15, intraoperative ureter identification, and lymph node dissection16.
Currently, the fluorescence brightness of ICG is evaluated only qualitatively and subjectively by the surgeon, basing on experience. Indeed, at the state of the art, there are no systems or techniques used to quantify it, and to objectively support the surgeons in their assessment. At the state of the art, several attempts to design systems capable to help the surgeons in the assessment of perfusion quality have been made17,18,19,20. These approaches are mostly based on the diffusion speed of the indocyanine in the tissues. The perfusion of the colorectal segment is estimated by looking at the gradient of the intensity of ICG fluorescence brightness captured by the camera. The output provided by these systems is a heat map highlighting the intestinal portions characterized by a faster increase of ICG fluorescence brightness after the injection. Furthermore, they try to correlate the heat map with the post-surgery result. However, the methods described in these works are not able to automatically assess if the perfusion is good or not in the analyzed area, as they only provide a graphical output that has to be interpreted by the surgeons subjectively.
To overcome this issue, the branch of technology that is becoming increasingly popular is Artificial Intelligence (AI)21,22, and in particular Machine Learning (ML). The capillary diffusion of powerful calculators, along with the effort of researchers to develop effective algorithms23,24,25 have contributed to the widespread adoption of this technology in a large variety of application contexts, such as healthcare26,27. At the state of art, there are many examples of ML-based approaches used in the medical field, as proof that this technology can help and assist the operators to minimize the risks for the patients and prevent complications. For example, a decision support system based on AI was used by Cahill et al.28 in colorectal cancer intra-operative tissue classification. Instead, Park et al.29 adopted AI to evaluate the feasibility of AI-based real-time analysis of microperfusion to predict the risk of anastomotic complication in the patient with laparoscopic colorectal cancer surgery. According to Igaki et al.30, a first study to use an image-guided navigation system with total mesorectal excision was conducted. Moreover, Sanchez et al.31 provided a systematic literature review regarding the use of AI to find colorectal polyps in colonoscopy. Finally, Kitaguchi et al.32 used AI to identify laparoscopic surgical videos, in order to facilitate the automation of time-consuming manual processes, such as video analysis, indexing, and video-based skill assessment. Nevertheless, there are still no methods based on AI that automatically assess the quality of perfusion in the analyzed area.
Starting from these considerations, in this paper, a ML-based system to objectively assess if an intestinal sector is adequately perfused after an injection of ICG is proposed. This system is used to (i) acquire information related to perfusion during laparoscopic colorectal surgery, and (ii) objectively support the surgeon in assessing the outcome of the procedure. In particular, the algorithm provides an output that classifies the perfusion as adequate or inadequate.
From an implementation point of view, the system works on a video extracted from a laparoscopic camera and a Region of Interest (ROI). The ROI is selected by a member of the operating room team (e.g., an assistant surgeon) and contains the area to be assessed. Then, it adopts a set of pre-processing steps to build the input of a Feed Forward Neural Network used to evaluate the quality of perfusion. The precise tuning of the neural network hyper-parameters allows the proposed architecture to have a prediction accuracy high enough to anticipate the possible adoption of the system as a standard routine to be applied during surgery. As a proof-of-concept demonstration, the case study, based on perfusion analysis applied to abdominal laparoscopic surgery at University Hospital Federico II in Naples, Italy, is reported. The feasibility of such approach in real time is proven with optimal performance. Thus, the system can represent an effective decision support for both less-experienced surgeons and those at the beginning of the learning curve.
The problem addressed in this work can be formally expressed as a two-class classification problem involving frames (from a video streaming) corresponding to an adequately – or inadequately – perfused area. To this purpose, the idea was to develop a system that automatically assesses the quantity of ICG present in the ROI, by computing the histogram of the green band of the acquired frames, then providing an output corresponding to an adequate or to an inadequate perfusion.
The study was conducted according to the guidelines of the Declaration of Helsinki. Because the study does not include a pharmacological experimentation, using medical devices or patient data, but only the computer analysis of video material (collected during routine clinical practice), approval by the Ethics Committee is not necessary. Each patient signed an informed consent for the surgical procedure and approved the use of their data by third parties.
The overall architecture of the proposed system is shown in Fig. 1.
Block architecture of the proposed algorithm. Three main blocks are outlined: (i) a fast tracking algorithm to track the selected ROI, (ii) a feature extraction block to pre-process the available frames and (iii) a ML-based classifier to provide the output in terms of quality of perfusion.
The input of the system are (i) the frames coming from the Video streaming, and (ii) the ROI identified as a rectangular box selected by the user. The ROI, which identifies the portion of the frame to be analyzed, is selected by the OR operator using the mouse or the track-pad on the computer when starting the algorithm.
The architecture, from left to right, is composed of the following three functional blocks:
The first block consists of a Fast tracking algorithm, which is used to track the selected ROI during the video execution. In particular, the Minimum Output Sum of Squared Error (MOSSE) tracker33 was exploited, as it uses adaptive correlation to track objects, resulting in a better robustness to variations in lightning, pose, scale and non rigid transformations. The MOSSE implements also an auto pause and resume functionality if the object to track disappear (for example, if the surgeon covers it) and then it reappears again. Moreover, the exploited tracker can work at high frame rates (more than 450 fps).
Once the frames containing the ROI are extracted from the video source, the second block performs a Features extraction. The frames are in the RGB format, namely the colored image is obtained by a combination of three images, one for each color channel: red, green and blue. Each pixel has an 8-bit resolution; this value represents the intensity of the pixel. Afterwards, the ROI of each frame is divided into 20 vertical equal slices and, for each of them, the histogram of the green band and its area is computed according to Eq. 1:
where (A_i) is the generic element of the features vector corresponding to the slice i; b(l) is the bin value at the level l; (count_i(l)) is the number of occurrences of the green intensity at the level l for the slice i; and k is a parameter used to exclude pixels with low values of green. In this work, (k=25) was chosen as it guaranteed the best classification performance. Finally, a vector of 20 elements (features vector) is obtained. This vector becomes the input to the last functional block.
This step of the process binarily Classifies the feature vector and establishes whether it corresponds to an adequate/1 or to a inadequate/0 perfusion of the colorectal portion. This classifier is obtained by a Feed Forward Neural Network. The binary cross-entropy was selected as a loss function, and the optimizer Adam34,35 was chosen.
The following Neural Networks (NN) were evaluated as classifiers:
One-hidden-layer NN: In this case, a classic feed forward neural network (FFNN) with one hidden layer was used. The output layer had a single neuron with a sigmoidal activation function. The hidden layer was preliminarly tested with (i) 20 neurons and a Rectifier Linear Unit (ReLU) activation function, and (ii) 80 neurons and a Tanh activation function. After, this network was further tested with the following activation functions: Tanh, Sigmoid, and Rectifier Linear Unit (ReLU): here, for each activation function the number of neurons changed between 10 and 100, with step 10.
Two-hidden-layer NN: In this case, a FFNN composed by 2 hidden layers was used. Different combinations of ReLU, Sigmoid and Tanh activation function, considering a number of neurons equal to 50, 70, and 90, were tested. SoftMax activation function, and two neurons were used for the output layer.
Therefore, the tuned hyper-parameters were (i) activation functions, and (ii) number of neurons for each hidden layer. Moreover, Support Vector Machine (SVM)36 method was used with linear and Gaussian kernel as a baseline classifier. For each hyper-parameter configuration, all the aforementioned ML models were validated on the entire data set using the K-fold Cross Validation (CV)37 with K=10 folds. K-fold CV is a standard approach to assess and select a ML model in a statistically significant manner and without overfitting38. The data set is divided in K folds and the network is trained K times for each combination of hyper-parameters. Each time the network is trained, one of the K folds of the data set is used as test set and all the remaining K-1 as training set. The selection of the best model was conducted according to mean of the obtained accuracies over the K test folds, defined as the percentage of correct classification. After the selection, the chosen model was trained again on the entire data set in order to extract as much information as possible from the data39, to use it in real-time in an actual surgery scenario.
The proposed algorithm was developed in Python 2.7 on Windows 10. The open-source framework and libraries used are TensorFlow, Keras, and OpenCV. The training of the proposed NNs was conducted by setting a number of epochs equal to 100 and the batch size equal to 5.
The study was conducted according to the guidelines of the Declaration of Helsinki Approval of the institutional review committee was not required because the data of the present study were collected during routine clinical practice. Each patient signed an informed consent for the surgical procedure and approved the use of their data by third parties.
In this section, first, the laboratory experimental validation is described: during this phase, a data set provided by the surgeons was used to train and validate the ML classifiers adopted by the proposed algorithm. Then, an online validation in OR was carried out, by employing the best ML model obtained after the training.
Intraoperative use of ICG technology. Fluorescence angiography shows the vascular perfusion of the intestinal segment that delimits the section point (a). Anastomosis is performed with the residual colon (b).
Setup
A total of 11 videos in .M4V format were provided by the surgeons: the videos were collected and labelled by the medical staff during routine clinical practice. An anonymisation procedure was applied to protect patients privacy. In particular, any metadata was removed from the original files. These files contain the video, acquired directly from the endoscope during surgery, related to the portions of the intestine where the anastomosis was being performed. When the ICG was injected, the portion that was well perfused became fluorescent. An example of frames extracted from the dataset is shown in Fig. 2, which shows the intraoperative use of ICG technology.
In particular, Fig. 2a refers to the fluorescence angiography which shows the vascular perfusion of the intestinal segment that delimits the section point. Fig. 2b shows the anastomosis being performed with the residual colon. Fluorescence angiography was performed using a laparoscopic system (Olympus OTV-S300, Olympus Europe SE & Co. KG, Hamburg, Germany) with a light source (Olympus CLV-S200 IR) which allowed the use of both visible and near-infrared light.
Results
The performance of the developed NNs was validated by processing the 11 videos available from the dataset. From each video of the data set, different frames were extracted. To properly train the model in assessing the quality of perfusion, frames containing ROIs with clear evidence of ICG were selected as well as more tricky ones. From each ROI, 20 features vectors were obtained. The total size of the dataset is constituted by 470 frames. Figure 3 illustrates the overall process of features extraction. The considered frame is shown in Step 1. As aforementioned, the ROI selected by the user is splitted into 20 slices (Step 2); therefore, for each obtained slice (Step 3), the histogram of green, showing the occurrence of each level of green, is computed (Step 4). The area of the histogram constitutes an element of the feature vector. The extracted features are given as input to the aforementioned Classifiers: in Table 1 the chosen set of hyper parameters for each of the classifiers with the corresponding obtained accuracy, in terms of means and (1-sigma) repeatability, is summarized.
The obtained experimental results show that the one-hidden layer (L1) NN with 20 neurons and ReLU as activation function is the one that achieves the best performance, with an average accuracy of (99.9%) and a (1-sigma) repeatability of (1.9%). It outperforms the results obtained with the use of Sigmoidal and Tanh activation functions even with more neurons in the hidden layer. It was also found that both SVM and two-hidden layer NN exhibit worse results than the one-hidden layer networks. In fact, with SVM the best achieved accuracy is (54.5%), while with the two-hidden layer NN the best accuracy reached (85.2%).
Since the NN with one hidden layer achieved the best results, further tests were dedicated to fine tune the number of neurons. Table 2 and Fig. 4 summarize the detail of the performance for the one-hidden layer networks as both the number of neurons and the activation function are varied.
It can be observed that the best results are always obtained using ReLU as activation function. The number of neurons which achieved the greatest accuracy was confirmed to be 20.
The results reported in Table 2 were statistically validated by means of One-Way ANOVA and Fischer test, by verifying the statistical significance of the differences between the mean accuracies obtained by the three different activation functions used (Tanh, Sigmoid, and ReLU). The chosen null hypothesis (H_0) was that the groups belonged to the same population with a significance level (alpha) = 1.0%. The test rejected the null hypothesis with a P-value = (0.0%).
Therefore, the Paired t-test was carried out to understand which of the three groups is different from the others. The significance level (alpha) was again set equal to 1.0%. For all three tests the hypothesis (H_0), that assumed the groups were identical, was rejected. The analysis was conducted by means of the online tool Statistic Kindgom40. Further details are reported in Table 3.
The tests confirmed that the results obtained with ReLU activation function and 20 neurons are statistically relevant. In fact, there is a significant difference between this model and the others with different activation functions. Hence, this model was chosen for the classification stage of the proposed system and used in the prototype version.
For the sake of example, Fig. 5 shows the results obtained by applying the proposed algorithm to frames of the data set with good and bad perfusion.
For each considered frame, the corresponding ROI is indicated. In particular, Fig. 5a and d have ROI classified with good perfusion (good amount of green) and prediction 1. On the other hand, Fig. 5b and c has prediction output equal to 0 because the ROI are considered by the algorithm as bad perfused due to a low grade of green or due to a not uniform presence of green in the selected ROI. In each of the considered cases, the correctness of the classification was confirmed by the surgeons.
Details about the features extraction: Step 1: selection of the ROI. Step 2: the ROI is divided in 20 slice. Step 3: for each slice, the histogram of the green band is evaluated. Step 4: the amount of green of each histogram is evaluated and the obtained feature is used to build the feature vector sent to the ML classifier.
Comparison of (a) accuracy, and (b) 1-(sigma) repeatability for the three activation functions used with different neurons: Tanh (orange), Sigmoid (red), Rectifier Linear Unit (blue).
Four frames from the data set with respective ROI: (a) and (d) have ROI with adequate perfusion (high amount of green) and prediction 1. (b) and (c) have prediction 0 because the ROIs are inadequately perfused (low amount of green and/or not uniform ICG diffusion).
Setup
After the offline validation, the algorithm was further validated using the equipment available at the University Hospital Federico II in Naples, Italy. The aim was to ensure the possibility of interfacing the proposed system with the medical equipment. An additional aspect to consider is, in fact, the real-time interfacing with the endoscope. The endoscope used was the Olympus Visera Elite II. It is an imaging platform for general surgery, urology, gynecology, and more, which links the OR to other devices and facilities around the hospital. An S-video to USB adapter was used to connect the endoscope to a PC equipped with Windows 10 and Python 2.7. The video captured from the endoscope was transmitted in real time to an elaboration unit kept outside the OR. Therefore, surgeons who did not take part in the operation were asked to select the ROIs. However, this workflow can be also conducted by the main surgical team inside the OR.
Results
The algorithm was able to receive and process at least 30 frame per seconds (fps) from the video source: this frame rate was considered acceptable for the surgeons to select the ROI and use the system. Moreover, the output provided by the algorithm might effectively help surgeons to take the decision even in unclear situations (i.e., low brightness).
These additional trials demonstrated the feasibility of the practical implementation of the proposed ML-based algorithm Fig. 6, the output of the system working under three different levels of green brightness is shown. This also demonstrates that the proposed system is able to correctly classify the frame regardless of the level of green brightness.
Online validation: Frames characterized by different brightness levels acquired directly from the endoscope during the online validation: (a) is characterized by high brightness, (b) by medium brightness, and (c) by low brightness. Nevertheless, the real-time prediction works even in a low brightness scenario.
A system based on ML classifiers is proposed to assist the surgeons during laparoscopic colorectal surgery. It is a decision-support system able to automatically asses if the quality of the perfusion is adequate or inadequate after an injection of indiocyainine green dye. Different models of classifiers were tested on a dataset of videos of several anastomoses carried out at the Federico II Hospital. Overall, the one-hidden-layer NN with 20 neurons and ReLU activation function achieved the best performance. In fact, the obtained results showed a prediction accuracy of (99.9%) with a (1-sigma) repeatability of (1.9%). These results were statistically validated by means of (i) ANOVA, and (ii) Fischer and Paired-t tests. Therefore, this model was selected for the system implementation.
The proposed system was successfully validated also in relation to the interfacing with actual equipment used at the University Hospital Federico II in Naples, Italy. It can represent an important decision support to surgeons during the operation, especially in condition of uncertainty – where it is not clear whether the blood perfusion is adequate or not – due to an unclear presence of ICG.
Future work will be addressed to overcome the current research weakness, by (i) introducing more levels between adequate and inadequate perfusion, in order to increase the resolution of the assessment and further enhance accuracy of prediction, (ii) identifying a method to automatically select the ROIs, and (iii) enrich the dataset by facing circumstances when the blood perfusion is impaired by underlying pathologies (e.g., atherosclerosis). In this case, in fact, both the classifier and the surgeon are not trained to correctly assess whether perfusion is adequate or not.
The corresponding author is responsible for the safekeeping of all data, available upon request.
Donnachie, E. M. et al. Indocyanine green interference in the Kodak Ektachem determination of total bilirubin. Clin. Chem. 35, 899–900. https://doi.org/10.1093/clinchem/35.5.899 (1989).
CAS Article PubMed Google Scholar
Benya, R., Quintana, J. & Brundage, B. Adverse reactions to indocyanine green: a case report and a review of the literature. Cathet. Cardiovasc. Diagn. 17, 231–233 (1989).
CAS Article Google Scholar
Cherrick, G. R. et al. Indocyanine green: observations on its physical properties, plasma decay, and hepatic extraction. J. Clin. Investig. 39, 592–600 (1960).
CAS Article Google Scholar
Faybik, P. & Hetz, H. Plasma disappearance rate of indocyanine green in liver dysfunction. In Transplantation Proceedings, vol. 38, 801–802 (Elsevier, 2006).
De Nardi, P. et al. Intraoperative angiography with indocyanine green to assess anastomosis perfusion in patients undergoing laparoscopic colorectal resection: results of a multicenter randomized controlled trial. Surg. Endosc. 34, 53–60 (2020).
Article Google Scholar
Watanabe, J. et al. Indocyanine green fluorescence imaging to reduce the risk of anastomotic leakage in laparoscopic low anterior resection for rectal cancer: a propensity score-matched cohort study. Surg. Endosc. 34, 202–208 (2020).
Article Google Scholar
Hasegawa, H. et al. Impact of intraoperative indocyanine green fluorescence angiography on anastomotic leakage after laparoscopic sphincter-sparing surgery for malignant rectal tumors. Int. J. Colorectal Dis. 35, 471–480 (2020).
Article Google Scholar
Schwarz, C. et al. The value of indocyanine green clearance assessment to predict postoperative liver dysfunction in patients undergoing liver resection. Sci. Rep. 9, 1–7 (2019).
ADS Article Google Scholar
Ushijima, H. et al. Visualization of lymphatic flow in laparoscopic colon cancer surgery using indocyanine green fluorescence imaging. Sci. Rep. 10, 1–7 (2020).
Article Google Scholar
Merola, G. et al. Is robotic right colectomy economically sustainable? A multicentre retrospective comparative study and cost analysis. Surg. Endosc. 34, 4041–4047 (2020).
Article Google Scholar
Ahn, H.-M. et al. Optimization of indocyanine green angiography for colon perfusion during laparoscopic colorectal surgery. Colorectal Dis. 23, 1848–1859 (2021).
Article Google Scholar
Zhang, W. & Che, X. Effect of indocyanine green fluorescence angiography on preventing anastomotic leakage after colorectal surgery: a meta-analysis. Surg. Today 51, 1415–1428 (2021).
CAS Article Google Scholar
Emile, S. H., Khan, S. M. & Wexner, S. D. Impact of change in the surgical plan based on indocyanine green fluorescence angiography on the rates of colorectal anastomotic leak: a systematic review and meta-analysis. Surg. Endosc. 1–13 (2022).
Boni, L. et al. Indocyanine green-enhanced fluorescence to assess bowel perfusion during laparoscopic colorectal resection. Surg. Endosc. 30, 2736–2742 (2016).
Article Google Scholar
Dalli, J. et al. Digital dynamic discrimination of primary colorectal cancer using systemic indocyanine green with near-infrared endoscopy. Sci. Rep. 11, 1–12 (2021).
Article Google Scholar
Peltrini, R. et al. Intraoperative use of indocyanine green fluorescence imaging in rectal cancer surgery: The state of the art. World J. Gastroenterol. 27, 6374 (2021).
CAS Article Google Scholar
Wada, T. et al. Icg fluorescence imaging for quantitative evaluation of colonic perfusion in laparoscopic colorectal surgery. Surg. Endosc. 31, 4184–4193 (2017).
Article Google Scholar
Zehetner, J. et al. Intraoperative assessment of perfusion of the gastric graft and correlation with anastomotic leaks after esophagectomy. Ann. Surg. 262, 74 (2015).
Article Google Scholar
Diana, M. et al. Intraoperative fluorescence-based enhanced reality laparoscopic real-time imaging to assess bowel perfusion at the anastomotic site in an experimental model. J. Brit. Surg. 102, e169–e176 (2015).
CAS Article Google Scholar
Lütken, C. D., Achiam, M. P., Svendsen, M. B., Boni, L. & Nerup, N. Optimizing quantitative fluorescence angiography for visceral perfusion assessment. Surg. Endosc. 1–11 (2020).
Kaul, V., Enslin, S. & Gross, S. A. The history of artificial intelligence in medicine. Gastrointest. Endosc. (2020).
Krittanawong, C. et al. Integrating blockchain technology with artificial intelligence for cardiovascular medicine. Nat. Rev. Cardiol. 17, 1–3 (2020).
Article Google Scholar
Benjamens, S., Dhunnoo, P. & Meskó, B. The state of artificial intelligence-based fda-approved medical devices and algorithms: an online database. NPJ Digit. Med. 3, 1–8 (2020).
Article Google Scholar
Wang, Y. et al. Augmenting vascular disease diagnosis by vasculature-aware unsupervised learning. Nat. Mach. Intell. 2, 337–346 (2020).
Article Google Scholar
von der Emde, L. et al. Artificial intelligence for morphology-based function prediction in neovascular age-related macular degeneration. Sci. Rep. 9, 1–12 (2019).
ADS Article Google Scholar
Yu, K.-H., Beam, A. L. & Kohane, I. S. Artificial intelligence in healthcare. Nat. Biomed. Eng. 2, 719–731 (2018).
Article Google Scholar
Parmeggiani, D. et al. Artificial intelligence against breast cancer (annes-bc-project). Ann. Ital. Chir. 83, 1–5 (2012).
PubMed Google Scholar
Cahill, R. et al. Artificial intelligence indocyanine green (icg) perfusion for colorectal cancer intra-operative tissue classification. Br. J. Surg. 108, 5–9 (2021).
ADS CAS Article Google Scholar
Park, S.-H. et al. Artificial intelligence based real-time microcirculation analysis system for laparoscopic colorectal surgery. World J. Gastroenterol. 26, 6945 (2020).
CAS Article Google Scholar
Igaki, T. et al. Artificial intelligence-based total mesorectal excision plane navigation in laparoscopic colorectal surgery. Dis. Colon Rectum 65, e329–e333 (2022).
Article Google Scholar
Sánchez-Peralta, L. F., Bote-Curiel, L., Picón, A., Sánchez-Margallo, F. M. & Pagador, J. B. Deep learning to find colorectal polyps in colonoscopy: A systematic literature review. Artif. Intell. Med. 108, 101923 (2020).
Article Google Scholar
Kitaguchi, D. et al. Automated laparoscopic colorectal surgery workflow recognition using artificial intelligence: Experimental research. Int. J. Surg. 79, 88–94 (2020).
Article Google Scholar
Yao, R., Xia, S., Shen, F., Zhou, Y. & Niu, Q. Exploiting spatial structure from parts for adaptive kernelized correlation filter tracker. IEEE Signal Process. Lett. 23, 658–662 (2016).
ADS Article Google Scholar
Zhang, Z. Improved adam optimizer for deep neural networks. In 2018 IEEE/ACM 26th International Symposium on Quality of Service (IWQoS), 1–2 (IEEE, 2018).
Bock, S. & Weiß, M. A proof of local convergence for the adam optimizer. In 2019 International Joint Conference on Neural Networks (IJCNN), 1–8 (IEEE, 2019).
Ma, Y. & Guo, G. Support vector machines applications, vol. 649 (Springer, 2014).
Kohavi, R. et al. A study of cross-validation and bootstrap for accuracy estimation and model selection. In Ijcai, vol. 14, 1137–1145 (Montreal, Canada, 1995).
Arlot, S. et al. A survey of cross-validation procedures for model selection. Stat. Surv. 4, 40–79 (2010).
MathSciNet Article Google Scholar
Raschka, S. Model evaluation, model selection, and algorithm selection in machine learning. arXiv preprintarXiv:1811.12808 (2018).
Statistics Calculators. Statistics Kingdom. http://www.statskingdom.com.
Download references
This work was carried out as part of the “ICT for Health” project, which was financially supported by the Italian Ministry of Education, University and Research (MIUR), under the initiative ‘Departments of Excellence’ (Italian Law no. 232/2016), through an excellence grant awarded to the Department of Information Technology and Electrical Engineering of the University of Naples Federico II, Naples, Italy.
University of Naples Federico II – Interdepartmental Research Center in Health Management and Innovation in Healthcare (CIRMIS), Naples, 80131, Italy
Pasquale Arpaia, Umberto Bracale, Francesco Corcione, Egidio De Benedetto & Roberto Prevete
Department of Information Technology and Electrical Engineering, University of Naples Federico II, Naples, 80125, Italy
Pasquale Arpaia, Egidio De Benedetto, Alessandro Di Bernardo, Vincenzo Di Capua & Roberto Prevete
Department of Advanced Biomedical Sciences, University of Naples Federico II, Naples, 80131, Italy
Umberto Bracale
Department of Public Health, University of Naples Federico II, Naples, 80131, Italy
Francesco Corcione & Roberto Peltrini
Department of Electronics and Telecommunications, Polytechnic University of Turin, Turin, 10129, Italy
Luigi Duraccio
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
P.A. and F.C. conceived and supervised the study. U.B., E.D.B., R.P. and R.P. designed the experiments. V.D.C. , A.D.B. and L.D. performed the experiments. U.B., E.D.B., V.D.C., L.D., R. Peltrini and R. Prevete analyzed data. V.D.C. and L.D. wrote the manuscript. V.D.C. and A.D.B. prepared the figures. All authors reviewed the manuscript.
Correspondence to Pasquale Arpaia.
The authors declare no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Reprints and Permissions
Arpaia, P., Bracale, U., Corcione, F. et al. Assessment of blood perfusion quality in laparoscopic colorectal surgery by means of Machine Learning. Sci Rep 12, 14682 (2022). https://doi.org/10.1038/s41598-022-16030-8
Download citation
Received: 30 November 2021
Accepted: 04 July 2022
Published: 29 August 2022
DOI: https://doi.org/10.1038/s41598-022-16030-8
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.
Advertisement
Advanced search
© 2022 Springer Nature Limited
Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.