33 results on '"surgical tool detection"'
Search Results
2. Leveraging active learning techniques for surgical instrument recognition and localization.
- Author
-
PIOTROWSKI, Bartłomiej, Oszczak, Jakub, SAWICKI, Krzysztof, SIEMIĄTKOWSKA, Barbara, and CURATOLO, Andrea
- Subjects
- *
SURGICAL equipment , *OPHTHALMIC surgery , *RECURRENT neural networks , *SURGICAL instruments , *ARTIFICIAL intelligence - Abstract
The field of ophthalmic surgery demands accurate identification of specialized surgical instruments. Manual recognition can be time-consuming and prone to errors. In recent years neural networks have emerged as promising techniques for automating the classification process. However, the deployment of these advanced algorithms requires the collection of large amounts of data and a painstaking process of tagging selected elements. This paper presents a novel investigation into the application of neural networks for the detection and classification of surgical instruments in ophthalmic surgery. The main focus of the research is the application of active learning techniques, in which the model is trained by selecting the most informative instances to expand the training set. Various active learning methods are compared, with a focus on their effectiveness in reducing the need for significant data annotation -- a major concern in the field of surgery. The use of convolutional neural networks (CNNs) and recurrent neural networks (RNNs) to achieve high performance in the task of surgical tool detection is outlined. The combination of artificial intelligence (AI), machine learning, and Active Learning approaches, specifically in the field of ophthalmic surgery, opens new perspectives for improved diagnosis and surgical planning, ultimately leading to an improvement in patient safety and treatment outcomes. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Artificial intelligence model for automated surgical instrument detection and counting: an experimental proof-of-concept study
- Author
-
Ekamjit S. Deol, Grant Henning, Spyridon Basourakos, Ranveer M. S. Vasdev, Vidit Sharma, Nicholas L. Kavoussi, R. Jeffrey Karnes, Bradley C. Leibovich, Stephen A. Boorjian, and Abhinav Khanna
- Subjects
Retained surgical items ,Computer vision ,Artificial Intelligence ,Surgical tool detection ,Surgical safety ,Surgery ,RD1-811 - Abstract
Abstract Background Retained surgical items (RSI) are preventable events that pose a significant risk to patient safety. Current strategies for preventing RSIs rely heavily on manual instrument counting methods, which are prone to human error. This study evaluates the feasibility and performance of a deep learning-based computer vision model for automated surgical tool detection and counting. Methods A novel dataset of 1,004 images containing 13,213 surgical tools across 11 categories was developed. The dataset was split into training, validation, and test sets at a 60:20:20 ratio. An artificial intelligence (AI) model was trained on the dataset, and the model’s performance was evaluated using standard object detection metrics, including precision and recall. To simulate a real-world surgical setting, model performance was also evaluated in a dynamic surgical video of instruments being moved in real-time. Results The model demonstrated high precision (98.5%) and recall (99.9%) in distinguishing surgical tools from the background. It also exhibited excellent performance in differentiating between various surgical tools, with precision ranging from 94.0 to 100% and recall ranging from 97.1 to 100% across 11 tool categories. The model maintained strong performance on a subset of test images containing overlapping tools (precision range: 89.6–100%, and recall range 97.2–98.2%). In a real-time surgical video analysis, the model maintained a correct surgical tool count in all non-transition frames, with a median inference speed of 40.4 frames per second (interquartile range: 4.9). Conclusion This study demonstrates that using a deep learning-based computer vision model for automated surgical tool detection and counting is feasible. The model’s high precision and real-time inference capabilities highlight its potential to serve as an AI safeguard to potentially improve patient safety and reduce manual burden on surgical staff. Further validation in clinical settings is warranted.
- Published
- 2024
- Full Text
- View/download PDF
4. Artificial intelligence model for automated surgical instrument detection and counting: an experimental proof-of-concept study.
- Author
-
Deol, Ekamjit S., Henning, Grant, Basourakos, Spyridon, Vasdev, Ranveer M. S., Sharma, Vidit, Kavoussi, Nicholas L., Karnes, R. Jeffrey, Leibovich, Bradley C., Boorjian, Stephen A., and Khanna, Abhinav
- Subjects
- *
ARCHITECTURE , *COMPUTER simulation , *GRAPHIC arts , *RESEARCH funding , *ARTIFICIAL intelligence , *DESCRIPTIVE statistics , *EXPERIMENTAL design , *DEEP learning , *HYPOTHESIS , *SURGICAL count procedure , *SURGICAL instruments , *AUTOMATION , *ACCURACY , *DATA analysis software , *VIDEO recording , *REGRESSION analysis - Abstract
Background: Retained surgical items (RSI) are preventable events that pose a significant risk to patient safety. Current strategies for preventing RSIs rely heavily on manual instrument counting methods, which are prone to human error. This study evaluates the feasibility and performance of a deep learning-based computer vision model for automated surgical tool detection and counting. Methods: A novel dataset of 1,004 images containing 13,213 surgical tools across 11 categories was developed. The dataset was split into training, validation, and test sets at a 60:20:20 ratio. An artificial intelligence (AI) model was trained on the dataset, and the model's performance was evaluated using standard object detection metrics, including precision and recall. To simulate a real-world surgical setting, model performance was also evaluated in a dynamic surgical video of instruments being moved in real-time. Results: The model demonstrated high precision (98.5%) and recall (99.9%) in distinguishing surgical tools from the background. It also exhibited excellent performance in differentiating between various surgical tools, with precision ranging from 94.0 to 100% and recall ranging from 97.1 to 100% across 11 tool categories. The model maintained strong performance on a subset of test images containing overlapping tools (precision range: 89.6–100%, and recall range 97.2–98.2%). In a real-time surgical video analysis, the model maintained a correct surgical tool count in all non-transition frames, with a median inference speed of 40.4 frames per second (interquartile range: 4.9). Conclusion: This study demonstrates that using a deep learning-based computer vision model for automated surgical tool detection and counting is feasible. The model's high precision and real-time inference capabilities highlight its potential to serve as an AI safeguard to potentially improve patient safety and reduce manual burden on surgical staff. Further validation in clinical settings is warranted. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Real-time active constraint generation and enforcement for surgical tools using 3D detection and localisation network.
- Author
-
Souipas, Spyridon, Nguyen, Anh, Laws, Stephen G., Davies, Brian L., y Baena, Ferdinando Rodriguez, Raudonis, Vidas, and Li, Zhen
- Subjects
SURGICAL equipment ,INDUSTRIAL robots ,HAPTIC devices ,SURGICAL robots ,MULTI-degree of freedom ,ORTHOPEDIC surgery ,PATIENT safety - Abstract
Introduction: Collaborative robots, designed to work alongside humans for manipulating end-effectors, greatly benefit from the implementation of active constraints. This process comprises the definition of a boundary, followed by the enforcement of some control algorithm when the robot tooltip interacts with the generated boundary. Contact with the constraint boundary is communicated to the human operator through various potential forms of feedback. In fields like surgical robotics, where patient safety is paramount, implementing active constraints can prevent the robot from interacting with portions of the patient anatomy that shouldn't be operated on. Despite improvements in orthopaedic surgical robots, however, there exists a gap between bulky systems with haptic feedback capabilities and miniaturised systems that only allow for boundary control, where interaction with the active constraint boundary interrupts robot functions. Generally, active constraint generation relies on optical tracking systems and preoperative imaging techniques. Methods: This paper presents a refined version of the Signature Robot, a three degrees-of-freedom, hands-on collaborative system for orthopaedic surgery. Additionally, it presents a method for generating and enforcing active constraints "on-the-fly" using our previously introduced monocular, RGB, camera-based network, SimPS-Net. The network was deployed in real-time for the purpose of boundary definition. This boundary was subsequently used for constraint enforcement testing. The robot was utilised to test two different active constraints: a safe region and a restricted region. Results: The network success rate, defined as the ratio of correct over total object localisation results, was calculated to be 54.7% ± 5.2%. In the safe region case, haptic feedback resisted tooltip manipulation beyond the active constraint boundary, with a mean distance from the boundary of 2.70 mm ± 0.37 mm and a mean exit duration of 0.76 s ± 0.11 s. For the restricted-zone constraint, the operator was successfully prevented from penetrating the boundary in 100% of attempts. Discussion: This paper showcases the viability of the proposed robotic platform and presents promising results of a versatile constraint generation and enforcement pipeline. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. Real-time active constraint generation and enforcement for surgical tools using 3D detection and localisation network
- Author
-
Spyridon Souipas, Anh Nguyen, Stephen G. Laws, Brian L. Davies, and Ferdinando Rodriguez y Baena
- Subjects
active constraints ,surgical robotics ,surgical tool detection ,3D pose estimation ,surgical tool localisation ,Mechanical engineering and machinery ,TJ1-1570 ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
Introduction: Collaborative robots, designed to work alongside humans for manipulating end-effectors, greatly benefit from the implementation of active constraints. This process comprises the definition of a boundary, followed by the enforcement of some control algorithm when the robot tooltip interacts with the generated boundary. Contact with the constraint boundary is communicated to the human operator through various potential forms of feedback. In fields like surgical robotics, where patient safety is paramount, implementing active constraints can prevent the robot from interacting with portions of the patient anatomy that shouldn’t be operated on. Despite improvements in orthopaedic surgical robots, however, there exists a gap between bulky systems with haptic feedback capabilities and miniaturised systems that only allow for boundary control, where interaction with the active constraint boundary interrupts robot functions. Generally, active constraint generation relies on optical tracking systems and preoperative imaging techniques.Methods: This paper presents a refined version of the Signature Robot, a three degrees-of-freedom, hands-on collaborative system for orthopaedic surgery. Additionally, it presents a method for generating and enforcing active constraints “on-the-fly” using our previously introduced monocular, RGB, camera-based network, SimPS-Net. The network was deployed in real-time for the purpose of boundary definition. This boundary was subsequently used for constraint enforcement testing. The robot was utilised to test two different active constraints: a safe region and a restricted region.Results: The network success rate, defined as the ratio of correct over total object localisation results, was calculated to be 54.7% ± 5.2%. In the safe region case, haptic feedback resisted tooltip manipulation beyond the active constraint boundary, with a mean distance from the boundary of 2.70 mm ± 0.37 mm and a mean exit duration of 0.76 s ± 0.11 s. For the restricted-zone constraint, the operator was successfully prevented from penetrating the boundary in 100% of attempts.Discussion: This paper showcases the viability of the proposed robotic platform and presents promising results of a versatile constraint generation and enforcement pipeline.
- Published
- 2024
- Full Text
- View/download PDF
7. Identification of Surgical Forceps Using YOLACT++
- Author
-
Memida, Shoko, Miura, Satoshi, Ceccarelli, Marco, Series Editor, Agrawal, Sunil K., Advisory Editor, Corves, Burkhard, Advisory Editor, Glazunov, Victor, Advisory Editor, Hernández, Alfonso, Advisory Editor, Huang, Tian, Advisory Editor, Jauregui Correa, Juan Carlos, Advisory Editor, Takeda, Yukio, Advisory Editor, Petuya, Victor, editor, Quaglia, Giuseppe, editor, Parikyan, Tigran, editor, and Carbone, Giuseppe, editor
- Published
- 2023
- Full Text
- View/download PDF
8. Collaborative robot acting as scrub nurse for cataract surgery (CRASCS)
- Author
-
Rekha, D. and Kaliyappan, Harish Kumar
- Published
- 2024
- Full Text
- View/download PDF
9. A semi-supervised Teacher-Student framework for surgical tool detection and localization.
- Author
-
Teevno, Mansoor Ali, Ochoa-Ruiz, Gilberto, and Ali, Sharib
- Subjects
SUPERVISED learning ,MINIMALLY invasive procedures - Abstract
Surgical tool detection in minimally invasive surgery is an essential part of computer-assisted interventions. Current approaches are mostly based on supervised methods requiring large annotated datasets. However, labelled datasets are often scarce. Semi-supervised learning (SSL) has recently emerged as a viable alternative showing promise in producing models retaining competitive performance to supervised methods. Therefore, this paper introduces an SSL framework in the surgical tool detection paradigm, which aims to mitigate training data scarcity and data imbalance problems through a knowledge distillation approach. In the proposed work, we train a model with labelled data which initialises the Teacher-Student joint learning, where the Student is trained on Teacher-generated pseudo-labels from unlabelled data. We also propose a multi-class distance with a margin-based classification loss function in the region-of-interest head of the detector to segregate the foreground-background region effectively. Our results on m2cai16-tool-locations dataset indicates the superiority of our approach on different supervised data settings (1%, 2%, 5% and 10% of annotated data) where our model achieves overall improvements of 8%, 12%, and 27% in mean average precision on 1% labelled data over the state-of-the-art SSL methods and the supervised baseline, respectively. The code is available at . [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
10. Automated Surgical Procedure Assistance Framework Using Deep Learning and Formal Runtime Monitoring
- Author
-
Gupta, Gaurav, Shankar, Saumya, Pinisetty, Srinivas, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Dang, Thao, editor, and Stolz, Volker, editor
- Published
- 2022
- Full Text
- View/download PDF
11. Domain generalization improves end-to-end object detection for real-time surgical tool detection.
- Author
-
Reiter, Wolfgang
- Abstract
Purpose: Computer assistance for endoscopic surgery depends on knowledge about the contents in an endoscopic scene. An important step of analysing the video contents is real-time surgical tool detection. Most methods for tool detection nevertheless depend on multi-step algorithms building upon prior knowledge like anchor boxes or non-maximum suppression which ultimately decrease performance. A real-world difficulty encountered by learning-based methods are limited datasets. Training a neural network on data matching a specific distribution (e.g. from a single hospital or showing a specific type of surgery) can result in a lack of generalization. Methods: In this paper, we propose the application of a transformer based architecture for end-to-end tool detection. This architecture promises state-of-the-art accuracy while decreasing the complexity resulting in improved run-time performance. To improve the lack of cross-domain generalization due to limited datasets, we enhance the architecture with a latent feature space via variational encoding to capture common intra-domain information. This feature space models the linear dependencies between domains by constraining their rank. Results: The trained neural networks show a distinct improvement on out-of-domain data indicating better generalization to unseen domains. Inference with the end-to-end architecture can be performed at up to 138 frames per second (FPS) achieving a speedup in comparison to older approaches. Conclusions: Experimental results on three representative datasets demonstrate the performance of the method. We also show that our approach leads to better domain generalization. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
12. Surgical Tool Detection in Laparoscopic Videos by Modeling Temporal Dependencies Between Adjacent Frames
- Author
-
Jalal, N. A., Abdulbaki Alshirbaji, T., Docherty, P. D., Neumuth, T., Moeller, K., Magjarevic, Ratko, Series Editor, Ładyżyński, Piotr, Associate Editor, Ibrahim, Fatimah, Associate Editor, Lackovic, Igor, Associate Editor, Rock, Emilio Sacristan, Associate Editor, Jarm, Tomaz, editor, Cvetkoska, Aleksandra, editor, Mahnič-Kalamiza, Samo, editor, and Miklavcic, Damijan, editor
- Published
- 2021
- Full Text
- View/download PDF
13. Surgical Tool Detection in Open Surgery Videos.
- Author
-
Fujii, Ryo, Hachiuma, Ryo, Kajita, Hiroki, and Saito, Hideo
- Subjects
MINIMALLY invasive procedures ,VIDEOS ,CATARACT surgery ,SURGERY - Abstract
Detecting surgical tools is an essential task for analyzing and evaluating surgical videos. However, most studies focus on minimally invasive surgery (MIS) and cataract surgery. Mainly because of a lack of a large, diverse, and well-annotated dataset, research in the area of open surgery has been limited so far. Open surgery video analysis is challenging because of its properties: varied number and roles of people (e.g., main surgeon, assistant surgeons, and nurses), a complex interaction of tools and hands, various operative environments, and lighting conditions. In this paper, to handle these limitations and difficulties, we introduce an egocentric open surgery dataset that includes 15 open surgeries recorded with a head-mounted camera. More than 67k bounding boxes are labeled to 19k images with 31 surgical tool categories. Finally, we present a surgical tool detection baseline model based on recent advances in object detection. The results of our new dataset show that our presented dataset provides enough interesting challenges for future methods and that it can serve as a strong benchmark to address the study of tool detection in open surgery. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
14. Assessing Generalisation Capabilities of CNN Models for Surgical Tool Classification
- Author
-
Alshirbaji Tamer Abdulbaki, Jalal Nour Aldeen, Docherty Paul D., Neumuth Thomas, and Moeller Knut
- Subjects
cnn generalisability ,surgical tool detection ,laparoscopic images. ,Medicine - Abstract
Accurate recognition of surgical tools is a crucial component in the development of robust, context-aware systems. Recently, deep learning methods have been increasingly adopted to analyse laparoscopic videos. Existing work mainly leverages the ability of convolutional neural networks (CNNs) to model visual information of laparoscopic images. However, the performance was evaluated only on data belonging to the same dataset used for training. A more comprehensive evaluation of CNN performance on data from other datasets can provide a more rigorous assessment of the approaches. In this work, we investigate the generalisation capability of different CNN architectures to classify surgical tools in laparoscopic images recorded at different institutions. This research highlights the need to determine the effect of using data from different surgical sites on CNN generalisability. Experimental results imply that training a CNN model using data from multiple sites improves generalisability to new surgical locations.
- Published
- 2021
- Full Text
- View/download PDF
15. Gauze Detection and Segmentation in Minimally Invasive Surgery Video Using Convolutional Neural Networks.
- Author
-
Sánchez-Brizuela, Guillermo, Santos-Criado, Francisco-Javier, Sanz-Gobernado, Daniel, de la Fuente-López, Eusebio, Fraile, Juan-Carlos, Pérez-Turiel, Javier, and Cisnal, Ana
- Subjects
- *
CONVOLUTIONAL neural networks , *MINIMALLY invasive procedures , *MEDICAL equipment , *SURGICAL robots , *IMAGE segmentation , *OPERATING rooms - Abstract
Medical instruments detection in laparoscopic video has been carried out to increase the autonomy of surgical robots, evaluate skills or index recordings. However, it has not been extended to surgical gauzes. Gauzes can provide valuable information to numerous tasks in the operating room, but the lack of an annotated dataset has hampered its research. In this article, we present a segmentation dataset with 4003 hand-labelled frames from laparoscopic video. To prove the dataset potential, we analyzed several baselines: detection using YOLOv3, coarse segmentation, and segmentation with a U-Net. Our results show that YOLOv3 can be executed in real time but provides a modest recall. Coarse segmentation presents satisfactory results but lacks inference speed. Finally, the U-Net baseline achieves a good speed-quality compromise running above 30 FPS while obtaining an IoU of 0.85. The accuracy reached by U-Net and its execution speed demonstrate that precise and real-time gauze segmentation can be achieved, training convolutional neural networks on the proposed dataset. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
16. Graph Convolutional Nets for Tool Presence Detection in Surgical Videos
- Author
-
Wang, Sheng, Xu, Zheng, Yan, Chaochao, Huang, Junzhou, Hutchison, David, Editorial Board Member, Kanade, Takeo, Editorial Board Member, Kittler, Josef, Editorial Board Member, Kleinberg, Jon M., Editorial Board Member, Mattern, Friedemann, Editorial Board Member, Mitchell, John C., Editorial Board Member, Naor, Moni, Editorial Board Member, Pandu Rangan, C., Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Terzopoulos, Demetri, Editorial Board Member, Tygar, Doug, Editorial Board Member, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Chung, Albert C. S., editor, Gee, James C., editor, Yushkevich, Paul A., editor, and Bao, Siqi, editor
- Published
- 2019
- Full Text
- View/download PDF
17. A convolutional neural network with a two-stage LSTM model for tool presence detection in laparoscopic videos
- Author
-
Tamer Abdulbaki Alshirbaji, Jalal Nour Aldeen, and Möller Knut
- Subjects
convolutional neural network (cnn) ,endoscopic video ,spatiotemporal feature ,surgical tool detection ,Medicine - Abstract
Surgical tool presence detection in laparoscopic videos is a challenging problem that plays a critical role in developing context-aware systems in operating rooms (ORs). In this work, we propose a deep learning-based approach for detecting surgical tools in laparoscopic images using a convolutional neural network (CNN) in combination with two long short-term memory (LSTM) models. A pre-trained CNN model was trained to learn visual features from images. Then, LSTM was employed to include temporal information through a video clip of neighbour frames. Finally, the second LSTM was utilized to model temporal dependencies across the whole surgical video. Experimental evaluation has been conducted with the Cholec80 dataset to validate our approach. Results show that the most notable improvement is achieved after employing the two-stage LSTM model, and the proposed approach achieved better or similar performance compared with state-of-the-art methods.
- Published
- 2020
- Full Text
- View/download PDF
18. Real-Time Surgical Tool Detection in Minimally Invasive Surgery Based on Attention-Guided Convolutional Neural Network
- Author
-
Pan Shi, Zijian Zhao, Sanyuan Hu, and Faliang Chang
- Subjects
Attention mechanism ,convolutional neural network ,light-head module ,real-time ,surgical tool detection ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
To enhance surgeons' efficiency and safety of patients, minimally invasive surgery (MIS) is widely used in a variety of clinical surgeries. Real-time surgical tool detection plays an important role in MIS. However, most methods of surgical tool detection may not achieve a good trade-off between detection speed and accuracy. We propose a real-time attention-guided convolutional neural network (CNN) for frame-by-frame detection of surgical tools in MIS videos, which comprises a coarse (CDM) and a refined (RDM) detection modules. The CDM is used to coarsely regress the parameters of locations to get the refined anchors and perform binary classification, which determines whether the anchor is a tool or background. The RDM subtly incorporates the attention mechanism to generate accurate detection results utilizing the refined anchors from CDM. Finally, a light-head module for more efficient surgical tool detection is proposed. The proposed method is compared to eight state-of-the-art detection algorithms using two public (EndoVis Challenge and ATLAS Dione) datasets and a new dataset we introduced (Cholec80-locations), which extends the Cholec80 dataset with spatial annotations of surgical tools. Our approach runs in real-time at 55.5 FPS and achieves 100, 94.05, and 91.65% mAP for the above three datasets, respectively. Our method achieves accurate, fast, and robust detection results by end-to-end training in MIS videos. The results demonstrate the effectiveness and superiority of our method over the eight state-of-the-art methods.
- Published
- 2020
- Full Text
- View/download PDF
19. An Anchor-Free Convolutional Neural Network for Real-Time Surgical Tool Detection in Robot-Assisted Surgery
- Author
-
Yuying Liu, Zijian Zhao, Faliang Chang, and Sanyuan Hu
- Subjects
Anchor-free ,center point ,RAS ,single-stage ,stacked hourglass network ,surgical tool detection ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Robot-assisted surgery (RAS), a type of minimally invasive surgery, is used in a variety of clinical surgeries because it has a faster recovery rate and causes less pain. Automatic video analysis of RAS is an active research area, where precise surgical tool detection in real time is an important step. However, most deep learning methods currently employed for surgical tool detection are based on anchor boxes, which results in low detection speeds. In this paper, we propose an anchor-free convolutional neural network (CNN) architecture, a novel frame-by-frame method using a compact stacked hourglass network, which models the surgical tool as a single point: the center point of its bounding box. Our detector eliminates the need to design a set of anchor boxes, and is end-to-end differentiable, simpler, more accurate, and more efficient than anchor-box-based detectors. We believe our method is the first to incorporate the anchor-free idea for surgical tool detection in RAS videos. Experimental results show that our method achieves 98.5% mAP and 100% mAP at 37.0 fps on the ATLAS Dione and Endovis Challenge datasets, respectively, and truly realizes real-time surgical tool detection in RAS videos.
- Published
- 2020
- Full Text
- View/download PDF
20. Surgical Tool Detection in Open Surgery Videos
- Author
-
Ryo Fujii, Ryo Hachiuma, Hiroki Kajita, and Hideo Saito
- Subjects
surgical tool detection ,open surgery ,egocentric camera ,surgical video analysis ,deep neural network ,Technology ,Engineering (General). Civil engineering (General) ,TA1-2040 ,Biology (General) ,QH301-705.5 ,Physics ,QC1-999 ,Chemistry ,QD1-999 - Abstract
Detecting surgical tools is an essential task for analyzing and evaluating surgical videos. However, most studies focus on minimally invasive surgery (MIS) and cataract surgery. Mainly because of a lack of a large, diverse, and well-annotated dataset, research in the area of open surgery has been limited so far. Open surgery video analysis is challenging because of its properties: varied number and roles of people (e.g., main surgeon, assistant surgeons, and nurses), a complex interaction of tools and hands, various operative environments, and lighting conditions. In this paper, to handle these limitations and difficulties, we introduce an egocentric open surgery dataset that includes 15 open surgeries recorded with a head-mounted camera. More than 67k bounding boxes are labeled to 19k images with 31 surgical tool categories. Finally, we present a surgical tool detection baseline model based on recent advances in object detection. The results of our new dataset show that our presented dataset provides enough interesting challenges for future methods and that it can serve as a strong benchmark to address the study of tool detection in open surgery.
- Published
- 2022
- Full Text
- View/download PDF
21. Gauze Detection and Segmentation in Minimally Invasive Surgery Video Using Convolutional Neural Networks
- Author
-
Guillermo Sánchez-Brizuela, Francisco-Javier Santos-Criado, Daniel Sanz-Gobernado, Eusebio de la Fuente-López, Juan-Carlos Fraile, Javier Pérez-Turiel, and Ana Cisnal
- Subjects
convolutional neural networks ,image segmentation ,image object detection ,surgical tool detection ,minimally invasive surgery ,Chemical technology ,TP1-1185 - Abstract
Medical instruments detection in laparoscopic video has been carried out to increase the autonomy of surgical robots, evaluate skills or index recordings. However, it has not been extended to surgical gauzes. Gauzes can provide valuable information to numerous tasks in the operating room, but the lack of an annotated dataset has hampered its research. In this article, we present a segmentation dataset with 4003 hand-labelled frames from laparoscopic video. To prove the dataset potential, we analyzed several baselines: detection using YOLOv3, coarse segmentation, and segmentation with a U-Net. Our results show that YOLOv3 can be executed in real time but provides a modest recall. Coarse segmentation presents satisfactory results but lacks inference speed. Finally, the U-Net baseline achieves a good speed-quality compromise running above 30 FPS while obtaining an IoU of 0.85. The accuracy reached by U-Net and its execution speed demonstrate that precise and real-time gauze segmentation can be achieved, training convolutional neural networks on the proposed dataset.
- Published
- 2022
- Full Text
- View/download PDF
22. Surgical Tool Classification in Laparoscopic Videos Using Convolutional Neural Network
- Author
-
Abdulbaki Alshirbaji Tamer, Jalal Nour Aldeen, and Möller Knut
- Subjects
surgical tool detection ,loss-sensitive learning ,laparoscopic videos ,imbalanced data ,Medicine - Abstract
Laparoscopic videos are a very important source of information which is inherently available in minimally invasive surgeries. Detecting surgical tools based on that videos have gained increasing interest due to its importance in developing a context-aware system. Such system can provide guidance assistance to the surgical team and optimise the processes inside the operating room. Convolutional neural network is a robust method to learn discriminative visual features and classify objects. As it expects a uniform distribution of data over classes, it fails to identify classes which are under-presented in the training data. In this work, loss-sensitive learning approach and resampling techniques were applied to counter the negative effects of imbalanced laparoscopic data on training the CNN model. The obtained results showed improvement in the classification performance especially for detecting surgical tools which are shortly used in the procedure.
- Published
- 2018
- Full Text
- View/download PDF
23. Assessing Generalisation Capabilities of CNN Models for Surgical Tool Classification
- Author
-
Tamer Abdulbaki Alshirbaji, Nour Aldeen Jalal, Paul D. Docherty, Thomas Neumuth, and Knut Moeller
- Subjects
laparoscopic images ,cnn generalisability ,surgical tool detection ,Biomedical Engineering ,Medicine - Abstract
Accurate recognition of surgical tools is a crucial component in the development of robust, context-aware systems. Recently, deep learning methods have been increasingly adopted to analyse laparoscopic videos. Existing work mainly leverages the ability of convolutional neural networks (CNNs) to model visual information of laparoscopic images. However, the performance was evaluated only on data belonging to the same dataset used for training. A more comprehensive evaluation of CNN performance on data from other datasets can provide a more rigorous assessment of the approaches. In this work, we investigate the generalisation capability of different CNN architectures to classify surgical tools in laparoscopic images recorded at different institutions. This research highlights the need to determine the effect of using data from different surgical sites on CNN generalisability. Experimental results imply that training a CNN model using data from multiple sites improves generalisability to new surgical locations.
- Published
- 2021
24. Surgical Tool Classification in Laparoscopic Videos Using Convolutional Neural Network.
- Author
-
Alshirbaji, Tamer Abdulbaki, Jalal, Nour Aldeen, and Möller, Knut
- Published
- 2018
- Full Text
- View/download PDF
25. Addressing multi-label imbalance problem of surgical tool detection using CNN.
- Author
-
Sahu, Manish, Mukhopadhyay, Anirban, Szengel, Angelika, and Zachow, Stefan
- Abstract
Purpose: A fully automated surgical tool detection framework is proposed for endoscopic video streams. State-of-the-art surgical tool detection methods rely on supervised one-vs-all or multi-class classification techniques, completely ignoring the co-occurrence relationship of the tools and the associated class imbalance. Methods: In this paper, we formulate tool detection as a multi-label classification task where tool co-occurrences are treated as separate classes. In addition, imbalance on tool co-occurrences is analyzed and stratification techniques are employed to address the imbalance during convolutional neural network (CNN) training. Moreover, temporal smoothing is introduced as an online post-processing step to enhance runtime prediction. Results: Quantitative analysis is performed on the M2CAI16 tool detection dataset to highlight the importance of stratification, temporal smoothing and the overall framework for tool detection. Conclusion: The analysis on tool imbalance, backed by the empirical results, indicates the need and superiority of the proposed framework over state-of-the-art techniques. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
26. Real-Time Surgical Tool Detection in Minimally Invasive Surgery Based on Attention-Guided Convolutional Neural Network
- Author
-
Faliang Chang, Zijian Zhao, Sanyuan Hu, and Pan Shi
- Subjects
General Computer Science ,Computer science ,Attention mechanism ,convolutional neural network ,real-time ,02 engineering and technology ,light-head module ,Convolutional neural network ,030218 nuclear medicine & medical imaging ,03 medical and health sciences ,0302 clinical medicine ,RDM ,Atlas (anatomy) ,surgical tool detection ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,General Materials Science ,business.industry ,General Engineering ,Pattern recognition ,medicine.anatomical_structure ,Invasive surgery ,020201 artificial intelligence & image processing ,Artificial intelligence ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,business ,lcsh:TK1-9971 - Abstract
To enhance surgeons’ efficiency and safety of patients, minimally invasive surgery (MIS) is widely used in a variety of clinical surgeries. Real-time surgical tool detection plays an important role in MIS. However, most methods of surgical tool detection may not achieve a good trade-off between detection speed and accuracy. We propose a real-time attention-guided convolutional neural network (CNN) for frame-by-frame detection of surgical tools in MIS videos, which comprises a coarse (CDM) and a refined (RDM) detection modules. The CDM is used to coarsely regress the parameters of locations to get the refined anchors and perform binary classification, which determines whether the anchor is a tool or background. The RDM subtly incorporates the attention mechanism to generate accurate detection results utilizing the refined anchors from CDM. Finally, a light-head module for more efficient surgical tool detection is proposed. The proposed method is compared to eight state-of-the-art detection algorithms using two public (EndoVis Challenge and ATLAS Dione) datasets and a new dataset we introduced (Cholec80-locations), which extends the Cholec80 dataset with spatial annotations of surgical tools. Our approach runs in real-time at 55.5 FPS and achieves 100, 94.05, and 91.65% mAP for the above three datasets, respectively. Our method achieves accurate, fast, and robust detection results by end-to-end training in MIS videos. The results demonstrate the effectiveness and superiority of our method over the eight state-of-the-art methods.
- Published
- 2020
27. Surgical Tool Classification in Laparoscopic Videos Using Convolutional Neural Network
- Author
-
Tamer Abdulbaki Alshirbaji, Nour Aldeen Jalal, and Knut Möller
- Subjects
business.industry ,Computer science ,loss-sensitive learning ,Biomedical Engineering ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Pattern recognition ,Convolutional neural network ,Imbalanced data ,030218 nuclear medicine & medical imaging ,03 medical and health sciences ,0302 clinical medicine ,030220 oncology & carcinogenesis ,surgical tool detection ,laparoscopic videos ,Medicine ,imbalanced data ,Artificial intelligence ,business - Abstract
Laparoscopic videos are a very important source of information which is inherently available in minimally invasive surgeries. Detecting surgical tools based on that videos have gained increasing interest due to its importance in developing a context-aware system. Such system can provide guidance assistance to the surgical team and optimise the processes inside the operating room. Convolutional neural network is a robust method to learn discriminative visual features and classify objects. As it expects a uniform distribution of data over classes, it fails to identify classes which are under-presented in the training data. In this work, loss-sensitive learning approach and resampling techniques were applied to counter the negative effects of imbalanced laparoscopic data on training the CNN model. The obtained results showed improvement in the classification performance especially for detecting surgical tools which are shortly used in the procedure.
- Published
- 2018
28. A new weakly supervised strategy for surgical tool detection.
- Author
-
Xue, Yao, Liu, Siming, Li, Yonghui, Wang, Ping, and Qian, Xueming
- Subjects
- *
TOOLS , *VISION , *MINES & mineral resources - Abstract
Surgical tool detection is a recently active research area. It is the foundation to a series of advanced surgical support functions, such as image guided surgical navigation, forming safety zone between surgical tools and sensitive tissues. Previous methods rely on two types of information: tool locating signals and vision features. Collecting tool locating signals requires additional hardware equipments. Vision based methods train their detection models using strong annotations (e.g. bounding boxes), which are quite rare and expensive to acquire in the field of surgical image understanding. In this paper, we propose a Pseudo Supervised surgical Tool detection (PSTD) framework, which performs explicit detection refinement by three levels of associated measures (pseudo bounding box generation, real box regression, weighted boxes fusion) in a weakly supervised manner. On the basis of PSTD, we develop a Bi-directional Adaption Weighting (BAW) mechanism in our tool classifier for contextual information mining by creating competition or cooperation relationships between channels. By only using image-level tool category labels, the proposed method yields state-of-the-art results with 87.0% mAP on a mainstream surgical image dataset: Cheloc80. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
29. Video analysis for augmented cataract surgery
- Author
-
Al Hajj, Hassan, Laboratoire de Traitement de l'Information Medicale (LaTIM), Institut National de la Santé et de la Recherche Médicale (INSERM)-IMT Atlantique Bretagne-Pays de la Loire (IMT Atlantique), Institut Mines-Télécom [Paris] (IMT)-Institut Mines-Télécom [Paris] (IMT)-Centre Hospitalier Régional Universitaire de Brest (CHRU Brest)-Université de Brest (UBO)-Institut Brestois Santé Agro Matière (IBSAM), Université de Brest (UBO), Université de Bretagne occidentale - Brest, and Béatrice Cochener
- Subjects
Détection des outils chirurgicaux ,Convolutional and recurrent neural networks ,Surgical tool detection ,Chirurgie de la cataracte ,Réseaux de neurones récurrents et à convolutions ,Analyse vidéo ,Video analysis ,Cataract surgery ,[SDV.MHEP]Life Sciences [q-bio]/Human health and pathology - Abstract
The digital era is increasingly changing the world due to the sheer volume of data produced every day. The medical domain is highly affected by this revolution, because analysing this data can be a source of education/support for the clinicians. In this thesis, we propose to reuse the surgery videos recorded in the operating rooms for computer-assisted surgery system. We are chiefly interested in recognizing the surgical gesture being performed at each instant in order to provide relevant information. To achieve this goal, this thesis addresses the surgical tool recognition problem, with applications in cataract surgery. The main objective of this thesis is to address the surgical tool recognition problem in cataract surgery videos.In the surgical field, those tools are partially visible in videos and highly similar to one another. To address the visual challenges in the cataract surgical field, we propose to add an additional camera filming the surgical tray. Our goal is to detect the tool presence in the two complementary types of videos: tool-tissue interaction and surgical tray videos. The former records the patient's eye and the latter records the surgical tray activities.Two tasks are proposed to perform the task on the surgical tray videos: tools change detection and tool presence detection.First, we establish a similar pipeline for both tasks. It is based on standard classification methods on top of visual learning features. It yields satisfactory results for the tools change task, howev-lateer, it badly performs the surgical tool presence task on the tray. Second, we design deep learning architectures for the surgical tool detection on both video types in order to address the difficulties in manually designing the visual features.To alleviate the inherent challenges on the surgical tray videos, we propose to generate simulated surgical tray scenes along with a patch-based convolutional neural network (CNN).Ultimately, we study the temporal information using RNN processing the CNN results. Contrary to our primary hypothesis, the experimental results show deficient results for surgical tool presence on the tray but very good results on the tool-tissue interaction videos. We achieve even better results in the surgical field after fusing the tool change information coming from the tray and tool presence signals on the tool-tissue interaction videos.; L’ère numérique change de plus en plus le monde en raison de la quantité de données récoltées chaque jour. Le domaine médical est fortement affecté par cette explosion, car l’exploitation de ces données est un véritable atout pour l’aide à la pratique médicale. Dans cette thèse, nous proposons d’utiliser les vidéos chirurgicales dans le but de créer un système de chirurgie assistée par ordinateur. Nous nous intéressons principalement à reconnaître les gestes chirurgicaux à chaque instant afin de fournir aux chirurgiens des recommandations et des informations pertinentes. Pour ce faire, l’objectif principal de cette thèse est de reconnaître les outils chirurgicaux dans les vidéos de chirurgie de la cataracte. Dans le flux vidéo du microscope, ces outils sont partiellement visibles et certains se ressemblent beaucoup. Pour relever ces défis, nous proposons d'ajouter une caméra supplémentaire filmant la table opératoire. Notre objectif est donc de détecter la présence des outils dans les deux types de flux vidéo : les vidéos du microscope et les vidéos de la table opératoire. Le premier enregistre l'oeil du patient et le second enregistre les activités de la table opératoire. Deux tâches sont proposées pour détecter les outils dans les vidéos de la table : la détection des changements et la détection de présence d'outil. Dans un premier temps, nous proposons un système similaire pour ces deux tâches. Il est basé sur l’extraction des caractéristiques visuelles avec des méthodes de classification classique. Il fournit des résultats satisfaisants pour la détection de changement, cependant, il fonctionne insuffisamment bien pour la tâche de détection de présence des outils sur la table. Dans un second temps, afin de résoudre le problème du choix des caractéristiques, nous utilisons des architectures d’apprentissage profond pour la détection d'outils chirurgicaux sur les deux types de vidéo. Pour surmonter les défis rencontrés dans les vidéos de la table, nous proposons de générer des vidéos artificielles imitant la scène de la table opératoire et d’utiliser un réseau de neurones à convolutions (CNN) à base de patch. Enfin, nous exploitons l'information temporelle en utilisant un réseau de neurones récurrent analysant les résultats de CNNs. Contrairement à notre hypothèse, les expérimentations montrent des résultats insuffisants pour la détection de présence des outils sur la table, mais de très bons résultats dans les vidéos du microscope. Nous obtenons des résultats encore meilleurs dans les vidéos du microscope après avoir fusionné l’information issue de la détection des changements sur la table et la présence des outils dans l’oeil.
- Published
- 2018
30. Analyse vidéo pour la chirurgie de la cataracte augmentée
- Author
-
Al Hajj, Hassan, Laboratoire de Traitement de l'Information Medicale (LaTIM), Université de Brest (UBO)-Institut National de la Santé et de la Recherche Médicale (INSERM)-Centre Hospitalier Régional Universitaire de Brest (CHRU Brest)-IMT Atlantique Bretagne-Pays de la Loire (IMT Atlantique), Institut Mines-Télécom [Paris] (IMT)-Institut Mines-Télécom [Paris] (IMT)-Institut Brestois Santé Agro Matière (IBSAM), Université de Brest (UBO), Université de Bretagne occidentale - Brest, and Béatrice Cochener
- Subjects
Détection des outils chirurgicaux ,Convolutional and recurrent neural networks ,Surgical tool detection ,Chirurgie de la cataracte ,Réseaux de neurones récurrents et à convolutions ,Analyse vidéo ,Video analysis ,Cataract surgery ,[SDV.MHEP]Life Sciences [q-bio]/Human health and pathology - Abstract
The digital era is increasingly changing the world due to the sheer volume of data produced every day. The medical domain is highly affected by this revolution, because analysing this data can be a source of education/support for the clinicians. In this thesis, we propose to reuse the surgery videos recorded in the operating rooms for computer-assisted surgery system. We are chiefly interested in recognizing the surgical gesture being performed at each instant in order to provide relevant information. To achieve this goal, this thesis addresses the surgical tool recognition problem, with applications in cataract surgery. The main objective of this thesis is to address the surgical tool recognition problem in cataract surgery videos.In the surgical field, those tools are partially visible in videos and highly similar to one another. To address the visual challenges in the cataract surgical field, we propose to add an additional camera filming the surgical tray. Our goal is to detect the tool presence in the two complementary types of videos: tool-tissue interaction and surgical tray videos. The former records the patient's eye and the latter records the surgical tray activities.Two tasks are proposed to perform the task on the surgical tray videos: tools change detection and tool presence detection.First, we establish a similar pipeline for both tasks. It is based on standard classification methods on top of visual learning features. It yields satisfactory results for the tools change task, howev-lateer, it badly performs the surgical tool presence task on the tray. Second, we design deep learning architectures for the surgical tool detection on both video types in order to address the difficulties in manually designing the visual features.To alleviate the inherent challenges on the surgical tray videos, we propose to generate simulated surgical tray scenes along with a patch-based convolutional neural network (CNN).Ultimately, we study the temporal information using RNN processing the CNN results. Contrary to our primary hypothesis, the experimental results show deficient results for surgical tool presence on the tray but very good results on the tool-tissue interaction videos. We achieve even better results in the surgical field after fusing the tool change information coming from the tray and tool presence signals on the tool-tissue interaction videos.; L’ère numérique change de plus en plus le monde en raison de la quantité de données récoltées chaque jour. Le domaine médical est fortement affecté par cette explosion, car l’exploitation de ces données est un véritable atout pour l’aide à la pratique médicale. Dans cette thèse, nous proposons d’utiliser les vidéos chirurgicales dans le but de créer un système de chirurgie assistée par ordinateur. Nous nous intéressons principalement à reconnaître les gestes chirurgicaux à chaque instant afin de fournir aux chirurgiens des recommandations et des informations pertinentes. Pour ce faire, l’objectif principal de cette thèse est de reconnaître les outils chirurgicaux dans les vidéos de chirurgie de la cataracte. Dans le flux vidéo du microscope, ces outils sont partiellement visibles et certains se ressemblent beaucoup. Pour relever ces défis, nous proposons d'ajouter une caméra supplémentaire filmant la table opératoire. Notre objectif est donc de détecter la présence des outils dans les deux types de flux vidéo : les vidéos du microscope et les vidéos de la table opératoire. Le premier enregistre l'oeil du patient et le second enregistre les activités de la table opératoire. Deux tâches sont proposées pour détecter les outils dans les vidéos de la table : la détection des changements et la détection de présence d'outil. Dans un premier temps, nous proposons un système similaire pour ces deux tâches. Il est basé sur l’extraction des caractéristiques visuelles avec des méthodes de classification classique. Il fournit des résultats satisfaisants pour la détection de changement, cependant, il fonctionne insuffisamment bien pour la tâche de détection de présence des outils sur la table. Dans un second temps, afin de résoudre le problème du choix des caractéristiques, nous utilisons des architectures d’apprentissage profond pour la détection d'outils chirurgicaux sur les deux types de vidéo. Pour surmonter les défis rencontrés dans les vidéos de la table, nous proposons de générer des vidéos artificielles imitant la scène de la table opératoire et d’utiliser un réseau de neurones à convolutions (CNN) à base de patch. Enfin, nous exploitons l'information temporelle en utilisant un réseau de neurones récurrent analysant les résultats de CNNs. Contrairement à notre hypothèse, les expérimentations montrent des résultats insuffisants pour la détection de présence des outils sur la table, mais de très bons résultats dans les vidéos du microscope. Nous obtenons des résultats encore meilleurs dans les vidéos du microscope après avoir fusionné l’information issue de la détection des changements sur la table et la présence des outils dans l’oeil.
- Published
- 2018
31. Addressing Hospital Staffing Shortages: Dynamic Surgical Tool Tracking and Delivery Using Baxter
- Author
-
Sthitapragyan Parida
- Subjects
intelligent systems ,Engineering ,Artificial neural network ,gesture-based input ,business.industry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Staffing ,Intelligent decision support system ,Economic shortage ,neural networks ,computer vision ,multi-ANN systems ,Robot Operating System (ROS) ,Engineering management ,machine learning ,surgical tool delivery ,surgical tool detection ,Baxter ,tool classification ,Operations management ,Tracking (education) ,business - Abstract
Several hospitals face nurse staffing shortages for surgeries. This research focuses on building a system with a Baxter robot capable of identifying surgical tools using computer vision and delivering them to the surgeon on demand. This would deal with the issue of nurse unavailability during simple surgical procedures. The key aspects of the project were testing the accuracies of various artificial neural networks (ANNs) in classifying surgical instruments and programming Baxter to implement a surgical tool delivery system using magnets at the tip of its 7 degrees of freedom (DOF) robotic arms. The methodology consisted of first implementing algorithms to enable Baxter to do pick and deliver tasks for surgical tools, and second, gathering HuMoments of various tools using the cameras on Baxter's arm, which were then used to train the ANNs. Tool detection accuracies of ANNs with hidden layer neuron number varying from 5–50 and learning rates varying from 0.005‒0.1 were collected. Then, the tool identification and tool delivery system were merged together to create a turn-by-turn dynamic tool tracking and delivery system, which retrieved tools, based on the surgeons input, through a Leap Motion Controller. In addition to delivery, the system was modified to retrieve used tools from the surgeon, using a computer vision-based approach. The optimal ANN configuration consisted of an ensemble of various ANNs working together and achieved a detection accuracy of 93%. The average time taken for a mock abdominal incision surgery with the system is expected to be around 10 minutes and 30 seconds.
- Published
- 2015
32. Addressing Hospital Staffing Shortages: Dynamic Surgical Tool Tracking and Delivery Using Baxter
- Author
-
Parida, Sthitapragyan and Parida, Sthitapragyan
- Abstract
Several hospitals face nurse staffing shortages for surgeries. This research focuses on building a system with a Baxter robot capable of identifying surgical tools using computer vision and delivering them to the surgeon on demand. This would deal with the issue of nurse unavailability during simple surgical procedures. The key aspects of the project were testing the accuracies of various artificial neural networks (ANNs) in classifying surgical instruments and programming Baxter to implement a surgical tool delivery system using magnets at the tip of its 7 degrees of freedom (DOF) robotic arms. The methodology consisted of first implementing algorithms to enable Baxter to do pick and deliver tasks for surgical tools, and second, gathering HuMoments of various tools using the cameras on Baxter's arm, which were then used to train the ANNs. Tool detection accuracies of ANNs with hidden layer neuron number varying from 5–50 and learning rates varying from 0.005‒0.1 were collected. Then, the tool identification and tool delivery system were merged together to create a turn-by-turn dynamic tool tracking and delivery system, which retrieved tools, based on the surgeons input, through a Leap Motion Controller. In addition to delivery, the system was modified to retrieve used tools from the surgeon, using a computer vision-based approach. The optimal ANN configuration consisted of an ensemble of various ANNs working together and achieved a detection accuracy of 93%. The average time taken for a mock abdominal incision surgery with the system is expected to be around 10 minutes and 30 seconds.
- Published
- 2015
33. Dynamic Surgical Tool Tracking and Delivery System using Baxter Robot
- Author
-
Parida, Sthitapragyan, Wachs, Juan Pablo, and Cabrera, Maria Eugenia
- Subjects
intelligent systems ,surgical tool detection ,Industrial Engineering ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Baxter ,Robotics ,artificial neural networks ,computer vision - Abstract
Several hospitals face nurse staffing shortages for surgeries. This research focuses on building a system with Baxter robot capable of identifying surgical tools using computer vision and delivering them to the surgeon on demand. This would deal with the issue of nurse unavailability during simple surgical procedures. The key aspects of the project were: testing the accuracies of various Artificial Neural Networks (ANNs) in classifying surgical instruments, and programming Baxter to implement a surgical tool delivery system using magnets at the tip of its 7-DOF robotic arms. The methodology consisted of, first, implementing algorithms to enable Baxter to do pick and deliver tasks for surgical tools, and second, gathering Hu-Moments of various tools using the cameras on Baxter's arm, which were then used to train the ANNs. Tool detection accuracies of ANNs with hidden layer neuron number varying from 5 – 50 and learning rates varying from 0.005 - 0.1 were collected. Then, the tool identification and tool delivery system were merged together to create a turn by turn dynamic tool tracking and delivery system, which retrieved tools, based on the surgeons input, through a Leap Motion sensor. In addition to delivery, the system was modified to retrieve used tools back from the surgeon, using a computer vision based approach. The optimal ANN configuration consisted of an ensemble of various ANNs working together and achieved a detection accuracy of 93%. The average time taken for a mock abdominal incision surgery with the system is expected to be around 10 min and 30 sec.
- Published
- 2014
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.