Loading…

Surgical gestures can be used to assess surgical competence in robot-assisted surgery

To collect validity evidence for the assessment of surgical competence through the classification of general surgical gestures for a simulated robot-assisted radical prostatectomy (RARP). We used 165 video recordings of novice and experienced RARP surgeons performing three parts of the RARP procedur...

Full description

Saved in:
Bibliographic Details
Published in:Journal of robotic surgery 2024-12, Vol.18 (1), p.47
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue 1
container_start_page 47
container_title Journal of robotic surgery
container_volume 18
description To collect validity evidence for the assessment of surgical competence through the classification of general surgical gestures for a simulated robot-assisted radical prostatectomy (RARP). We used 165 video recordings of novice and experienced RARP surgeons performing three parts of the RARP procedure on the RobotiX Mentor. We annotated the surgical tasks with different surgical gestures: dissection, hemostatic control, application of clips, needle handling, and suturing. The gestures were analyzed using idle time (periods with minimal instrument movements) and active time (whenever a surgical gesture was annotated). The distribution of surgical gestures was described using a one-dimensional heat map, snail tracks. All surgeons had a similar percentage of idle time but novices had longer phases of idle time (mean time: 21 vs. 15 s, p < 0.001). Novices used a higher total number of surgical gestures (number of phases: 45 vs. 35, p < 0.001) and each phase was longer compared with those of the experienced surgeons (mean time: 10 vs. 8 s, p < 0.001). There was a different pattern of gestures between novices and experienced surgeons as seen by a different distribution of the phases. General surgical gestures can be used to assess surgical competence in simulated RARP and can be displayed as a visual tool to show how performance is improving. The established pass/fail level may be used to ensure the competence of the residents before proceeding with supervised real-life surgery. The next step is to investigate if the developed tool can optimize automated feedback during simulator training.
doi_str_mv 10.1007/s11701-023-01807-4
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_3148950010</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3148950010</sourcerecordid><originalsourceid>FETCH-LOGICAL-p157t-c8b1e9c4a2d01fb5cb0ec732f88f95ffa810b5e3a5c2f67408876c93dd05eeae3</originalsourceid><addsrcrecordid>eNo9zz9PwzAQBXALgUQpfAEmS8yGuziOnRFVQJEqMUDnynHOVaqSBJ8z8O0J4s90b_jpPZ0Q1wi3CGDvGNECKii0AnRgVXkiFugqrYqyxtP_7PS5uGA-ABhrNC7E9nVK-y74o9wT5ykRy-B72ZCcmFqZB-mZiVnynwvD-0iZ-kCy62UamiGr2XScZ_-tKH1eirPoj0xXv3cpto8Pb6u12rw8Pa_uN2pEY7MKrkGqQ-mLFjA2JjRAweoiOhdrE6N3CI0h7U0oYmVLcM5WodZtC4bIk16Km5_eMQ0f0_zA7jBMqZ8ndxpLVxsABP0FntxU1w</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3148950010</pqid></control><display><type>article</type><title>Surgical gestures can be used to assess surgical competence in robot-assisted surgery</title><source>Springer Nature</source><description>To collect validity evidence for the assessment of surgical competence through the classification of general surgical gestures for a simulated robot-assisted radical prostatectomy (RARP). We used 165 video recordings of novice and experienced RARP surgeons performing three parts of the RARP procedure on the RobotiX Mentor. We annotated the surgical tasks with different surgical gestures: dissection, hemostatic control, application of clips, needle handling, and suturing. The gestures were analyzed using idle time (periods with minimal instrument movements) and active time (whenever a surgical gesture was annotated). The distribution of surgical gestures was described using a one-dimensional heat map, snail tracks. All surgeons had a similar percentage of idle time but novices had longer phases of idle time (mean time: 21 vs. 15 s, p &lt; 0.001). Novices used a higher total number of surgical gestures (number of phases: 45 vs. 35, p &lt; 0.001) and each phase was longer compared with those of the experienced surgeons (mean time: 10 vs. 8 s, p &lt; 0.001). There was a different pattern of gestures between novices and experienced surgeons as seen by a different distribution of the phases. General surgical gestures can be used to assess surgical competence in simulated RARP and can be displayed as a visual tool to show how performance is improving. The established pass/fail level may be used to ensure the competence of the residents before proceeding with supervised real-life surgery. The next step is to investigate if the developed tool can optimize automated feedback during simulator training.</description><identifier>ISSN: 1863-2483</identifier><identifier>EISSN: 1863-2491</identifier><identifier>DOI: 10.1007/s11701-023-01807-4</identifier><language>eng</language><publisher>Heidelberg: Springer Nature B.V</publisher><subject>Annotations ; Bladder ; Control equipment ; Dissection ; Feedback ; Phases ; Python ; Robotic surgery ; Software ; Surgeons ; Surgical instruments ; Sutures ; Training simulators ; Virtual reality</subject><ispartof>Journal of robotic surgery, 2024-12, Vol.18 (1), p.47</ispartof><rights>Copyright Springer Nature B.V. Dec 2024</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><title>Surgical gestures can be used to assess surgical competence in robot-assisted surgery</title><title>Journal of robotic surgery</title><description>To collect validity evidence for the assessment of surgical competence through the classification of general surgical gestures for a simulated robot-assisted radical prostatectomy (RARP). We used 165 video recordings of novice and experienced RARP surgeons performing three parts of the RARP procedure on the RobotiX Mentor. We annotated the surgical tasks with different surgical gestures: dissection, hemostatic control, application of clips, needle handling, and suturing. The gestures were analyzed using idle time (periods with minimal instrument movements) and active time (whenever a surgical gesture was annotated). The distribution of surgical gestures was described using a one-dimensional heat map, snail tracks. All surgeons had a similar percentage of idle time but novices had longer phases of idle time (mean time: 21 vs. 15 s, p &lt; 0.001). Novices used a higher total number of surgical gestures (number of phases: 45 vs. 35, p &lt; 0.001) and each phase was longer compared with those of the experienced surgeons (mean time: 10 vs. 8 s, p &lt; 0.001). There was a different pattern of gestures between novices and experienced surgeons as seen by a different distribution of the phases. General surgical gestures can be used to assess surgical competence in simulated RARP and can be displayed as a visual tool to show how performance is improving. The established pass/fail level may be used to ensure the competence of the residents before proceeding with supervised real-life surgery. The next step is to investigate if the developed tool can optimize automated feedback during simulator training.</description><subject>Annotations</subject><subject>Bladder</subject><subject>Control equipment</subject><subject>Dissection</subject><subject>Feedback</subject><subject>Phases</subject><subject>Python</subject><subject>Robotic surgery</subject><subject>Software</subject><subject>Surgeons</subject><subject>Surgical instruments</subject><subject>Sutures</subject><subject>Training simulators</subject><subject>Virtual reality</subject><issn>1863-2483</issn><issn>1863-2491</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNo9zz9PwzAQBXALgUQpfAEmS8yGuziOnRFVQJEqMUDnynHOVaqSBJ8z8O0J4s90b_jpPZ0Q1wi3CGDvGNECKii0AnRgVXkiFugqrYqyxtP_7PS5uGA-ABhrNC7E9nVK-y74o9wT5ykRy-B72ZCcmFqZB-mZiVnynwvD-0iZ-kCy62UamiGr2XScZ_-tKH1eirPoj0xXv3cpto8Pb6u12rw8Pa_uN2pEY7MKrkGqQ-mLFjA2JjRAweoiOhdrE6N3CI0h7U0oYmVLcM5WodZtC4bIk16Km5_eMQ0f0_zA7jBMqZ8ndxpLVxsABP0FntxU1w</recordid><startdate>20241201</startdate><enddate>20241201</enddate><general>Springer Nature B.V</general><scope>K9.</scope></search><sort><creationdate>20241201</creationdate><title>Surgical gestures can be used to assess surgical competence in robot-assisted surgery</title></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-p157t-c8b1e9c4a2d01fb5cb0ec732f88f95ffa810b5e3a5c2f67408876c93dd05eeae3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Annotations</topic><topic>Bladder</topic><topic>Control equipment</topic><topic>Dissection</topic><topic>Feedback</topic><topic>Phases</topic><topic>Python</topic><topic>Robotic surgery</topic><topic>Software</topic><topic>Surgeons</topic><topic>Surgical instruments</topic><topic>Sutures</topic><topic>Training simulators</topic><topic>Virtual reality</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><collection>ProQuest Health &amp; Medical Complete (Alumni)</collection><jtitle>Journal of robotic surgery</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Surgical gestures can be used to assess surgical competence in robot-assisted surgery</atitle><jtitle>Journal of robotic surgery</jtitle><date>2024-12-01</date><risdate>2024</risdate><volume>18</volume><issue>1</issue><spage>47</spage><pages>47-</pages><issn>1863-2483</issn><eissn>1863-2491</eissn><abstract>To collect validity evidence for the assessment of surgical competence through the classification of general surgical gestures for a simulated robot-assisted radical prostatectomy (RARP). We used 165 video recordings of novice and experienced RARP surgeons performing three parts of the RARP procedure on the RobotiX Mentor. We annotated the surgical tasks with different surgical gestures: dissection, hemostatic control, application of clips, needle handling, and suturing. The gestures were analyzed using idle time (periods with minimal instrument movements) and active time (whenever a surgical gesture was annotated). The distribution of surgical gestures was described using a one-dimensional heat map, snail tracks. All surgeons had a similar percentage of idle time but novices had longer phases of idle time (mean time: 21 vs. 15 s, p &lt; 0.001). Novices used a higher total number of surgical gestures (number of phases: 45 vs. 35, p &lt; 0.001) and each phase was longer compared with those of the experienced surgeons (mean time: 10 vs. 8 s, p &lt; 0.001). There was a different pattern of gestures between novices and experienced surgeons as seen by a different distribution of the phases. General surgical gestures can be used to assess surgical competence in simulated RARP and can be displayed as a visual tool to show how performance is improving. The established pass/fail level may be used to ensure the competence of the residents before proceeding with supervised real-life surgery. The next step is to investigate if the developed tool can optimize automated feedback during simulator training.</abstract><cop>Heidelberg</cop><pub>Springer Nature B.V</pub><doi>10.1007/s11701-023-01807-4</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1863-2483
ispartof Journal of robotic surgery, 2024-12, Vol.18 (1), p.47
issn 1863-2483
1863-2491
language eng
recordid cdi_proquest_journals_3148950010
source Springer Nature
subjects Annotations
Bladder
Control equipment
Dissection
Feedback
Phases
Python
Robotic surgery
Software
Surgeons
Surgical instruments
Sutures
Training simulators
Virtual reality
title Surgical gestures can be used to assess surgical competence in robot-assisted surgery
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-08T03%3A30%3A59IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Surgical%20gestures%20can%20be%20used%20to%20assess%20surgical%20competence%20in%20robot-assisted%20surgery&rft.jtitle=Journal%20of%20robotic%20surgery&rft.date=2024-12-01&rft.volume=18&rft.issue=1&rft.spage=47&rft.pages=47-&rft.issn=1863-2483&rft.eissn=1863-2491&rft_id=info:doi/10.1007/s11701-023-01807-4&rft_dat=%3Cproquest%3E3148950010%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-p157t-c8b1e9c4a2d01fb5cb0ec732f88f95ffa810b5e3a5c2f67408876c93dd05eeae3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3148950010&rft_id=info:pmid/&rfr_iscdi=true