Preprint
Review

This version is not peer-reviewed.

Surgeon Training in the Era of Computer Enhanced Simulation Robotics and Emerging Technologies: A Narrative Review

A peer-reviewed article of this preprint also exists.

Submitted:

22 January 2025

Posted:

22 January 2025

You are already at the latest version

Abstract
Background: Teaching methodology has recently undergone significant evolution from traditional apprenticeship models as we adapt to ever increasing rates of technological advancement. Big data, artificial intelligence and machine learning are on the precipice of revolutionizing all aspects of surgical practice with far reaching implications. Robotic platforms will increase in autonomy as machine learning rapidly becomes more sophisticated and therefore training requirements will no longer slow innovation. Materials and Methods: Search of published studies discussing surgeon training and computer enhanced simulation robotics and emerging technologies using MEDLINE, PubMed, EMBASE, Scopus, CRANE, CINAHL and Web of Science was performed January 2024. Online resources associated with proprietary technologies related to the subject matter were also utilised. Results: Following review of 3209 articles, 91 published, relevant articles on aspects of robotics-based computer enhanced simulation, technologies and education were included. Publications ranged from RCTs, cohort studies, meta-analysis and systematic reviews. Eight medical technology-based website content were analysed and included in this review to ensure the most up-to-date information was analysed. Discussion: Surgeons should aim to be at the forefront of this revolution for the ultimate benefit of patients. Surgical exposure will no longer be due to incidental experiences. Rather, surgeons and trainees will have access to a complete database of simulated minimally invasive procedures and procedural simulation certification will likely become a requisite from graduation to live operating to maintain rigorous patient safety standards. This review provides a comprehensive outline of the current and future status of surgical training in the robotic and digital era.
Keywords: 
;  ;  ;  ;  ;  ;  ;  ;  ;  

1. Introduction

Surgical practice has undergone a dramatic evolution in the past 30 years. Rapid advances in computing and imaging technology have enabled the rise of minimally invasive surgery (MIS) as an alternative to traditional open procedures. MIS approaches have subsequently become the gold standard across many surgical disciplines owing to the perioperative benefits, improved cosmesis and functional recovery. Initially, laparoscopy with “straight stick” instruments surged in popularity through the 1990s and subsequently became the most common approach to MIS worldwide. Robotic surgery presents an alternative MIS approach that promises to eclipse laparoscopy and become the new standard of care in the near future. Purported benefits include visual enhancement with 3D stereoscopic vision and magnification up to 10x, tremor reduction, greatly enhanced dexterity through articulated instruments offering seven degrees of freedom, improved ergonomics and a resultant decrease in operator fatigue, both mental and physical.
The DaVinci platform from Intuitive Surgical has dominated the robotics market to date. Adoption has been slow but steady internationally, owing primarily to the associated costs when compared to laparoscopy. However, as patents expire on many robotic technologies, numerous competitors are now entering the robotic marketplace. This promises to drive down prices, improve availability thrust robotic surgery to the forefront of global MIS. It is therefore critical to examine how best to learn and teach in this exciting new surgical era.

2. Materials and Methods

We searched for publications on MEDLINE (EBSCO), PubMed, EMBASE, Scopus, CRANE Central registry of controlled trials, CINAHL (EBSCO) and Web of Science database using the keywords robotic surgery, surgical education AND simulation, virtual reality, 3D imaging, augmented reality, telesurgery, artificial intelligence, 3D printing, dual-console training OR telementoring. The search was limited to articles from 2002-2023 and performed in January 2024.

3. Results

Given the broad nature of the topic, 3209 potential articles were assessed and narrowed down to 91 publications that best addressed the main facets of this review.
As a rapidly evolving field driven largely by medical technology companies, 8 medical technology-based website content were analysed and added to this review to ensure the most up-to-date information was analysed.

4. Discussion:

4.1. Curriculum Development – The Apprentice Model

Traditionally, open surgical training represented an apprenticeship model. Trainees observed and assisted their surgical mentors in performing a procedure. Through mimicry and direct supervision, trainees gradually acquired operative skill and increasing independence. Competence was subjectively determined by surgical mentors throughout training, with little in the way of objective procedure-specific assessment of technical proficiency. However, modern surgical training poses significant challenges to this model. Restricted working hours, increased subspecialisation and a rapid expansion in technology and procedures dictate that this model is no longer able to guarantee training adequacy. Furthermore, as established surgeons adopt new technologies, it is imperative that gold standard outcomes are maintained through thorough training. These issues are compounded by a lack of standardisation in hospital accreditation processes and pressures from patients, peers, media, medical technology companies and healthcare administrators. Thus, surgical training models must adapt in order to maintain standards of care. Zorn and colleagues highlighted the extent of the problem in 2009, noting that an estimated 85% of radical prostatectomies in the US the preceding year were performed with robotic assistance despite the lack of any formalised accreditation or training process [1].
Currently, Intuitive Surgical provides a recommended framework for robotic accreditation, including online learning, in-service training, bedside assistance and primary operating [2]. Trainees subsequently provide a letter supporting their robotic competence from their supervisor, at which point Da Vinci will issue a certificate of system training. Completion of Da Vinci surgical simulator skills sessions are also recommended, though not required. However, one must note that this certificate is only intended to show competence in the use of the robot itself. The ability to perform a range of procedures requires far greater surgical experience with in-depth knowledge of anatomy, tissue handling and pathophysiology.
In the past decade there has been a deluge of studies reporting on the robotic learning curve and its potential utility in determining competence. A recent systematic review by Soomro et al. showed that the majority of literature was of poor quality with a large variety of outcome measures and methodologies rendering it difficult to make any meaningful conclusions for the purposes of implementing safe training in robotic MIS [3]. The challenge lies in determining at what point an individual has reached sufficient proficiency to practice independently without undue harm to patients and it seems that case numbers alone are insufficient to determine this.
Therefore, considerable effort has been placed into the development of structured training programmes for the safe implementation of robotic surgery. The Society of American Gastrointestinal and Endoscopic Surgeons (SAGES) and the Minimally Invasive Robotic Association (MIRA) first published a consensus statement regarding guidelines for training and credentialing in 2008 [4]. These guidelines emphasised the importance of a combination of didactic teaching, live case observation and hands-on experience, both of simulations and in vivo, though specific indicators of proficiency were not discussed. Lee and colleagues expanded on this concept, publishing best practice guidelines for robotic training and credentialing in 2011 [5]. They divided the training process into preclinical and clinical phases. Preclinical training involved didactic teaching and online learning modules followed by acquisition of basic robotic skills through dry labs and simulation. Trainees subsequently graduated to the clinical phase of teaching which progressed through case observation, live cases and expert discussion, bedside assisting, and finally operating as console surgeon with procedural breakdown into measurable steps of increasing complexity. Critically, the authors recognised the inadequacy of using case numbers to determine competence. They therefore recommended an objective, outcomes-based assessment of proficiency prior to credentialing. These recommendations form the backbone of modern robotic training curricula.
This has been further refined over time. In 2014, 14 multinational surgical societies developed the “Fundamentals of Robotic Surgery” (FRS) curriculum [6]. This programme is web-based and has the formidable aspiration of generalisability to any robotics platform and any surgical discipline. It includes 25 perioperative outcome measures and is divided into 3 sections: cognitive skills, psychomotor skills and team training and communication. Assessment is competency-based rather than time-based, with trainees required to reach benchmark “pass” values to complete the course. A multicentre RCT showed that significant improvements in task completion time and error rate were made following FRS training. These were comparable to a control group consisting of surgeons who had completed their own local institution specific training, confirming that the FRS was at least comparable with other common forms of training [7].
Alternative validated training curricula include the Robotic Training network (RTN) and the Fundamental Skills of Robotic Surgery (FSRS). However, these have their own limitations including limited international availability and, for FRS, requirement for the specific RoSS surgical simulator [8,9].
While online curricula are more accessible and offer greater flexibility to participants, a perceived benefit of on-site training is the ability of supervisors to give advice on correcting technique and thus improve efficiency. However, this supposition has been challenged by a study comparing expert preceptorship with an educational video on skills acquisition [10]. Both groups showed significant improvements with training and no significant differences between groups. Thus, although direct comparisons between each training curricula have not been performed, this is unlikely to be of clinical significance. The most important factor is the completion of one form of validated, proficiency-based skills curriculum prior to embarking on further robotics training. The ‘best’ programme for any given surgeon is likely that which they have ready access to.
The robotic section of the EAU, known as ERUS, ultimately published the first standardised 12-week robotic training curriculum in 2015 [11]. The curriculum included an initial e-learning module on principles of robotics followed by operative observation and assisting, simulation-based training incorporating VR simulation, dry and wet lab activities and supervised, modular training with a progression through increasingly complex steps as proficiency increased. All participants showed significant improvement in dVSS simulator performance through the course of the training program. Face, content and construct validity were all confirmed. However, only 80% of participants were deemed competent to independently perform robot assisted radical prostatectomy (RARP) by their mentors at the conclusion of the programme. Independent assessors similarly scored 80% as safe and competent. Further, expert mentors felt that only 30% were capable of safely and independently completing a complex case. This once again highlights the inherent variability in learning curve and the importance of using objective performance-based outcomes over case numbers to determine competency.
The following year, the Clinical Robotic Surgery Association (CRSA) published specific recommendations on structured training in colorectal surgery [12]. These guidelines focused on a stepwise objective assessment of competency. Basic training is divided into sequential stages and each must be passed objectively before being allowed to progress to live operating. Each operation is broken down into steps and trainees are rated as independent, requiring prompting, or unable to perform for each step. Procedures are taught in order of increasing complexity until the trainee is ultimately deemed independent and credentialed.
Numerous surgical societies have subsequently followed suit and published their own guidelines for safe introduction of robotic surgery. However, the recommendations of the CRSA arguably represent the greatest paradigm shift toward objective, procedure specific, proficiency-based accreditation to date.

4.2. Novel Training Modalities

Having now defined the aims of robotic training, one must examine the most effective methods for achieving proficiency while minimising risk of harm to patients. Fortunately, robotic surgery is uniquely suited to innovative new training models.

4.2.1. Virtual Reality

While VR is not a new concept, having first arisen in the 1980s, imaging technology and processing power have only recently advanced sufficiently to allow virtual reality to enter the mainstream. Driven by video game development, there are now numerous open surgical simulators based on the oculus and HTC platforms. However, much of the skill in open surgery centres around tissue handling and manipulation that is well beyond the capabilities of current platforms, severely limiting their utility in open surgical training. Similarly, in performing laparoscopic surgery there remains a significant component of haptic feedback and off-screen movement that influences surgical technique and is difficult to simulate. By comparison, modern robotic surgery lacks haptic feedback. While this is often described as a limitation in current robotic surgery, for the purposes of simulation and training, these become strengths. Technique is entirely guided by on-screen visual cues and, therefore, simulation has the potential to most accurately reproduce the robotic operative environment.
First generation VR simulators include the DaVinci Skills Simulator (dVSS), Mimic DaVinci trainer (dVT) FlexVR, the ProMIS simulator, the Simsurgery Educational Platform (SEP) and the Robotic Surgical Simulator (RoSS) [13]. These platforms focus primarily on basic skill acquisition through skills drills, though a limited degree of operative simulation is offered. Second generation platforms include the RobotiX Mentor (RM), the RoSS II/II lite, and SimNow by Da Vinci [13,14,15]. These next generation platforms greatly expand the utility of simulation. Photorealism is vastly improved and it is now possible to perform complete operative procedures in the simulated environment, either in guided fashion or free-hand with in-depth assessment scores on completion. RoSS also offers an alternative unique module known as Hands-on Surgical Training (HoST). Here, surgical videos from real robotic procedures are displayed on screen. Through haptic feedback of the controls, the trainee is able to experience the exact hand movements of the operating surgeon in synchrony with the surgical footage [14].
In order to be of value, simulation must lead to a demonstrable improvement in surgical skills. This may be described across several validity domains (Table 1).
The efficacy of virtual reality simulation in robotic skills training is well established. Each of the aforementioned simulators has demonstrated face, content and construct validity in randomised controlled trials across individuals with varying levels of surgical ability [13,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32].
In considering the superiority of one simulator over another, few head-to-head comparisons have been performed. Hertz and colleagues recently compared face and content validity between dVT, dVSS and RM. The dVSS was found to be superior to dVT, while no other significant differences were identified between platforms [33]. This potentially reflects the aging nature of dVT, now in use since 2007, and the use of the master console with dVSS, improving simulation realism. A further comparison performed by Tanaka et al. compared face, content and construct validity between dVT, DVSS and RoSS [34]. dVSS and dVT both significantly outperformed RoSS in face and content validity. Additionally, dVSS and dVT both showed good construct validity, while RoSS was unable to distinguish between novice and expert users. Further support for the dVSS simulator was published by Ahmad and colleagues who showed that fellows trained on the dVT scored significantly lower on both pre and post-test assessment, had a lower average curriculum score and spent more time completing each assessment.
These results appear to translate to meaningful skill acquisition. Hoogenes et al. compared the performance of junior and senior trainees in performing vesicourethral anastomoses following completion of an identical simulator-training curriculum on dVT or dVSS [35]. Junior trainees showed significantly better performance following dVSS training compared with dVT training, while senior trainees who had greater previous robotic experience showed no difference in performance between programmes. Therefore, it seems that dVSS may represent a superior training tool when compared to dVT, particularly in robotic novices. Unfortunately, there is no data comparing the next generation simulators described above with which to make a similarly informed decision.
Of note, the majority of aforementioned studies focus on face, content and construct validity in performing basic skills such as object manipulation and knot tying. While such tasks are attractive for research purposes due to their simplistic nature and ease of outcome comparison between groups, they do not necessarily reflect the transfer of skills to the operating room (concurrent and predictive validity) which is of critical importance in determining the value of simulation in training. A recent meta-analysis by Schmidt et al. has examined the role of simulator skills transfer to live operating, identifying 8 studies for review [36]. They concluded that VR skills acquisition is transferrable to the OR and that simulator performance on DVT and dVSS demonstrates concurrent validity, although their findings were limited by the small number and heterogeneous nature of included studies.
Furthermore, logic dictates that realistic operative simulation is likely to dramatically increase the utility of virtual reality platforms. This technology is already available and will only continue to advance in coming years as processing power and imaging technology allow true photorealistic re-creation. To date, this has only been assessed in one randomised trial by Raison and colleagues [37]. Novice participants received no training (n=9), basic simulation training (n=13) or procedural simulation training (n=13). Subsequently, each participant performed robotic radical prostatectomy on a cadaver and was assessed by blinded expert reviewers. Completion of either training model resulted in significantly higher scores than no training, demonstrating concurrent validity. Furthermore, procedural training resulted in significantly higher scores than basic training, demonstrating superiority as a training tool.

4.2.2. Animal/Cadaver Models

Of course, there are alternatives to virtual reality simulation that must also be considered in determining the most appropriate teaching tool for robotic surgery. Animal and cadaveric models are well recognised in surgical training [38,39]. These models represent the most realistic form of surgical simulation in terms of anatomical authenticity and tissue handling properties. This is particularly true of living tissue handling in animal models, though at the expense of anatomical variance. To bridge this divide, cadaveric models may be enhanced with re-establishment of simulated perfusion, as outlined in a systematic review by Bellier et al. [40]. However, despite their longstanding usage and validated nature, direct comparisons between animal, cadaveric and VR simulation in MIS are severely limited. Just two papers comparing laparoscopic VR with cadaveric simulation have been published [41,42]. Both focused strongly on subjective participant satisfaction, finding a preference for cadaveric training. VR simulation found greater acceptance for basic task training with junior trainees, and was found to be less complex than cadaveric dissection in procedural assessment.
Specific to robotics, literature is similarly limited. Bertolo and colleagues conducted a single robotic training session on fresh-frozen human cadavers for surgical residents with limited robotic experience [43]. They found a high degree of satisfaction amongst participants, who showed subjective and objective skill improvement following the session. Furthermore, participants rated the activity as superior to both VR and porcine training sessions.
Thus, very limited evidence may suggest an advantage to cadaveric simulation, particularly for more complex procedural tasks. However, as noted above, VR technology has advanced dramatically in the past decade and the relevance of the above results when compared to more modern VR simulators is unknown.

4.2.3. 3-D Printing

Another novel alternative has been the construction of artificial models through 3D printing and polymer moulding. Several studies have recently assessed the value of 3D printed models in robotic training. These have demonstrated the face, content and construct validity of such models [44,45,46].
As with VR, modelling techniques have greatly advanced in recent years. Significant focus has been placed on the realism of tissue reproduction, paving the way for valuable, high fidelity surgical simulation. Models can now conduct diathermy in a realistic manner and can ‘bleed’ due to artificial perfusion with solutions of similar viscosity to blood. In addition to creating realistic models for training purposes, patient specific models have been developed from CT reconstructions, allowing 3-dimensional tumour assessment to assist in operative planning, and even to rehearse procedures. This was first demonstrated by von Rundstedt et al. who showed that operative times and tumour characteristics were remarkably similar when comparing 3D printed complex renal tumours to the in vivo specimen [47].
Further work was performed by Ghazi and colleagues in establishing the validity of such high-fidelity organ models. Their models allow ultra-realistic simulation of entire surgical procedures within a replica abdomen or pelvis [48]. Reviewers consistently rated these models as superior to porcine or cadaveric models and perfusion was considered to be a particularly important element. Experts significantly outperformed novices in performing RALPN over a variety of validated scoring systems, confirming the construct validity.
Each of these papers serve as an impressive demonstration of how far 3D printed simulation has progressed and its great potential for teaching and operative planning. However, animal, cadaveric and 3D models are not without other practical limitations and VR simulators demonstrate several benefits here. Training in animals and cadavers is costly, there is limited availability and there are many ethical considerations. A complete Da Vinci robot must be available in a wet laboratory environment and instruments must also be made available at considerable cost. As a result, access to training with these models is greatly restricted, and is primarily only available in dedicated teaching institutions. In the case of 3D printed models, the requirement for a complete robot and training instruments remains a limitation, though instrument sterility and infection control issues are eliminated. Models can only be used once, consume considerable time and resources in preparation and generate significant waste. The more complex and realistic the model, the greater the effort required in manufacture. In fact, Witthaus et al. noted that each model took a skilled biomedical engineer approximately 5.5hrs to construct. Therefore, commercialisation is likely required in order to develop mainstream utility. Unfortunately, this often comes at considerable additional cost, particularly in the case of medical technology. By comparison, although initial purchase costs are high, VR simulators may subsequently be made available 24 hours a day, 7 days a week. They can be reused an unlimited number of times with virtually no operational costs. The main limitation is in the fidelity of procedural simulation. As VR comes ever closer to achieving photorealism, it is highly likely to declare itself as the predominant training modality.

4.2.4. Dual-Console Training

Following graduation from simulation, trainees must transition to in vivo console operation. During the early transition phase, the presence of an on-site preceptor is considered critical. This individual should be an experienced robotic surgeon, able to offer guidance and supervision to maximise patient safety. With the initial Da Vinci platform, this was a cumbersome process, as described by doctors Crawford and Dwyer [49]. In addition to the provision of verbal advice, the preceptor would often act as bedside assistant, providing the ability to assist or point laparoscopically. However, the preceptor was commonly required to “break scrub” and temporarily act as console surgeon, a time consuming and frustrating process. Fortunately, progressive technological advancement has significantly enhanced the means of interaction between supervisor and mentee since this time.
Hanly and colleagues first described the concept of dual console platforms in 2006, linking two surgeon consoles together via a special-purpose connection [50]. This enabled collaborative teaching through the fluid exchange of instrument control between the trainer and trainee, in addition to providing a form of haptic feedback, allowing both surgeons to simultaneously feel the movement of the instruments. Thus, the trainee could be guided in the performance of precise tasks such as intracorporeal suturing. This teaching model was much more akin to open teaching techniques than those employed at laparoscopy and was subjectively perceived as highly advantageous. It offered vast improvements in the ability to teach robotic surgery efficiently while reducing some of the associated anxiety involved in teaching MIS. Several studies have subsequently reported the safety and training benefits of the dual-console model [51,52,53].

4.2.5. Augmented Reality

The second major advance in robotic teaching was the introduction of telestration [49]. This likely represented the first practical application of augmented reality (AR) in robotic surgery. Telestration enabled the preceptor to direct the trainee by marking on the laparoscopic image on a touch-screen display. The markings were then reproduced on the surgeon console display. Alternatively, in the dual console model, the preceptor was given the ability to control a virtual pointer on-screen in real time to serve the same purpose, but with the added benefit of 3D. Therefore, the trainee could be provided with visual guidance without need to leave the console, greatly improving communication and efficiency.
Jarc et al. further advanced telestration techniques with the introduction of “ghost tools” [54,55]. Ghost tools offered 3D telestration abilities to proctors through use of a 3D pointer, 3D hands with the ability to point or simulate grasping, and 3D instruments that could be manipulated in similar fashion to actual operating instruments. Both proctors and trainees demonstrated a preference for 3D hands and 3D instruments over traditional 2D telestration, finding them to be more effective demonstration tools despite an increase in complexity of use for the proctors. Further objective research demonstrated proctors made good use of the enhanced manipulation abilities on offer [55].
Though not specific to teaching, there are several other current and future augmented reality technologies that offer benefit to both trainee and mentor alike. First amongst these is image enhancement through the use of immunofluorescence. The technology, known as “Firefly”, involves the administration of indocyanine green (ICG) followed by filtering the endoscope image for near infrared light wavelengths. The technique can be used to highlight underlying critical vasculature, biliary structures and ureters, to delineate hepatic tumours and to assess tissue perfusion during anastomosis. Since its introduction in 2011 its use has become commonplace in many surgical procedures [56].
Recently, Activ Surgical have released an augmented reality endoscope attachment that allows real-time assessment of tissue perfusion without the need to inject dye [57]. Though this is currently designed for laparoscopic surgery, there is little doubt that robotic offerings will be quick to follow and further refinement may see this become the new standard of care.
3D reconstructions of staging imaging can prove particularly helpful in providing a surgeon or trainee with a greater understanding of the patient’s anatomy preoperatively, enhancing surgical planning and surgical safety. Currently, the TilePro function of the Da Vinci platform allows the operator to display and manipulate this imaging intraoperatively alongside the endoscopic display to guide dissection. Intraoperative ultrasound may also be displayed on TilePro, allowing sonographically detectible lesions in solid organs to be marked out, allowing maximal preservation of critical neurovasculature while reducing the risk of an involved resection margin [58]. The next phase of augmented reality involves overlaying this information onto the surgical field in real time for enhanced identification of critical anatomy and improved efficiency. Proof of concept has already been successfully demonstrated in urologic and hepatic surgeries, with recent work suggesting improved accuracy in resection margins for hepatic tumours compared to the current gold standard of intraoperative ultrasound [59,60,61,62,63,64]. The next major hurdle lies in adjusting for real-time tissue deformation and manipulation intraoperatively.
In the near future, augmented reality will offer even greater value through the use of artificial intelligence (AI) and machine learning. Machine learning may be supervised, whereby a human input labelled data into a programme to teach it to differentiate between structures, or unsupervised, where unlabelled data is fed into the algorithm, which then attempts to identify the abnormality. In re-enforcement learning, the AI is then set a task and gains further data points based on its successes or failures [65]. Essentially, these methods allow AI programs to ‘learn’ to analyse data and identify the desired abnormality with increasing accuracy. In minimally invasive surgery, this technology can be applied to display an intraoperative on-screen visual representation of areas of safe dissection and “no go” zones containing underlying critical anatomy in order to improve patient safety. Such algorithms have already been successfully applied to laparoscopic cholecystectomy to provide on-screen guidance around safe dissection of Calot’s triangle and avoidance of portal structures with high levels of efficacy [66,67]. While the application of AI to surgical training is still in its infancy, machine learning holds the ability to advance at a rate far outstripping human learning and this will no doubt become an extremely powerful surgical tool.

4.2.6. Telementoring

Since the advent of robotic surgery there has been great interest in the potential applications of telesurgery in revolutionising healthcare. The master-slave nature of the Da Vinci platform is perfectly suited to the performance of telesurgery, whereby the operating surgeon controls the robotic instruments from a remote location. Potential applications included the delivery of healthcare to poorly serviced areas or adverse environments (e.g., warzones), to offer highly subspecialised services from a central “institute of excellence” without requiring the patient or surgeon to travel and, in the age of Covid-19, to reduce the risk of surgeon exposure to transmissible diseases [68]. The first telerobotic cholecystectomy was performed in 2001 [69]. However, adoption and practical application has been limited due to concerns over network stability, latency times, medicolegal issues, the risk of cybersecurity threats and establishment costs [68]. Optimal latencies are considered to be below 200-300ms, while latencies greater than 700-1500ms make surgical performance challenging and likely unsafe [70,71,72,73].
Fortunately, these limitations are less restrictive in the case of telementoring, and this technology has been utilised with good effect in robotic surgical training. A systematic review by Bilgic et al. has confirmed the safety and efficacy of telementoring in surgery [74]. The authors examined papers comparing on-site mentoring with telementoring and found 11 studies of 453 cases that were suitable for inclusion. No differences in perioperative complication rates were encountered in any study. 90% reported comparable operating times between groups, with 1 study showing a longer operating time due to telementoring. Technical difficulties were encountered in 3% of telementored cases. Subjective analysis of trainee satisfaction revealed no difference between on-site mentoring and telementoring, while objective improvements in operating times across the learning curve were also comparable between groups.
Subsequent publications by Papalois et al. and Artsen et al. have placed further support behind telementoring. Papalois and colleagues developed a surgical curriculum delivered in mixed reality through use of Microsoft’s Hololens [75]. The curriculum focused on surgical decision-making, operative anatomy and expert “tips and tricks”. 93% of students and 100% of tutors felt that virtual mentorship was of use in future surgical training, while 73% agreed/strongly agreed that their understanding of anatomy and decision-making rationale was improved by the module. Artsen et al. compared a series of teleproctored robotic gynaecologic cases with historical controls performed with in-person proctoring. They found high satisfaction rates amongst surgeons and no change in perioperative complication rates [76].
Several new commercial platforms have recently become available that significantly improve on existing telementoring software. These include Orpheus medical (recently acquired by Intuitive Surgical,) Proximie, and Reacts (recently acquired by Philips) [77,78,79]. Each of these platforms allows real-time collaboration and consultation for the purpose of live telementoring and utilises augmented reality overlays for advanced telestration and annotation.
Telementoring therefore offers great potential. As the technology matures, it may be utilised at any ability level, from residents receiving didactic teaching and supervised simulation training to a specialist surgeon consulting peers regarding a particular intraoperative quandary. This will aid in the standardisation of gold-standard surgical techniques globally by vastly increasing access to world leaders in any given subspecialty. This is particularly important in an age where new procedures and technologies are in constant development. In order to maintain currency, there is a strong requirement for the establishment of global surgical collaborative networks.

4.2.7. Surgical Videos

In the past two decades, online data transfer rates have accelerated dramatically. This has greatly improved the ability to share high-definition surgical footage globally. This footage offers a valuable learning tool, particularly in the case of robotic surgery. As discussed previously, due to haptic feedback limitations of current robotic platforms, the surgeon is entirely guided by on-screen visual cues. Thus, observers have access to identical sensory inputs afforded to the operating surgeon. This may be capitalised upon for learning purposes.
Video-based learning has become an extremely common method of surgical learning. In a recent survey of residents and surgeons, Mota et al. found that 98.6% of respondents had made use of videos in preparation for surgery [80]. Furthermore, 57% noted that surgical video was their preferred method of surgical preparation. This was particularly true of younger, less experienced respondents.
Video material may be used at different stages of teaching. Demonstration videos can be effective in teaching simulation skills for robotic surgery, as shown by Shim et al. [10]. Educational video was shown to be as effective as on-site mentoring in mastering robotic vesicourethral anastomosis, while both learning techniques were superior to self-directed learning. Video based learning has also shown superiority when compared to hands-on practical training in the case of laparoscopic cholecystectomy. Pape-Koehler and colleagues performed a randomised controlled trial comparing a multimedia, video-based learning module with practical training or no training [81]. Participants who underwent multimedia-based training showed a significant improvement in surgical performance compared to both their practically trained and untrained colleagues.
In the near future, surgical video may be further enhanced through immersive footage that incorporates haptic feedback. Pandya and colleagues developed a novel recording system that synchronized robotic arm and surgeon-console interactions with operative footage [82]. The material could then be replayed to an observer at the surgeon console, allowing them to feel the movements of the operating surgeon’s hands to gain a deeper appreciation of correct operative technique and shorten the learning curve associated with complex procedures. This concept has already been applied to simulation training in the form of the HoST platform on the RoSS virtual simulator described earlier [14].
Huynh and colleagues offer a word of caution in the application of surgical videos [83]. The authors reviewed the most viewed 50 YouTube videos relating to MIS inguinal hernia repair and rated their surgical performance in comparison to the “9 commandments” of safe technique as defined by Daes and Felix [84]. Only 16% of videos demonstrated all 9 commandments, with significant differences between laparoscopic and robotic approaches. Furthermore, 46% of videos were considered to display unsafe techniques through dangerous mesh fixation, risks to critical structures or inappropriate tissue handling.
To address the variable quality of online material, an international multidisciplinary consensus group have recently published guidelines on the appropriate reporting of educational videos [85]. These guidelines cover 36 recommendations including requirement for video introduction and information about the authors, case presentation and staging imaging, robotic setup, procedural demonstration +/- stepwise teaching and telestration of relevant anatomy, review of postoperative outcomes, and confirmation of high-quality footage. This standardised approach to reporting aims to improve the general quality of online material as a teaching resource. Knowledge of these guidelines will also help a prospective student to select appropriate videos from which to learn.
Several comprehensive clinical media platforms are now available that facilitate data sharing in the surgical community. Examples include Orpheus, Proximie and Touch Surgery Enterprise. These platforms simplify the process of video storage, editing and retrieval, while offering rapid, secure, de-identified sharing of material between colleagues. Furthermore, they each possess AI algorithms for the automatic segmentation and labelling of surgical footage [77,78,86]. This further improves the utility of videos through rapid access to relevant sections of a procedure.

4.3. Technical Assessment and Tracked Metrics

As previously described, most modern robotic surgery curricula focus on objective assessment of proficiency prior to graduation. This is critical to patient safety. Technical skills assessment from operative video has been directly associated with perioperative complications, morbidity and mortality [87,88]. Assessments in robotic surgery publications have most commonly been delivered through validated scoring systems by mentors or independent reviewers. Common examples include GEARS, R-OSATS and PACE scores [89,90,91]. However, this process is time intensive and relies on considerable goodwill on behalf of assessors. Therefore, several potential alternatives have been explored. Crowdsourcing is one such example. This technique utilises the ready availability of large numbers of individuals in an online forum. Non-surgical crowdworkers are briefly trained in video evaluation through an online module and then score performances using a rating scale such as GEARS. Crowdsourcing results in much more rapid responses than the use of expert reviewers at minimal cost and studies have shown good correlation between scores by expert reviewers and crowdworkers, even in the assessment of complex procedures such as prostatectomy [92,93,94,95].
Automated performance metrics (APMs) represent an increasingly valuable means of assessment. APMs are a set of data points relating to various aspects of an operation that are routinely collected throughout a robotic procedure. They are readily available with minimal effort or cost. For the purposes of training, each surgical simulation platform automatically assesses user data and provides performance scores following the completion of a task, as compared to expert benchmarks. In order to be of value, these metrics must be shown to hold practical utility. Several studies on the validity of simulation have shown that automated metrics correlate highly with expert GEARS scores in simulation exercises. Importantly, tracked metrics were also shown to correlate well with subsequent intraoperative performance assessment [96].
Chen et al. have also proven the ability of intraoperative APMs to differentiate between expert and novice robotic surgeons in the performance of a vesicourethral anastomosis during RARP [97]. By combining the APM data with clinicopathologic characteristics in a deep learning algorithm, they were subsequently able to accurately predict postoperative continence rates [98]. This supports the value of APMs and deep learning algorithms in assessing surgical quality. Multiple subsequent studies of deep learning algorithms have unanimously proven to be highly accurate in predicting surgical skill level when compared to structured assessments by expert reviewers [99].
Currently, this remains a new and novel technology and how best to incorporate it into proficiency assessment has yet to be clearly defined. Nevertheless, APMs and deep learning algorithms are an area of burgeoning research. Due to the rapid availability of results, cost-efficacy and accuracy, it is highly likely that these will ultimately become the primary method of proficiency-based assessment for surgical trainees and will have a significant role in hospital accreditation and credentialing processes.

4.4. Artificial Intelligence, Machine Learning and Big Data

Several major device manufacturers have identified the potential of tracked metrics and machine learning. Previously discussed media platforms such as Orpheus, Proximie and Touch Surgery currently offer their users procedural analysis based on APMs. In their current states, the information produced is useful for benchmarking and identifying areas for individual improvement. However, their true potential lies in future applications. The routine recording and upload of cases will rapidly lead to vast troves of data, knows as “big data”. In combination with electronic medical records, these can be analysed by deep learning algorithms such as convolutional neural networks to identify associations between surgical technique and outcomes that cannot be appreciated by traditional research methods or simple observation. As a result, trainees and surgeons will be provided with a constant source of feedback on subtle areas for self-improvement, to the benefit of their patients.
Optimal patient management will also become far more standardised globally based on best evidence from hundreds of thousands of cases, overcoming many of the current limitations of surgical research. While big data is already utilised in the production of treatment algorithms such as CeDAR or NELA, these only utilise a handful of data points to guide clinical decision making. Deep learning algorithms will have the ability to advise on the most appropriate treatment for a patient based on innumerable individualised data points in a far more personalised manner and will have a profound effect on the way we practice surgery.
An additional significant benefit arises in cost and time savings. Traditional research methods such as randomised controlled trials frequently take years to produce clinically relevant conclusions at great cost and often with limited generalisability. The field of surgery is rapidly expanding with procedures undergoing constant evolution. Traditional clinical trial models can no longer keep up with this rate of advancement. By comparison, machine learning possesses the ability to produce clinically relevant outcome data at a rate that keeps pace with surgical advancement, guiding clinical practice in a current and meaningful manner.
The storage of big data does raise ethical and medicolegal issues around the potential identification of the underperforming surgeon. However, it is currently of critical importance that the information produced is utilised for self-improvement purposes alone, rather than in any disciplinary manner. Deep learning algorithms are entirely dependent on the purity of input data in order to produce accurate results. The fear that participation may result in negative repercussions may otherwise lead to data contamination through inaccurate data input or data omission. This would severely compromise the reliability of results and potentially even lead to patient harm. As a self–improvement tool, all surgeons are afforded the opportunity to improve their own practice, thus providing widespread benefits to patients.

5. Conclusion:

This thesis has provided a comprehensive outline of the current and future status of surgical training in the robotic and digital era. Teaching methodology has recently undergone significant evolution from traditional apprenticeship models as we look to adapt to ever increasing rates of technological advancement. Big data, AI and machine learning are on the precipice of revolutionizing all aspects of surgical practice with far reaching implications.
The future procedural surgical training model will likely commence with recorded didactic teaching and demonstrations by expert surgeons. International collaboration and deep learning will provide a better appreciation of the gold standard approach to be taught. Trainees will complete basic simulation training followed by procedural simulation with constant AI guidance around safe planes, structures to avoid, and overall detailed grading of the quality of the procedure at conclusion. Procedures will be repeated as many times as necessary until proficiency is achieved. In-person tutors will not be required, saving significantly on costs and resources. Additionally, innovators will be able to trial new approaches in the simulated surgical environment without risk to patients. Surgical exposure will no longer be “pot luck”, with experience dependent on the procedures coming in the door. Rather, surgeons and trainees will have access to a complete database of simulated minimally invasive procedures and procedural simulation certification will likely become a requisite for graduation to live operating in order to maintain rigorous patient safety standards. This will be important not just for the new surgical trainee, but also for the established surgeon adopting new techniques or technologies. The profession is now advancing at such a rate that constant re-training will be required throughout each of our careers.
Previously, adopting new techniques into practice has been hampered by the considerable associated learning curve. Advances in AI will facilitate this in future. Learning curves will be shortened through the increased utilisation of AI in performing guided surgical procedures. Robotic platforms will increase in autonomy as machine learning rapidly becomes more sophisticated and therefore training requirements will no longer slow innovation.
Surgeons should aim to be at the forefront of this revolution for the ultimate benefit of our patients. In many countries, public access to robotic simulators and operating consoles remains limited, creating a training bottleneck. This must be overcome through collaboration between surgical training bodies and device manufacturers. Governance measures should be implemented for the safe introduction of this exciting technology.

Funding sources

None

Acknowledgements

None

Conflict of interests

None

Disclosure

None

References

  1. Zorn KC, Gautam G, Shalhav AL, et al. Training, credentialing, proctoring and medicolegal risks of robotic urological surgery: recommendations of the society of urologic robotic surgeons. J Urol. Sep 2009;182(3):1126-32. [CrossRef]
  2. Da Vinci residency/fellowship training. Da Vinci residency/fellowship training. 2020. Accessed 12/01/2022. davincisurgerycommunity.com.
  3. Soomro NA HD, Porteous AJ et al. Systematic review of learning curves in robot-assisted surgery. BJS Open. Feb 2020;4(1):27-44. [CrossRef]
  4. Herron DM MM, SAGES-MIRA robotic surgery consensus group. A consensus document on robotic surgery. Surg Endosc. Feb 2008;22(2):313-25. [CrossRef]
  5. Lee JY MP, Sundaram CP, McDougall EM. Best practices for robotic surgery training and credentialing. J Urol. Apr 2011;185(4):1191-1197. [CrossRef]
  6. Smith R PV, Satava R. Fundamentals of robotic surgery: a course of basic robotic surgery skills based upon a 14-society consensus template of outcomes measures and curriculum development. Int J Med Robot. Sep 2014;10(3):379-384. [CrossRef]
  7. Satava RM SD, Levy DS et al. Proving the effectiveness of the Fundamentals of Robotic Surgery (FRS) skills curriculum: A single-blinded, multispecialty, multi-institutional randomised control trial. Ann Surg. Aug 2020;272(2):384-392. [CrossRef]
  8. Stegemann AP AK, Syed JR et al. Fundamental skills of robotic surgery: a multi-institutional randomized controlled trial for validation of a simulation-based curriculum. Urology. Apr 2013;81(4):767-74. [CrossRef]
  9. Siddiqui N GM, Geller E et al. Validity and reliability of the robotic objective structured assessment of technical skills. Obstet Gynecol. Jun 2014;123(6):1193-99. [CrossRef]
  10. Shim JS KJ, Pyun JH et al. Comparison of effective teaching methods to achieve skill acquisition using a robotic virtual reality simulator: Expert proctoring versus an educational video versus independent training. Medicine (Baltimore). Dec 2018;97(51):e12569. [CrossRef]
  11. Volpe A AK, Dasgupta P et al. Pilot validation study of the European Association of Urology robotic training curriculum. Eur Urol. Aug 2015;68(2):292-299. [CrossRef]
  12. Petz W SG, Choi G et al. Structured training and competence assessment in colorectal robotic surgery. Results of a consensus experts round table. Int J Med Robot. Dec 2016;12(4):634-641. [CrossRef]
  13. MacCraith E FJ, Davis NF. Robotic simulation training for urological trainees: a comprehensive review on cost, merits and challenges. . Journal of robotic surgery. June 2019;13(3):371-377. [CrossRef]
  14. RoSS II. Accessed 02/01/23. http://simulatedsurgicals.com/projects/ross/.
  15. Da Vinci SimNow. Accessed 11/01/22. https://www.intuitive.com/en-us/products-and-services/da-vinci/education/simnow.
  16. Hung AJ ZP, Patil MB. Face, content and construct validity of a novel robotic surgery simulator. J Urol. 2011;186(3):1019-1024. [CrossRef]
  17. Kelly DC MA, Kundavaram CR. Face, content and construct validity of the da Vinci Skills Simulator. J Urol. 2012;79(5):1068-1072. [CrossRef]
  18. Liss MA AC, Quach S. Validation, correlation and comparison of the da Vinci Trainer and the da Vinci surgical skills simulator using the Mimic softward for urologic robotic surgical education. J Endourol. 2012;26(12):1629-1634. [CrossRef]
  19. Lendvay CP SR, Peters C. Initial validation of a virtual-reality robotic simulator. J Robot Surg. 2008;2(3):145-149. [CrossRef]
  20. Kenney P WM, Gould J, Libertino J, Moinzadeh A. Face, content and construct validity of dV-trainer: a novel virtual reality simulator for robotic surgery. J Urol. 2009;73(6):1288-1292. [CrossRef]
  21. Sethi AS PW, Mohammadi Y. Validation of a novel virtual reality simulator. J Endourol. 2009;23(3):503-508. [CrossRef]
  22. Perrenot C PM, Tran N, Jehl JP, Felblinger J, Bresler L, Hubert J. The virtual reality simulator dV-Trainer is a valid assessment tool for robotic surgical skills. J Surg Endosc. 2012;26(9):2587-2593. [CrossRef]
  23. Korets R MA, Graversen JA, Gupta M, Benson MC, Cooper KL, Landman J, Badani KK. Validating the use of teh mimic dV-trainer for robotic surgery skill acquistion among urology residents. J Urol. 2011;78(6):1326-1330. [CrossRef]
  24. Lee JY MP, Kerbl DC. Validation study of a virtual reality robotic simulator - role as an assessment tool? J Urol. 2012;187(3):998-1002. [CrossRef]
  25. Schreuder HWR PJ, Wolswijk RGH, Ihse I, Schijven MP, Verheijen RHM. Validation of a novel virtual reality simulator for robotic surgery. Sci World J. 2014;2014:10. [CrossRef]
  26. Seixas-Mikelus SA KT, Srimathveeravalli G. Face validation of a novel robotic surgical simulator. J Urol. 2010;76(2):357-360. [CrossRef]
  27. Chowriappa AJ YS, Raza SJ et al. Development and validation of a composite scoring system for robot-assisted surgical training - the Robotic Skills Assessment Score. J Surg Res. Dec 2013;185(2):561-569. [CrossRef]
  28. Seixas-Mikelus SA SA, Kesvadas T et al. Content validation of a novel robotic surgical simulator. BJU Int. Apr 2011;107(7):1130-5. [CrossRef]
  29. Hung AJ PM, Zender P et al. Concurrent and predictive validation of a novel robotic surgery simulator: a prospective randomized study. J Urol. 2012;187(2):630-637. [CrossRef]
  30. Colaco M BA, Su D et al. Initial experiences with RoSS surgical simulator in residency training: a validity and model analysis. J Robot Surg. 2012;7(1):71-75. [CrossRef]
  31. Finnegan KT MA, Staff I et al. da Vinci skills simulator construct validation study: correlation of prior robotic experience with overall score and time score simulator performance. J Urol. 2012;80(2):330-335. [CrossRef]
  32. Lerner M AM, Peine W, Sundaram C. Does training on a virtual reality robotic simulator improve performamnce on the da Vinci surgical system? J Endourol. 2010;24(3):467-472. [CrossRef]
  33. Hertz AM GE, Vaccaro CM, Brand TC. Head-to-head comparison of three virtual-reality robotic surgery simulators. JSLS. Jan 2018;22(1):e2017.00081. [CrossRef]
  34. Tanaka A GC, Simpson K, Perez M, Truong M, Smith R. Robotic surgery simulation validity and usability comparative analysis. Surgical endoscopy. 2016;30(9):3720-3729. [CrossRef]
  35. Hoogenes J WN. A randomized comparison of two robotic virtual reality simulators and evaluation of trainees’ skills transfer to a simulated robotic urethrovescical anastomosis task. Urology. Jan 2018;111:110-115. [CrossRef]
  36. Schmidt MW KK, Fan C, Kowalewski KF, Schmidt LP, Vey J, Proctor T, Probst P, Bintintan VV, Muller-Stich BP, Nickel F. Virtual reality simulation in robot-assisted surgery: meta-analysis of skill transfer and predictability of skill. BJS Open. March 2021;5(2):zraa066. [CrossRef]
  37. Raison N HP, Takashige A, Abdullatif, A Ahmed K, Dasgupta P. Procedural virtual reality simulation training for robotic surgery: a randomised controlled trial. Surg Endosc. Jan 2021:epub ahead of print. [CrossRef]
  38. Costello DM HI, Burke G et al. A review of simulation training and new 3D computer-generated synthetic organs for robotic surgery education. J Robot Surg. Sep 2021:1-15. [CrossRef]
  39. James HK CA, Pattison GTR, Griffin DR, Fisher JD. Systematic review of the current status of cadaveric simulation for surgical training. BJS. 2019;106(13):1726-1734. [CrossRef]
  40. Bellier A CA, Belingheri P, Chaffanjon P. Techniques of cadaver perfusion for surgical training: a systematic review. Surg Radiol Anat. Apr 2018;40(4):439-448. [CrossRef]
  41. Sharma M HA. Comparison of fresh-frozen cadaver and high-fidelity virtual reality simulator as methods of laparoscopic training. World J Surg. Aug 2012;36(8):1732-7. [CrossRef]
  42. LeBlanc F CB, Augestad KM et al. A comparison of human cadaver and augmented reality simulator models for straight laparoscopic colorectal skill acquisition training. J AM Coll Surg. Aug 2010;211(2):250-5. [CrossRef]
  43. Bertolo R GJ, Dagenais J, Sagalovich D, Kaouk JH. Single session of robotic human cadaver training: the immediate impact on urology residents in a teaching hospital. J Laparoendosc Adv Surg Tech A. Oct 2018;28(10):1157-1162. [CrossRef]
  44. Shee K KK, Wu X, Ghali F, Halter R, Hyams E. A novel ex vivo trainer for robotic vesicourethral anastomosis. Journal of Robotic Surgery. 2020;(14):21-27. [CrossRef]
  45. Johnson BA TM, Steinberg RL, Kosemund M, Mueller B, Gahan JC. Design and validation of a low cost, high fidelity model for urethrovesical anastomosis in radical prostatectomy. J Endourol. Apr 2019;33(4):331-336. [CrossRef]
  46. Monda SM WJ, Anderson BG et al. Development and validity of a silicone renal tumor model for robotic partial nephrectomy training. Urology. Apr 2018;114:114-120. [CrossRef]
  47. von Rundstedt FC SJ, Agrawal S, Zaneveld J, Link RE. Utility of patient-specific silicone renal models for planning and rehearsal of complex tumour resections prior to robot-assisted laparoscopic partial nephrectomy. BJU international. Apr 2017;119(4):598-604. [CrossRef]
  48. Ghazi A, Melnyk R, Hung AJ, et al. Multi-institutional validation of a perfused robot-assisted partial nephrectomy procedural simulation platform utilizing clinically relevant objective metrics of simulators (CROMS). BJU Int. Jun 2021;127(6):645-653. [CrossRef]
  49. Crawford DL DA. Evolution and literature review of robotic general surgery resident training 2002-2018. Updates in surgery. Sep 2018;70(3):363-368. [CrossRef]
  50. Hanly EJ MB, Kumar R et al. Mentoring console improves collaboration and teaching in surgical robotics. J Laparoendosc Adv Surg Tech A. Oct 2006;16(5):445-51. [CrossRef]
  51. MT B. Expanded robotic training and education of residents and faculty surgeons using dual console robotic platforms utilizing aviation safety trans cockpit responsibility gradient comparisons. J Minim Invasive Gynecol. 2014;21(6):S5. [CrossRef]
  52. Smith AL SE, Krivak TC, Olawaiye AB, Chu T, Richard SD. Dual-console robotic surgery: a new teaching paradigm. J Robot Surg. 2013;7(2):113-118. [CrossRef]
  53. Morgan MSC SN, Garcia-Gil M et al. Single- versus dual-console robot assisted radical prostatectomy: impact on intraoperative and postoperative outcomes in a teaching institution. World J Urol. Jun 2015;33(6):781-786. [CrossRef]
  54. Jarc AM SS, Adebar T et al. Beyond 2D telestration: an evaluation of novel proctoring tools for robot-assisted minimally invasive surgery. J Robotic Surg. Jun 2016;10(2):103-109. [CrossRef]
  55. Jarc AM SA, Clifford T, Gill IS, Hung AJ. Proctors exploit three-dimensional ghost tools during clinical-like training scenarios: a preliminary study. World J Urol. Jun 2017;35(6):957-965. [CrossRef]
  56. Da Vinci Vision. Accessed 04/01/23. http://intuitive.com/en-us/products-and-services/da-vinci/vision.
  57. ActivSight by Activ Surgical. activsurgical.com/#technology.
  58. Gandaglia G SP, De Naeyer G, D’Hondt F, Mottrie A. Novel technologies in urologic surgery: a rapidly changing scenario. Curr Urol Rep. Mar 2016;17(3):19. [CrossRef]
  59. Adballah M EY, Clavet L et al. Augmented reality in laparoscopic liver resection evaluated on an ex-vivo animal model with pseudo-tumours. Surg Endosc. Nov 2021:epub ahead of print. [CrossRef]
  60. Soler L NS, Pessaux P, Mutter D, Marescaux J. Real-time 3D image reconstrution guidance in liver resection surgery. Hepatobiliary Surg Nutr. Apr 2014;3(2):73-81. [CrossRef]
  61. Bertrand LR OE, Koo B, Buc E, Bartoli A. Augmented reality guidance in laparoscopic hepatectomy with deformable semi-automatic computed tomography alignment. J Visc Surg. Jun 2019;156(3):261-262. [CrossRef]
  62. Bertrand LR AM, Espinel Y et al. A case series study of augmented reality in laparoscopic liver resection with a deformable preoperative model. Surg Endosc. Dec 2020;34(12):5642-5648. [CrossRef]
  63. Porpiglia F CE, Amparore D et al. Augmented-reality robot-assisted radical prostatectomy using hyper-accuracy three-dimensional reconstruction (HA3D) technology: a radiological and pathological study. BJU Int. May 2019;123(5):834-845. [CrossRef]
  64. Hughes-Hallett A ME, Marchus HJ et al. Augmented reality partial nephrectomy: Examining the current status and future perspectives. Urology. Feb 2014;83(2):266-273. [CrossRef]
  65. Hasimoto DA RG, Rus D, Meireles OR. Artificial intelligence in surgery: promises and perils. Ann Surg. 2018;268(1):70-76. [CrossRef]
  66. Tokuyasu T IY, Matsunobu Y et al. Development of an artifiial intelligence system using deep learning to indicate anatomical landmarks during laparoscopic cholecystectomy. Surg Endosc. Apr 2021;35(4):1651-1658. [CrossRef]
  67. Madani A NB, Altieri MS et al. Artificial intelligency for intraoperative guidance: Using semantic segmentation to identify surgical anatomy during laparoscopic cholecystectomy. Ann Surg. Nov 2020:epub ahead of print. [CrossRef]
  68. Mohan A WU, Shaikh MTA, Rahman RM, Zaidi ZA. Telesurgery and robotics: An improved and efficient era. Cureus. Mar 2021;13(3):e14124. [CrossRef]
  69. Marescaux J LJ, Rubino F et al. Transcontinental robot-assisted remote telesurgery: feasibility and potential applications. Ann Surg. Apr 2002;235(4):487-492. [CrossRef]
  70. Perez M XS, Chauhan S et al. Impact of delay on telesurgical performance: study on the robotic simulator dV-Trainer. Int J Comput Assist Radiol Surg. Apr 2016;11(4):581-587. [CrossRef]
  71. Xu S PM, Kun Y et al. Determination of the latency effects on surgical performance and the acceptable latency levels in telesurgery using the dV-Trainer simulator. Surg Endosc. Sep 2014;28(9):2569-2576. [CrossRef]
  72. Korte C NS, Nistor V, Low TP, Doarn CR, Schaffner G. Determining the threshold of time-delay for teleoperation accuracy and efficeincy in relation to telesurgery. Telemed J E Health. Dec 2014;20(12):1078-1086. [CrossRef]
  73. Sterbis JR HE, Herman BC et al. Transcontinental telesurgical nephrectomy using the da Vinci robot in a porcine model. Urology. 2008;71:971-973. [CrossRef]
  74. Bilgic E TS, Watanabe Y et al. Effectiveness of telementoring in surgery compared with on-site mentoring: A systematic review. Surg Innov. Aug 2017;24(4):379-385. [CrossRef]
  75. Papalois ZA AA, Khan A et al. HoloMentor: A mixed reality surgical anatomy curriculum for robot-assisted radical prostatectomy. Eur Surg Res. Nov 2021:epub ahead of print. [CrossRef]
  76. Artsen AM BS, Duvvuri U, Bonidie M. Surgeon satisfaction and outcomes of tele-proctoring for robotic gynecologic surgery. J robot surg. Jul 2021:epub ahead of print. [CrossRef]
  77. Orpheus, an Intuitive company. Accessed 01 01. orpheus-medical.com/surgical-solutions-enhancing-clinical-performance/.
  78. Proximie. Accessed 04/03/23. proximie.com.
  79. Reacts. Accessed 04/03/23. reacts.com.
  80. Mota P CN, Carvalho-Dias E, Costa MJ, Correia-Pinto J, Lima E. Video-based surgical learning: Improving trainee education and preparation for surgery. J Surg Educ. May 2018;75(3):828-835. [CrossRef]
  81. Pape-Koehler C IM, Sauerland S et al. Multimedia-based training on internet platforms improves surgical performance: a randomized controlled trial. Surg Endosc. 2013;27(5):1737-1747. [CrossRef]
  82. Pandya A ES, Ying H, Nokleby M, Reisner LA. A robotic recording and playback platform for training surgeons and learning autonomous behaviours using the da Vinci surgical system. Robotics. 2019;8(1):9-25. [CrossRef]
  83. Huynh D FN, Gok H, Wright A, Towfigh S. Thou shalt not trust online videos for inguinal hernia repair techniques. Surg Endosc. Oct 2021;35(10):5724-5728. [CrossRef]
  84. Daes J FE. Critical view of the myopectineal orifice. Ann Surg. Jul 2017;266(1):e1-e2. [CrossRef]
  85. Celentano V SN, McGrath J et al. How to report educational videos in robotic surgery: an international multidisciplinary consensus statement. Updates Surg. 2021;73(3):815-821. [CrossRef]
  86. Touch Surgery Enterprise. Accessed 01/01/23. touchsurgery.com/enterprise.
  87. Stulberg JJ HR, Kreutzer L et al. Association between surgeon technical skills and patient outcomes. JAMA Surg. Oct 2020;155(10):960-968. [CrossRef]
  88. Prebay ZJ PJ, Miller DC, Ghani KR. Video review for measuring and improving skill in urological surgery. Nat Rev Urol. Apr 2019;16(4):261-267. [CrossRef]
  89. Hussein AA GK, Peabody J et al. Development and validation of an objective scoring tool for robot-assisted radical prostatectomy: Prostatectomy Assessment and Competency Evaluation. J Urol. May 2017;197(5):1237-1244. [CrossRef]
  90. Goh AC GD, Sander JC, Miles BJ, Dunkin BJ. Global Evaluative Assessment of Robotic Skills: Validation of a clinical assessment tool to measure robotic surgical skills. J Urol. Jan 2012;187(1):247-252. [CrossRef]
  91. Siddiqui NY GM, Geller EJ et al. Validity and reliability of the Robotic Objective Structured Assessment of Technical Skills Obstet Gynecol. Jun 2014;123(6):1193-1199. [CrossRef]
  92. White LW KT, Dockter RL, Comstock B, Hannaford B, Lendvay TS. Crowd-sourced assessment of technical skill: A valid method for discriminating basic robotic surgical skills. J Endourol. Nov 2015;29(11):1295-1301. [CrossRef]
  93. Chen C WL, Kowaleksi T et al. Crowd-Sourced Assessment of Technical Skills: A novel method to evaluate surgical performance. J Surg Res. Mar 2014;187(1):65-71. [CrossRef]
  94. Polin MR SN, Comstock BA et al. Crowdsourcing: a valid alternative to expert evaluation of robotic surgery skills. Am J Obstet Gynecol. Nov 2016;215(5):644.e1-644.e7. [CrossRef]
  95. Ghani KR MD, Linsell S et al. Measuring to improve: Peer and crowd-sourced assessments of technical skill with robot-assisted radical prostatectomy. Eur Urol. Apr 2016;69(4):547-550. [CrossRef]
  96. Aghazadeh MA MM, Pan MM, Miles BJ, Goh AC. Performance of robotic simulated skills tasks is positively associated with clinical robotic surgical performance. BJU Int. Sep 2016;118(3):475-481. [CrossRef]
  97. Chen J OP, Cheng N et al. Use of autmoated performance metrics to measure surgeon performance during robotic vesicourethral anastomosis and methodical development of a training tutorial. J Urol. Oct 2018;200(4):895-902. [CrossRef]
  98. Hung AJ CJ, Ghodoussipour S et al. A deep-learning model using automated performance metrics and clinical features to predict urinary continence recovery after robot-assisted radical prostatectomy. BJU Int. Sep 2019;124(3):487-495. [CrossRef]
  99. Lee D, Yu HW, Kwon H, Kong HJ, Lee KE, Kim HC. Evaluation of Surgical Skills during Robotic Surgery by Deep Learning-Based Multiple Surgical Instrument Tracking in Training and Actual Operations. J Clin Med. Jun 23 2020;9(6)doi:10.3390/jcm9061964.
Table 1. Definitions of validity as related to simulation.
Table 1. Definitions of validity as related to simulation.
Face A subjective assessment of how well the simulator replicates the real world
Content A subjective assessment of whether the simulation exercise is providing an accurate assessment of the intended content
Construct An objective assessment of the ability of the simulator to differentiate a novice from an expert
Concurrent An objective assessment of how well the simulator results correlate with current operative performance
Predictive An objective assessment of how well simulator results can predict future operative performance
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated