Next Article in Journal
Development of a Framework for Echocardiographic Image Quality Assessment and Its Application in CRT-D/ICD Patients
Next Article in Special Issue
The Potential Role of an Artificial Intelligence-Driven Tool in Decision-Making for Mitral Valve Repair Surgery
Previous Article in Journal
Second Primary Malignancies in Patients with Colorectal Cancer: The Frequency, Types, and Timeline
Previous Article in Special Issue
Minimally Invasive Nephrectomy for the Management of Polycystic Kidney Disease: The Hilum-First Technique
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sensorized Vascular High-Fidelity Physical Simulator for Robot-Assisted Surgery Training: A Multisite Pilot Evaluation

by
Giulia Gamberini
1,2,3,*,†,
Alessandro Dario Mazzotta
2,3,4,†,
Angela Durante
1,5,
Selene Tognarelli
2,3,
Niccolò Petrucciani
6,
Gianluca Mennini
4,
Gianfranco Silecchia
6 and
Arianna Menciassi
1,2,3
1
Health Science Interdisciplinary Center, Sant’Anna School of Advanced Studies, 56127 Pisa, Italy
2
The BioRobotics Institute, Sant’Anna School of Advanced Studies, 56127 Pontedera, Italy
3
The Department of Excellence in Robotics & AI, Scuola Superiore Sant’Anna, 56127 Pisa, Italy
4
Department of General and Specialist Surgery, Sapienza University of Rome, 00185 Rome, Italy
5
Fondazione Toscana Gabriele Monasterio, 56124 Pisa, Italy
6
Department of Medico-Surgical Sciences and Translation Medicine, Faculty of Medicine and Psychology, St Andrea Hospital, Sapienza University of Rome, 00189 Rome, Italy
*
Author to whom correspondence should be addressed.
Both authors contributed equally to the work.
J. Clin. Med. 2026, 15(3), 1054; https://doi.org/10.3390/jcm15031054
Submission received: 23 December 2025 / Revised: 17 January 2026 / Accepted: 24 January 2026 / Published: 28 January 2026

Abstract

Background/Objectives: Robot-Assisted Surgery poses challenges in skill acquisition due to the lack of haptic feedback, which may lead to adverse intraoperative events. This study focused on a multisite pilot evaluation on the simulator’s ability to discriminate between different levels of expertise and the ability to explore potential differences between surgical specialties. Methods: We built a simulator that can replicate anatomies of vascular and adipose tissue. A resistive stretching sensor was integrated into a silicone vessel to objectively measure its deformation. A total of 18 males and 12 females, aged between 26 and 64 years old, participated to the study. In total, there were 30 participants, (21 general surgeons, 2 thoracic surgeons, 4 gynecologists, 3 urologists) and they performed two repetitions of a surgical task and filled in a questionnaire about face- and content validities and a system usability scale. The tests were conducted between February and October 2023. Results: The discriminant validity was positively assessed, considering the maximum deformation value (p-value = 0.0479) and the mean deformation value (p-value = 0.0317). Differences were found between urologists, (i) general surgeons (p-value = 0.0167) and, (ii) gynecologists (p-value = 0.0495). The face- and content validity of the simulator received 80% and 90% of positive answers, respectively. Conclusions: Future works will deal with the evaluation of the simulator abilities in surgical training by comparing surgeons trained on the simulator to those who are not.

1. Introduction

In recent years, the increasing adoption of Robot-Assisted Surgery (RAS) has resulted in the evolution of training methodologies that improve the role of simulation [1]. Robot-assisted surgical procedures are increasing every year with more than 2,680,000 procedures performed using a da Vinci Surgical System in 2024 (data taken from the Intuitive Surgical Inc., Annual Report). RAS allows us to achieve increased surgical precision, less discomfort after surgery, shorter hospital stays, and faster recovery time, all of which improve patient outcomes and lessen the strain on healthcare resources, which plays a critical role in public health [2]. Due to different control strategies, from more analogic “hands on” systems to digital devices and indirect consoles, the length of the learning curve has increased (40–70 cases to reach proficiency) [3]. Furthermore, the minimally invasive procedures reduced the haptic feedback in robotic surgery [4]. On this matter, a specific new approach is needed to ensure adequate medical training, which will maintain patient safety during robotic surgical procedures [5,6,7]. Since the introduction of simulations in healthcare, the skills required by a robotic surgeon can be trained using different approaches. Simulation training includes dry-laboratory activities, wet-laboratory activities, and virtual-reality (VR) simulators. It offers a risk-free environment for trainees and enables them to practice complex procedures, which can enhance clinical performance and patient outcomes [8,9]. In particular, when dealing with delicate tissues, such as blood vessels, the specific training is much more needed as the application of excessive force may cause intraoperative complications, like massive bleeding or damage to other tissues or organ, which may lead to the immediate conversion to open surgery [10,11,12,13].
The application of a large deformation leads to high strain values, which are a primary cause of vascular rupture and present a considerable risk of intraoperative hemorrhage. In clinical practice, excessive force applied during vessel manipulation, whether sudden or sustained, is a recognized factor contributing to bleeding and may require conversion to open surgery [14,15,16]. In addition, mean deformation is a crucial parameter for assessing the capabilities of handling vessels during the procedure [17,18]. Repeated or prolonged deformation, even without high peak events, can cause microtrauma to the vessel wall. Mechanical stress can trigger inflammation and compromise tissue integrity, increasing the risk of postoperative complications that may cause stenosis or pseudoaneurysm formation and/or anastomotic healing [19,20].
In order to reduce the number of such adverse events, surgeons must be able to visually assess tissue deformation—a concept known as ‘visual haptic’—in addition to precise hand–eye coordination [21]. To acquire visual-haptic competence, structured training is mandatory to achieve the combination of specific knowledge, a safe experience, and a productive reflection in a simulation environment [22,23]
However, few physical simulators are available to ensure the gaining of visual haptics, especially regarding vascular robotic surgery [24]. Thus, the aim of this article is to present a physical simulator for robotic surgery and to perform discriminant validity tests within different pools of end-users.

2. Materials and Methods

A high-fidelity physical simulator of the vascular structure was realized as reported in [24]. The simulator prototype under the da Vinci Xi (Intuitive Surgical Inc., Sunnyvale, CA, USA) during the validation tests is shown in Figure 1.
The simulator was meant to teach and train the isolation and resection of vascular structures, which is a common interventional task, while removing a target tissue due to structural abnormalities or tumors. Our simulator was meant to reproduce the structure of a main vein comparable with a pulmonary vein (length of 120 mm, diameter of 13 mm, and wall thickness of 1.5 mm) [25], and it was made through cast molding using silicone Ecoflex 00-30 (Smooth-On, Macungie, PA, USA)—chosen to match the mechanical properties of human tissues, as reported in [19].
In this framework, the specific aims of this study were to assess the following:
(i)
the discriminant validity of the simulator—defined in our case as the ability to discriminate between different levels of expertise between novices, fellows, and expert surgeons [26];
(ii)
the ability to explore, based on the tests results, the potential differences among the surgical specialties using a stratification of the sample.

2.1. Description of the Technical Features of the Vascular Simulator

As stated above, the high-fidelity physical simulator was designed to replicate the anatomic structure of a main vein and its adipose surrounding tissue [24]. The surrounding tissue was realized as a parallelepiped with an elliptical central hole for housing the vessel, and it was made using polyvinyl alcohol (PVA), allowing its electrocoagulation. A support structure was created to securely attach the simulator to the surgical table during the testing sessions and to store all the electronic components.
To measure vessel deformation during surgical manipulation and to evaluate forces induced in the vessel, a customized resistive stretching sensor was integrated into the vein. The sensor is soft, embeddable, and capable of detecting deformation of the vascular structure from 0 to 30% of strain. It is fabricated using a commercial fabric, i.e., Stretch Conductive Fabric (Less EMF, Latham, NY, USA), and deformation signals are acquired, recorded, and stored through a graphical user-interface program realized using LabVIEW 2019 software (National Instruments, Austin, TX, USA). As the training was aimed at assessing only technical manipulation, the blood stream was not included in the simulator.
The vessel and its adipose surrounding tissues were mechanically characterized by means of tensile tests. As shown in [24], our results are similar to 90.4% with respect to the pulmonary porcine vein and were obtained using an Instron machine (Instron, Norwood, MA, USA). The adipose surrounding tissue was characterized by performing compressive tests (controlled by a displacement percentage ranging from 0% to 90% at a rate of 200 mm/min) and it showed a Young’s modulus of 27.34 ± 2.79 kPa, that is comparable with the human adipose organ-surrounding tissue [27].

2.2. Population

A pre-statistical analysis was performed to determine the desired sample size. G*Power 3.1.9.7 software [28] was used and an ANOVA: Fixed-effects, omnibus, one-way was conducted with the following parameters:
  • Effect size equal to 0.6,
  • α = 0.05 (representing the significance level),
  • Power equals 0.8,
  • Number of groups equal to three.
The sample size was computed a priori, and we determined that the total sample size should be equal to 30 people.
Participants were selected from two academic institutions: the Research Institute Against Digestive Cancer (IRCAD), Strasbourg, France, and the Azienda Ospedaliera Universitaria (AOU) Sant’Andrea, La Sapienza, Rome, Italy. Specific ethical committee approvals were obtained, i.e., IRCAD—New Devices 2022–2025 and Sant’Anna School of Advanced Studies Authorization n°37/2023. All the experiments were performed in accordance with the General Data Protection Regulation (GDPR). All the recruited surgeons signed an informed consent form before starting the experiments. The evaluation was conducted in an operating theater using a da Vinci Xi surgical robot.
The surgeons were divided into three different groups according to the length of experience declared as the primary operator in RAS. They were classified as follows:
  • Novices: Junior residents with no RAS experience (0 years of RAS experience),
  • Fellows: Senior residents and young surgeons with medium experience (1–2 years of experience in RAS),
  • Experts: Surgeons with extensive RAS experience (at least 3 years of RAS experience).
In addition, participants were selected from different surgical specialties using a maximum variation sampling between those available in the hosting hospitals (general surgery, thoracic surgery, gynecological surgery, and urological surgery) [21]. The general surgery population includes surgery of the digestive system, oncological surgery, emergency surgery, endocrine surgery, reconstructive surgery, and organ transplant surgery.

2.3. Protocol

At the beginning of the testing session, all participants received a brief description of the surgical task through a slide presentation and a standardized instructional video; then, the study objectives and the safety- and the data-protection information were presented. After these preliminary phases, participants who were willing to participate in the tests were asked to sign an informed consent form. Demographic information and years of experience in RAS were collected anonymously using a questionnaire.
Participants underwent a familiarization session lasting 5 min. The familiarization session relied on guided steps to allow the users to understand the task workflow, the equipment, and the requirements. Moreover, they were instructed to operate the da Vinci Surgical System in safe conditions.
The simulated surgical task comprised three main steps:
  • Dissection and isolation of a target vessel.
  • Passage of vessel loops under the vessel to pull it up.
  • Insertion of a robotic stapler for vessel transection.
Two separate repetitions of the entire task per participant were performed. Sensor data were recorded to evaluate participants’ performance metrics that included the maximum vessel deformation, reached during the simulated surgical task, and the mean vessel deformation along the surgical task. Both maximum- and mean deformation signals were used as performance indicators of clinical competence during the simulated surgical task. At the end of the surgical tasks, the surgeons were asked to fill in an anonymous ad hoc questionnaire. The questionnaire contains 22 items, divided into three main sections. The three sections were about the face validity of the simulator, the content validity of the simulator, and the system usability scale (SUS). The face- and content validities were evaluated using Messick’s contemporary framework [29]. The other aspects of Messick’s framework were not considered in this work, as they do not apply to our specific case. Face validity, also known as Evidence based on Response Process Validity [30], is the subjective view end-users have of how realistic a simulator is [31]. Content validity, or Evidence based on Test Content validity [32], is defined as the extent to which a simulator’s content is representative of the knowledge or skills that have to be learned in a real clinical setting [33]. The SUS is a reliable, zero-cost psychometric tool composed of 10 alternating positive- and negative-formulated statements for which a respondent gives a subjective evaluation of a system’s usability [34]. The complete list of items divided into three sections is reported in Table 1. The answers were recorded through a Linkert scale ranging from 0 to 5, where 0 represents very-unrealistic-for-the-face-validity items and strongly-disagrees-with-the-content-validity-and-SUS items, while 5 represents very-realistic-for-the-face-validity items and strongly-agrees-with-the-content-validity-and-SUS items.

2.4. Statistical Analysis

2.4.1. Sensor Data Statistical Analysis

To evaluate discriminant validity and assess the ability of the simulator to distinguish between different surgical specialties, the Kruskal–Wallis ANOVA test [28] was performed in MATLAB R2024b software (MathWorks, Natick, MA, USA). Whenever statistically significant results were achieved, post hoc analyses were performed.

2.4.2. Questionnaire Data Statistical Analysis

In addition, a descriptive statistical analysis was performed for the validity items of the face and content, as well as for the SUS questions. To evaluate the overall internal consistency of the questionnaire for each of the three sections, McDonald’s omega coefficient was used [35]. For the content validity items, additional analyses were performed:
  • Calculation of the Content Validity Index (I-CVI) defined as the proportion of surgeons who rated an item as relevant.
    I - C V I = Number of surgeons rating   4   or   5 Total number of surgeons
  • Calculation of the Scale-level CVI (S-CVI) defined as the average of the I-CVIs across items [32].
Regarding the SUS, the total SUS score was calculated. The SUS is usually made up of both positive- and negative constructed elements; therefore, the total SUS score was calculated according to the literature [30] by reversing the negative elements, summing all the scores obtained, and scaling it to 100. In particular, the following should be considered:
  • Negative items are considered as 5 values achieved .
  • Positive items are considered as values achieved 1 .

3. Results

In total, 30 surgeons participated (13 novices, 8 fellows, and 9 experts). The main features of the three groups in terms of age, sex, dominant hand, specialties, and clinical-training experience in the simulation are reported in Table 2.
Although we tried to obtain at least one representative level of experience and/or surgical specialties (general surgery, thoracic surgery, gynecology, and urology), in the expert group, thoracic surgeons and urologists were missing. This was mainly due to the inaccessibility of experienced surgeons in these specialties in the surgical groups contacted. Regarding simulation experience, in the novice group, 69% of the subjects had previous experience with simulators for clinical training. In the fellow- and in the expert groups, this percentage reaches 88% and 100%, respectively.
Based on the sensor data, the vessel simulator was able to distinguish between the three different levels of experience, i.e., novices, fellows, and experts. Considering the maximum value of vessel deformation derived from continuously recorded data during the first repetition of the task, a statistically significant difference between the three groups was found—p-value equal to 0.0479, as shown in Figure 2A. The distribution of the data was evaluated, confirming a distribution not significantly different from a normal distribution (18% of difference). Thus, to verify whether the use of a non-parametric test was appropriate, Levene’s test for equality of variances was performed. A p-value of 0.018 was found, indicating that variances differ; therefore, the use of a non-parametric test is consistent with our aim (the full analysis of data distribution and variance is reported in Appendix A, Figure A1). The post hoc statistical analysis did not highlight any significant differences overall; however, when considering the individual pairs (experts vs. novices, experts vs. fellows, and fellows vs. novices), a statistically significant difference was found between experts and novices, with a p-value of 0.0178. Moreover, a statistical significance difference between the three levels of experience was also found by considering the mean value of the vessel deformation during the second task-repetition with a p-value of 0.0317 (Figure 2B). Data distribution does not differ significantly from a normal distribution (12% of difference). A p-value of 0.07 was found performing Levene’s test for equal variances; thus, variances do not differ, and neither a non-parametric test nor ANOVA can be used. To evaluate all the metrics with the same methodology, a non-parametric statical test was kept as evaluation performance metrics (the full analysis of data distribution and variance is reported in Appendix A, Figure A2). From the post hoc analysis, a statistically significant difference was found between mean ranks of expert and fellows (p-value = 0.0067).
By considering the results presented here, the discriminant validity in terms of levels of expertise was positively assessed. In addition, using the recruitment modality, the collected data was analyzed moving from surgical expertise to surgical specialties. An additional comparison was carried out by dividing the simulator data collected during the tests into four different groups, one for each clinical specialty only. This additional analysis was carried out to determine whether the simulator was also able to distinguish between them. A statistically significant difference cannot be achieved (p-value of 0.2057) when comparing all the specialties together. This could be justified by the unequal distribution of the users in the four specialties groups.
Due to the large difference in users’ numbers, and to better understand the ability of the simulator to discriminate between the different surgical specialties, a reduced sample of general surgeons was considered. By considering the number and the expertise distribution (percentage data are reported within round bracket in Table 2) of the other surgical specialties, a proportional number of general surgeons was randomly selected. By using a random-generator number extraction, four novices, two fellows, and one expert surgeon were randomly selected from the general-surgery population by performing a random under sampling in accordance with the literature [36]. It is important to note that the differentiation of the levels of the intra-user group was maintained. Then, this new user group (general surgery 2) was used to evaluate the ability of the simulator to differentiate between different surgical specialties. Nevertheless, no statistically significant difference was achieved (p-value = 0.071), Figure 3—top part. The failure to reject the null hypothesis of samples coming from different populations’ distribution could again be linked to the limited size of the population groups. A higher number of surgeons should be involved to further analyze the capabilities of the simulator in differentiating between surgical specialties.
However, to further analyze the ability of the simulator to discriminate between surgical specialties, a couples’ analysis was performed. In particular, the six different combinations of pairs of surgical specialties were compared and analyzed. Maximum vessel deformation data were used as reference during the second repetition of the surgical task (after familiarization tasks). This choice was made to reduce the effects of performance bias and variability due to unfamiliarity that may impact the initial attempt, typically affected by the user’s adaptation to the simulator and the surgical procedure.
The pairs tested are
  • General surgery vs. thoracic surgery;
  • General surgery vs. gynecological surgery;
  • Gynecological surgery vs. thoracic surgery;
  • Gynecological surgery vs. urological surgery;
  • General surgery vs. urological surgery;
  • Urological surgery vs. thoracic surgery.
Based on the analysis performed, a statistically significant difference was found between gynecological and urological surgeries (p-value of 0.0495) and between general and urological surgeries (p-value of 0.0167).
The results of the analysis for the comparisons of surgical specialties are depicted in Figure 3. For the analysis of the couple, only those with a statically significant difference were reported. In addition to maximum- and mean metrics, the area under the deformation–time curve (Figure 4A) was also analyzed in order to capture the cumulative strain applied to the vessel over the duration of the task.
From the analysis of the area under the curves, a statistically significant difference when comparing the different surgical specialties during the second repetition of the task was found (p = 0.0427), as reported in Figure 4B.
Finally, as mentioned above, general surgery was the specialty most represented in our population samples (21 users). To further assess the ability of the simulator, a comparison analysis restricted to only the general-surgery specialty was carried out. In particular, a comparison between the different levels of expertise in the general-surgery population was made through the Kruskal–Wallis ANOVA test. From the results obtained, a statistically significant difference between novices, fellows, and experts was found when considering the maximum deformation value of the vascular structure during the first repetition—p-value of 0.0327 (Figure 5A). Moreover, the difference in the mean vessel deformation during the second repetition was found to be statistically significant between the three levels of expertise with a p-value of 0.0288 (Figure 5B). From the post hoc analysis, a statistically significant difference was found between experts and novices in the maximum value of the first repetition, while for the mean value of the second repetition, a statistically significant difference was found between experts and fellows.
This finding can be explained by a procedural reason. In the initial repetition, higher vessel-deformation values—especially in the novice and fellow groups—can be attributed to their unfamiliarity with the surgical task. At this stage, less-experienced users often apply excessive or inadequately moderate force, leading to an increased maximum-deformation of the vascular structure. This behavior aligns with preliminary actions and an absence of advanced haptic perception and motor coordination.
In the second repetition, a performance-stabilization effect is noted: participants, having been previously exposed to the task, begin to modify their interaction with the simulator. As a result, peak deformation values become more consistent and no longer differ significantly among experience levels, while the mean deformation—reflecting overall precision and control—emerges as the more-sensitive parameter for discriminating between levels of expertise. This shift supports the simulator’s responsiveness not only to initial technical gaps but also to short-term learning- and adaptation processes.
The analysis of the results of the questionnaire—administered after the testing session—about evidence based on response processes (face validity), evidence based on test content validity, and the system usability scale is reported in Figure 6. Considering the face validity of the simulator, the overall global impression of the simulator was positively evaluated, with 29 surgeons declaring the simulator to be somewhat realistic or very realistic. Regarding the realism of the anatomical components, the simulator with the adipose tissue was considered more realistic than the version without the adipose tissue (Item 3 was rated higher than Item 2). Both the visual appearance of the vein and the adipose tissue (Items 4 and 5) were positively evaluated (somewhat realistic or very realistic) by 24 surgeons (80%). The haptic feedback when manipulated with hands, of both the vein and the adipose tissue, was evaluated positively by 20 and 23 surgeons (Items 6 and 7), while the interaction between the surgical instruments was positively evaluated by 23 and 25 surgeons for the vein and the adipose tissue, respectively (Items 8 and 9).
Regarding the content validity (Items 10–12), the simulator was thought to be useful in teaching:
  • Vascular structure isolation by 28 surgeons;
  • The stapling of vascular structure by 26 surgeons;
  • To minimize the forces applied to vascular structures by 28 surgeons.
The system usability scale of the simulator was performed to evaluate the ease of use of the simulator and the perceived availability in using the simulator by end-users. Twenty-four surgeons declared that they would like to use the simulator as a training platform (Item 13). Only two surgeons found the simulator more complex than expected (Item 14). Twenty-one surgeons thought that the simulator was easy to use (Item 15). Only six declared they would need a technician to support them in using the simulator, and five found difficulties in understanding the electronics connections (Items 16 and 18). The graphical user-interface was found understandable and clear by 21 surgeons (Item 17). A total of 28 and 27 users think that most of their colleagues will learn how to use the simulator very fast and that the system is intuitive and “plug and play” (Items 19 and 20). All surgeons agreed on being comfortable with the use of the simulator. Only seven surgeons declared that they would need more time to become familiar with the setup of the simulator.
Descriptive statistical analyses and McDonald’s omega analyses were performed both for face- and content validities, as well as for the system usability scale. Evidence based on response processes validity was used to evaluate whether the respondents interpreted the realism of the anatomical structures, mechanical haptics, and instrument–tissue interactions.
Table 3 reports the values of the mean and standard deviation for each item from one to nine (which are related to the evidence based on response processes validity). This is also graphically depicted in Figure A3. By the calculation of the McDonald’s omega for evidence based on response processes validity, we achieved an omega value equal to 0.978, thus representing an excellent internal consistency of the questionnaire items [37,38,39].
In Table 4, the values of the mean and standard deviation for each item from 10 to 12 (which are related to the evidence based on test content validity) are reported. This is also graphically depicted in Figure A4. By the calculation of the McDonald’s omega, for evidence based on response processes validity, we achieved an omega value equal to 0.940, thus representing an excellent internal consistency of the questionnaire items [37,38,39].
In addition, for the content validity, content validity indexes were calculated for each item of the questionnaire:
  • I-CVI for item 10: 0.933.
  • I-CVI for item 11: 0.867.
  • I-CVI for item 12: 0.933.
All the items have a CVI index greater than 0.78; thus, the items have excellent content validity, and they can be retained. Regarding the scale-level CVI that considers the average of all the I-CVI items, we achieved a S-CVI equal to 0.911 [32].
Table 5 reports the values of the mean and standard deviation for each time from 13 to 22 (which deal with the usability of the simulator). This is also graphically reported in Figure A5. By the calculation of the McDonald’s omega, for SUS we achieved an omega value equal to 0.978, thus representing an excellent internal consistency of the questionnaire items [32,33,34]. Moreover, the total SUS score was calculated, achieving a score equal to 75.92, showing an average usability of the system [40].

4. Discussion

This paper presents the pre-clinical assessment of a soft sensorized high-fidelity physical vascular structure with simulated adipose tissue by involving 30 surgeons with different levels of experience in RAS working in different surgical specialties. The simulator was designed to enable an objective- and proctor-independent evaluation of robotic-assisted surgical manipulation of delicate tissues, thanks to a resistive, customized stretching sensor embedded into the silicone vascular structure able to detect the vessel deformation through the simulated surgical task.
Although distinguishing between experience classes—particularly among experts, fellows, and novice—presents challenges and it is not usually presented, a statistically significant difference in maximum tissue-deformation was identified across the three users’ groups. Moreover, a statistically significant difference in mean deformation data was found for experts, fellows, and novices. This emphasizes the potential of sensor-based simulators to offer quantifiable metrics that can be used to objectively distinguish surgical-skill levels, thereby enabling targeted enhancements to training programs.
One of the main challenges of this study was recruiting fellows and expert surgeons, who are often difficult to involve. Fellows are in a transitional phase between training and independent practice, so their availability is limited. Expert surgeons are usually busy with clinical duties, which also makes their participation harder. As a result, these groups are often underrepresented, making it difficult to balance the sample size across different experience levels [41,42].
One of the main goals achieved by this work was the ability to extend the investigation to all three categories of surgeons, revealing significant differences in performance when comparing their data to those of both novices and experts. Future research could benefit from expanding recruiting criteria or collaborating with additional training facilities to obtain a more representative sample of fellows, thereby strengthening the validity of data comparisons across various experience levels. On the other hand, involving additional training facilities could be a solution for increasing the number of expert surgeons involved in the analysis.
In addition to the analysis based on surgical expertise, the study examined the simulator’s ability to differentiate among different surgical specialties [43]. Two pairwise comparisons demonstrated statistically significant differences in maximum vessel-elongation during the second task repetition: general surgery (gen2, n = 7) versus urology (n = 3) (p = 0.0167) and gynecology (n = 4) versus urology (n = 3) (p = 0.0495). This difference may be ascribed to the differing anatomical contexts and tissue types often encountered in their respective practices.
General surgeons frequently operate on tissues like the mesentery or bowel that exhibit relative mobility and are more amenable to manipulation. In contrast, urologists typically perform procedures in anatomically confined and fibrotic spaces, such as the retroperitoneum and pelvis, which demand a high degree of precision and delicate tissue handling. A similar pattern was noted between gynecologists and urologists, where the difference also reached statistical significance. Although both professions often function within the pelvis, the anatomical structures they engage with vary significantly. Gynecologic procedures usually involve more compliant responsive tissues, such as the uterus and adnexa, whereas urologic interventions involve denser- and less-elastic structures like the bladder and prostate. Nonetheless, these results should be regarded with caution because of the limited sample sizes of the urologic- and gynecologic cohorts. The statistical significance indicates the simulator’s ability to differentiate modest inter-specialty changes in tissue-manipulation behavior; nevertheless, more research with larger, more balanced populations is necessary to validate these findings and investigate their therapeutic implications more thoroughly [44,45].
Additionally, the comparison of the discriminant validity of the simulator when restricted to general surgery was carried out, and a statistically significant difference was found.
The obtained results are in line with the literature indications; Cundy P. et al. highlighted that (i) simulators can be used to enhance skills related to force sensing, and (ii) surgical skills are influenced by expertise level [45,46,47]. Our results confirm that surgical skills are influenced by the levels of experience. In fact, a statically significant difference was found between the three expertise levels. Future works will deal with the evaluation of the simulator abilities to enhance skills related to force sensing by comparing surgeons trained on the simulator to those who are not. In addition, thanks to the high-fidelity features of the simulated scenario, the proposed simulator allowed surgeons to train and test on a model that, not only from a mechanical point of view but also from a visual appeal, replicates the properties of human tissues, and thus allows them to train the so-called visual haptic. Moreover, even with the small number of involved users, statistically significant results were obtained. However, a larger number of surgeons could better demonstrate the training process highlighting more evident skill progression and reducing variability in statistical power. Moreover, a more balanced sample size across the different experience levels and the different surgical specialties could affect the statistical power of the analysis.
In addition, the increased adoption of artificial intelligence (AI) is also expanding in the field of surgical simulation [48]. AI-driven systems can include virtual-reality and augmented-reality surgical training modules as well as video-based assessment tools that can evaluate technique and efficiency, helping trainees to practice skills repeatedly in safe and controlled environments [49,50]. In our specific case, since the simulator is equipped with both visual and acoustic feedback to alert the user when a predefined strain-threshold is exceeded, it would be interesting to integrate artificial intelligence into training courses for clinicians by developing an AI-driven algorithm that adapts the deformation safety-threshold to the surgeon’s performance metrics.
From the questionnaire analysis, the overall global impression of the simulator was positively evaluated by 29 surgeons out of 30. The simulator with the adipose surrounding tissue was appreciated for the realism and the possibility of electrocoagulation given by the PVA-based tissue. The simulator’s face validity, or process-based validity evidence, was positively evaluated, with an average of 80% positive responses. With an average of 27 surgeons out of 30, the simulator was thought to be useful in teaching vascular-structure isolation, the stapling of the vessel, and the minimization of the forces applied. By considering these results, the content validity of the simulator was positively assessed. From the results of the SUS questions, the system was positively evaluated in terms of usability and ease of use. A total of 80% of the surgeons declared that they would like to use the simulator as a training platform, and 70% declared that the graphical user-interface was easy to understand. Overall, the total SUS score was equal to 75.92. In addition, all the surgeons were comfortable in using the simulator; thus, by considering the results reported above, the system usability was positively assessed.

5. Conclusions

To conclude, the work presented a multisite pilot evaluation of a sensorized high-fidelity physical simulator of vascular structure. The simulator was tested by 30 surgeons who have different levels of expertise and work in a variety of surgical fields.
A statistically significant difference in maximum- and mean tissue-deformation data was identified for experts, fellows, and novices. Moreover, statistically significant differences in maximum vessel-elongation during the second task repetition between general surgeons and urologists and between gynecologist and urologist were found.
In addition, from the questionnaire analysis, the overall global impression was positively evaluated by 29 surgeons out of 30. The simulator’s face validity obtained an average 80% of positive responses and the content validity was positively assessed. In addition, all the surgeons were comfortable in using the simulator; thus, the system usability was positively assessed.
The results align with the literature, confirming that surgical skills significantly vary with experience level. The high-fidelity simulator effectively supports training by realistically replicating mechanical- and visual tissue properties, enabling the development of force sensing and visual-haptic skills. Although statistically significant results were achieved with a small sample, larger and more balanced cohorts are needed to better demonstrate skill progression and improve statistical power.
Additionally, while the maximum- and mean deformation data provided quantifiable measures of tissue manipulation skills, they represent just some aspects of surgical proficiency. Other aspects, such as decision-making, also under pressure, and error correction were not assessed in this study. Including a broader range of evaluation criteria, a more comprehensive understanding of the competencies developed through robotic-assisted surgery training will be reached. Finally, from a technical point of view, additional efforts will be dedicated to fully replicating the complexity of a live surgical environment. Factors such as real-time bleeding, patient-specific anatomical variations, and more complex vascular structures have to be added to the current system to improve the training reality and potential translation to clinical settings.
Lastly, tracking the surgeons’ movements to understand the correlations between his/her arms movement and the vessel deformation could be an interesting future perspective of the proposed study. Motion-tracking sensors are becoming increasingly popular in medicine as a method for evaluating surgical psychomotor abilities. These sensors have been employed in numerous studies to compare the performance of novices and experts [51,52]. Combining motion tracking with strain measurement could further enhance the simulator’s ability to capture differences in skill levels, offering a more comprehensive understanding of a surgeon’s technical capabilities. This would also offer the possibility of translating clinical-simulation experience to the surgical environment.

Author Contributions

Conceptualization, G.G., S.T. and A.M.; methodology, G.G., A.D.M., S.T. and A.M.; software, G.G.; validation, G.G., A.D.M., S.T., N.P., G.M., G.S. and A.M.; formal analysis, G.G., A.D.M. and A.D.; resources, S.T. and A.M.; data curation, G.G., A.D. and A.D.M.; writing—original draft preparation, G.G. and A.D.M.; writing—review and editing, A.D., S.T. and A.M.; supervision, A.M.; funding acquisition, A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the BRIEF project (project identification code IR0000036) funded under the National Recovery and Resilience Plan (NRRP), Mission 4 Component 2 Investment 3.1 of the Italian Ministry of University and Research funded by the European Union. This work was also supported by FIS (Fondo Italiano per la Scienza), within the FIS00001278 IBRIS (Innovative BioRobotic artIficial organS) project framework.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the IRCAD and Sant’Anna School of Advanced Studies Ethical Committees (protocols: IRCAD—New Devices 2022–2025 and Sant’Anna School of Advanced Studies Authorization n°37/2023, date: 25 September 2023) for studies involving humans.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

The authors would like to thank the IRCAD group and facilities for the opportunity to conduct part of the present study with their support.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Data distribution divided into groups of the experience’s maximum deformation achieved during the first repetition is depicted in Figure A1 left, while the Kernel Density estimates are reported in Figure A1 right.
Figure A1. Data distribution of maximum deformation divided per novices (light blue), fellows (light pink) and experts (light yellow) is depicted on the left, while on the right, the kernel density estimates are plotted. According to the analysis, data did not differ significantly from a normal distribution.
Figure A1. Data distribution of maximum deformation divided per novices (light blue), fellows (light pink) and experts (light yellow) is depicted on the left, while on the right, the kernel density estimates are plotted. According to the analysis, data did not differ significantly from a normal distribution.
Jcm 15 01054 g0a1
The data distribution, divided into groups based on the mean deformation achieved during the second repetition, is shown in Figure A2 (left), while the Kernel Density estimate is reported in Figure A2 (right).
Figure A2. Data distribution of mean deformation divided per novices (light blue), fellows (light pink) and experts (light yellow) is depicted on the left, while on the right, the kernel density estimates are plotted. According to the analysis, data did not differ significantly from a normal distribution.
Figure A2. Data distribution of mean deformation divided per novices (light blue), fellows (light pink) and experts (light yellow) is depicted on the left, while on the right, the kernel density estimates are plotted. According to the analysis, data did not differ significantly from a normal distribution.
Jcm 15 01054 g0a2
The histogram plots of the different ratings of the questionnaire items are reported below. Figure A3, Figure A4 and Figure A5 illustrate the evaluation results: Figure A3 presents face validity, Figure A4 presents content validity, and Figure A5 presents the System Usability Scale scores.
Figure A3. Histogram plot of the mean and standard deviation of the results achieved for the face validity (Item 1 to Item 9).
Figure A3. Histogram plot of the mean and standard deviation of the results achieved for the face validity (Item 1 to Item 9).
Jcm 15 01054 g0a3
Figure A4. Histogram plot of the mean and standard deviation of the results achieved for the content validity (Item 10 to Item 12).
Figure A4. Histogram plot of the mean and standard deviation of the results achieved for the content validity (Item 10 to Item 12).
Jcm 15 01054 g0a4
Figure A5. Histogram plot of the mean and standard deviation of the results achieved for the System Usability Scale (Item 13 to Item 22).
Figure A5. Histogram plot of the mean and standard deviation of the results achieved for the System Usability Scale (Item 13 to Item 22).
Jcm 15 01054 g0a5

References

  1. Chen, I.-H.A.; Ghazi, A.; Sridhar, A.; Stoyanov, D.; Slack, M.; Kelly, J.D.; Collins, J.W. Evolving Robotic Surgery Training and Improving Patient Safety, with the Integration of Novel Technologies. World J. Urol. 2021, 39, 2883–2893. [Google Scholar] [CrossRef]
  2. Handa, A.; Gaidhane, A.; Choudhari, S.G. Role of Robotic-Assisted Surgery in Public Health: Its Advantages and Challenges. Cureus 2024, 16, e62958. [Google Scholar] [CrossRef] [PubMed]
  3. Wong, S.W.; Crowe, P. Factors Affecting the Learning Curve in Robotic Colorectal Surgery. J. Robot. Surg. 2022, 16, 1249–1256. [Google Scholar] [CrossRef] [PubMed]
  4. Colan, J.; Davila, A.; Hasegawa, Y. Tactile Feedback in Robot-Assisted Minimally Invasive Surgery: A Systematic Review. Robot. Comput. Surg. 2024, 20, e70019. [Google Scholar] [CrossRef] [PubMed]
  5. Alemzadeh, H.; Raman, J.; Leveson, N.; Kalbarczyk, Z.; Iyer, R.K. Adverse Events in Robotic Surgery: A Retrospective Study of 14 Years of FDA Data. PLoS ONE 2016, 11, e0151470. [Google Scholar] [CrossRef]
  6. Muaddi, H.; Stukel, T.A.; De Mestral, C.; Nathens, A.; Pautler, S.E.; Shayegan, B.; Hanna, W.C.; Schlachta, C.; Breau, R.H.; Hopkins, L.; et al. Adverse Events Following Robotic Surgery: Population-Based Analysis. Br. J. Surg. 2022, 109, 763–771. [Google Scholar] [CrossRef]
  7. Bergholz, M.; Ferle, M.; Weber, B.M. The Benefits of Haptic Feedback in Robot Assisted Surgery and Their Moderators: A Meta-Analysis. Sci. Rep. 2023, 13, 19215. [Google Scholar] [CrossRef]
  8. Schmelzle, M.; Malinka, T.; Bahra, M.; Pratschke, J. Teaching und Training in der robotergestützten Chirurgie. Zentralbl Chir 2020, 145, 271–277. [Google Scholar] [CrossRef]
  9. Rahimi, A.M.; Uluç, E.; Hardon, S.F.; Bonjer, H.J.; Van Der Peet, D.L.; Daams, F. Training in Robotic-Assisted Surgery: A Systematic Review of Training Modalities and Objective and Subjective Assessment Methods. Surg. Endosc. 2024, 38, 3547–3555. [Google Scholar] [CrossRef]
  10. Berry, M.F. Pulmonary Artery Bleeding During Video-Assisted Thoracoscopic Surgery. Thorac. Surg. Clin. 2015, 25, 239–247. [Google Scholar] [CrossRef]
  11. Awad, M.M.; Raynor, M.C.; Padmanabhan-Kabana, M.; Schumacher, L.Y.; Blatnik, J.A. Evaluation of Forces Applied to Tissues during Robotic-Assisted Surgical Tasks Using a Novel Force Feedback Technology. Surg. Endosc. 2024, 38, 6193–6202. [Google Scholar] [CrossRef]
  12. Uchida, S.; Matsunaga, T.; Tomita, H.; Fukui, M.; Hattori, A.; Takamochi, K.; Suzuki, K. Usefulness of Final Transection of the Proximal Pulmonary Artery in Robotic Left Upper Lobectomy. Interdiscip. Cardiovasc. Thorac. Surg. 2024, 38, ivae054. [Google Scholar] [CrossRef] [PubMed]
  13. Pagès, P.-B.; Gutierrez, I.; Baste, J.-M. Pulmonary Artery Repair during Robotic Lung Resection: Narrative Review and Management. J. Vis. Surg. 2022, 8, 3. [Google Scholar] [CrossRef]
  14. AlMasri, S.; Kraftician, J.; Zureikat, A.; Paniccia, A. Management of Intra-Operative Hemorrhage and Safe Venous Resection in Robotic-Assisted Pancreaticoduodenectomy: Techniques to Avoid Open Conversion. J. Gastrointest. Surg. 2023, 27, 1753–1756. [Google Scholar] [CrossRef] [PubMed]
  15. Gheza, F.; Esposito, S.; Gruessner, S.; Mangano, A.; Fernandes, E.; Giulianotti, P.C. Reasons for Open Conversion in Robotic Liver Surgery: A Systematic Review with Pooled Analysis of More than 1000 Patients. Robot. Comput. Surg. 2019, 15, e1976. [Google Scholar] [CrossRef]
  16. Pilz Da Cunha, G.; Sijberden, J.P.; Gobardhan, P.; Lips, D.J.; Terkivatan, T.; Marsman, H.A.; Patijn, G.A.; Leclercq, W.K.G.; Bosscha, K.; Mieog, J.S.D.; et al. Risk Factors and Outcomes of Conversions in Robotic and Laparoscopic Liver Resections: A Nationwide Analysis. Surgery 2025, 178, 108820. [Google Scholar] [CrossRef]
  17. Camara, M.; Mayer, E.; Darzi, A.; Pratt, P. Soft Tissue Deformation for Surgical Simulation: A Position-Based Dynamics Approach. Int. J. Comput. Assist. Radiol. Surg. 2016, 11, 919–928. [Google Scholar] [CrossRef]
  18. Witheford, M.; Borghese, O.; Mastracci, T.M.; Maurel, B. An Observational Assessment of Aortic Deformation during Infrarenal and Complex Endovascular Aortic Aneurysm Repair. J. Vasc. Surg. 2022, 76, 645–655.e3. [Google Scholar] [CrossRef]
  19. Sotelo, R.; Nunez Bragayrac, L.A.; Machuca, V.; Garza Cortes, R.; Azhar, R.A. Avoiding and Managing Vascular Injury during Robotic-Assisted Radical Prostatectomy. Ther. Adv. Urol. 2015, 7, 41–48. [Google Scholar] [CrossRef]
  20. Danjouma, C.H.; Arfanis, I.; Gronostayskiy, M. Preventing and Managing Intraoperative Catastrophic Bleeding in Robotic Lung Cancer Lobectomy: “LifeSaving Strategies”. J. Surg. 2023, 8, 1947. [Google Scholar] [CrossRef]
  21. Jourdes, F.; Valentin, B.; Allard, J.; Duriez, C.; Seeliger, B. Visual Haptic Feedback for Training of Robotic Suturing. Front. Robot. AI 2022, 9, 800232. [Google Scholar] [CrossRef] [PubMed]
  22. Vangone, I.; Caruso, R.; Magon, A.; Giamblanco, C.; Conte, G.; Baroni, I.; Russo, S.; Belloni, S.; Arrigoni, C. High-Fidelity Simulation and Its Impact on Non-Technical Skills Development among Healthcare Professionals: A Systematic Review and Meta-Analysis. Acta Biomed. Atenei Parm. 2024, 95, e2024095. [Google Scholar] [CrossRef]
  23. Porterfield, J.R.; Podolsky, D.; Ballecer, C.; Coker, A.M.; Kudsi, O.Y.; Duffy, A.J.; Meara, M.P.; Novitsky, Y.W. Structured Resident Training in Robotic Surgery: Recommendations of the Robotic Surgery Education Working Group. J. Surg. Educ. 2024, 81, 9–16. [Google Scholar] [CrossRef] [PubMed]
  24. Gamberini, G.; Maglio, S.; Mariani, A.; Mazzotta, A.D.; Forgione, A.; Marescaux, J.; Melfi, F.; Tognarelli, S.; Menciassi, A. Design and Preliminary Validation of a High-Fidelity Vascular Simulator for Robot-Assisted Manipulation. Sci. Rep. 2024, 14, 4779. [Google Scholar] [CrossRef]
  25. Kim, Y.-H.; Marom, E.M.; Herndon, J.E.; McAdams, H.P. Pulmonary Vein Diameter, Cross-Sectional Area, and Shape: CT Analysis. Radiology 2005, 235, 43–49. [Google Scholar] [CrossRef]
  26. Rönkkö, M.; Cho, E. An Updated Guideline for Assessing Discriminant Validity. Organ. Res. Methods 2022, 25, 6–14. [Google Scholar] [CrossRef]
  27. Sparks, J.L.; Vavalle, N.A.; Kasting, K.E.; Long, B.; Tanaka, M.L.; Sanger, P.A.; Schnell, K.; Conner-Kerr, T.A. Use of Silicone Materials to Simulate Tissue Biomechanics as Related to Deep Tissue Injury. Adv. Ski. Wound Care 2015, 28, 59–68. [Google Scholar] [CrossRef]
  28. Faul, F.; Erdfelder, E.; Buchner, A.; Lang, A.-G. Statistical Power Analyses Using G*Power 3.1: Tests for Correlation and Regression Analyses. Behav. Res. Methods 2009, 41, 1149–1160. [Google Scholar] [CrossRef]
  29. Dutt, D.D.; Hohnen, H.; Kulshrestha, S.; Razavi, H. Simulation Models in Direct Ophthalmoscopy Education: A Systematic Review. Can. Med. Educ. J. 2025, 16, 51–65. [Google Scholar] [CrossRef]
  30. Yusoff, M.S.B. ABC of Response Process Validation and Face Validity Index Calculation. Educ. Med. J. 2019, 11, 55–61. [Google Scholar] [CrossRef]
  31. Harris, D.J.; Bird, J.M.; Smart, P.A.; Wilson, M.R.; Vine, S.J. A Framework for the Testing and Validation of Simulated Environments in Experimentation and Training. Front. Psychol. 2020, 11, 605. [Google Scholar] [CrossRef] [PubMed]
  32. Yusoff, M.S.B. ABC of Content Validation and Content Validity Index Calculation. Educ. Med. J. 2019, 11, 49–54. [Google Scholar] [CrossRef]
  33. Alsalamah, A.; Campo, R.; Tanos, V.; Grimbizis, G.; Van Belle, Y.; Hood, K.; Pugh, N.; Amso, N. Face and Content Validity of the Virtual Reality Simulator ‘ScanTrainer®’. Gynecol. Surg. 2017, 14, 18. [Google Scholar] [CrossRef] [PubMed]
  34. Vlachogianni, P.; Tselios, N. Perceived Usability Evaluation of Educational Technology Using the System Usability Scale (SUS): A Systematic Review. J. Res. Technol. Educ. 2022, 54, 392–409. [Google Scholar] [CrossRef]
  35. Malkewitz, C.P.; Schwall, P.; Meesters, C.; Hardt, J. Estimating Reliability: A Comparison of Cronbach’s α, McDonald’s Ωt and the Greatest Lower Bound. Soc. Sci. Humanit. Open 2023, 7, 100368. [Google Scholar] [CrossRef]
  36. Ali, H.; Salleh, M.N.M.; Hussain, K.; Ahmad, A.; Ullah, A.; Muhammad, A.; Naseem, R.; Khan, M. A Review on Data Preprocessing Methods for Class Imbalance Problem. Int. J. Eng. Technol. 2019, 8, 390–397. [Google Scholar]
  37. Hornos, E.; Pleguezuelos, E.; Bala, L.; Collares, C.F.; Freeman, A.; Van Der Vleuten, C.; Murphy, K.G.; Sam, A.H. Reliability, Validity and Acceptability of an Online Clinical Reasoning Simulator for Medical Students: An International Pilot. Med. Teach. 2024, 46, 1220–1227. [Google Scholar] [CrossRef]
  38. Amirthalingam, P.; Pakkir Mohamed, S.H.; Veeramani, V.P.; Nagoor Thangam, M.M.; Alanazi, M.F.; Dhanasekaran, M.; Mani, V.; Ali, M.A.S. The Effectiveness of a Structured Validated Questionnaire to Assess Student Perception with Virtual Pharmacy Simulation in Pharmacy Practice Experiential Education. PLoS ONE 2024, 19, e0314117. [Google Scholar] [CrossRef]
  39. García-Álvarez, J.M.; García-Sánchez, A.; Molina-Rodríguez, A.; Suárez-Cortés, M.; Díaz-Agea, J.L. Validation of the Group Environment Questionnaire (GEQ) in a Simulated Learning Environment. Nurs. Rep. 2025, 15, 154. [Google Scholar] [CrossRef]
  40. Peres, S.C.; Pham, T.; Phillips, R. Validation of the System Usability Scale (SUS): SUS in the Wild. Proc. Hum. Factors Ergon. Soc. Annu. Meet. 2013, 57, 192–196. [Google Scholar] [CrossRef]
  41. Crocker, J.C.; Farrar, N.; Cook, J.A.; Treweek, S.; Woolfall, K.; Chant, A.; Bostock, J.; Locock, L.; Rees, S.; Olszowski, S.; et al. Recruitment and Retention of Participants in UK Surgical Trials: Survey of Key Issues Reported by Trial Staff. BJS Open 2020, 4, 1238–1245. [Google Scholar] [CrossRef] [PubMed]
  42. Williams, J.; Craig, T.J.; Robson, D. Barriers and Facilitators of Clinician and Researcher Collaborations: A Qualitative Study. BMC Health Serv. Res. 2020, 20, 1126. [Google Scholar] [CrossRef] [PubMed]
  43. Hughes, T.; Fennelly, J.T.; Patel, R.; Baxter, J. The Validation of Surgical Simulators: A Technical Report on Current Validation Terminology as a Reference for Future Research. Cureus 2022, 14, e31881. [Google Scholar] [CrossRef] [PubMed]
  44. Arcieri, M.; Cuman, M.; Restaino, S.; Tius, V.; Cianci, S.; Ronsini, C.; Martinelli, C.; Bordin, F.; Pregnolato, S.; Di Donato, V.; et al. Exploring Urinary Tract Injuries in Gynecological Surgery: Current Insights and Future Directions. Healthcare 2025, 13, 1780. [Google Scholar] [CrossRef]
  45. Golahmadi, A.K.; Khan, D.Z.; Mylonas, G.P.; Marcus, H.J. Tool-Tissue Forces in Surgery: A Systematic Review. Ann. Med. Surg. 2021, 65, 102268. [Google Scholar] [CrossRef]
  46. Cundy, T.P.; Thangaraj, E.; Rafii-Tari, H.; Payne, C.J.; Azzie, G.; Sodergren, M.H.; Yang, G.-Z.; Darzi, A. Force-Sensing Enhanced Simulation Environment (ForSense) for Laparoscopic Surgery Training and Assessment. Surgery 2015, 157, 723–731. [Google Scholar] [CrossRef]
  47. Bogar, P.Z.; Virag, M.; Bene, M.; Hardi, P.; Matuz, A.; Schlegl, A.T.; Toth, L.; Molnar, F.; Nagy, B.; Rendeki, S.; et al. Validation of a Novel, Low-Fidelity Virtual Reality Simulator and an Artificial Intelligence Assessment Approach for Peg Transfer Laparoscopic Training. Sci. Rep. 2024, 14, 16702. [Google Scholar] [CrossRef]
  48. Raghavan, N.; Patel, P.C.; Limon, D.; Morris, M.X.; Kempenich, J.W.; Rajesh, A. Artificial Intelligence in Surgical Education: A 2025 Update on Adaptive Training, Feedback, and Competency-Based Education. Am. Surg. 2025, 00031348251397597. [Google Scholar] [CrossRef]
  49. Varas, J.; Coronel, B.V.; Villagrán, I.; Escalona, G.; Hernandez, R.; Schuit, G.; Durán, V.; Lagos-Villaseca, A.; Jarry, C.; Neyem, A.; et al. Innovations in Surgical Training: Exploring the Role of Artificial Intelligence and Large Language Models (LLM). Rev. Colégio Bras. Cir. 2023, 50, e20233605. [Google Scholar] [CrossRef]
  50. Lee, F.; Wong, S.W. Surgical Education Reimagined: The Convergence of Learning Theories and Artificial Intelligence. BMC Med. Educ. 2025, 26, 38. [Google Scholar] [CrossRef]
  51. Kowalewski, K.-F.; Hendrie, J.D.; Schmidt, M.W.; Garrow, C.R.; Bruckner, T.; Proctor, T.; Paul, S.; Adigüzel, D.; Bodenstedt, S.; Erben, A.; et al. Development and Validation of a Sensor- and Expert Model-Based Training System for Laparoscopic Surgery: The iSurgeon. Surg. Endosc. 2017, 31, 2155–2165. [Google Scholar] [CrossRef]
  52. D’Angelo, A.-L.D.; Rutherford, D.N.; Ray, R.D.; Laufer, S.; Mason, A.; Pugh, C.M. Working Volume: Validity Evidence for a Motion-Based Metric of Surgical Efficiency. Am. J. Surg. 2016, 211, 445–450. [Google Scholar] [CrossRef][Green Version]
Figure 1. The simulator in the operating room during a testing session with all the system components and the surgical instruments used for the tests.
Figure 1. The simulator in the operating room during a testing session with all the system components and the surgical instruments used for the tests.
Jcm 15 01054 g001
Figure 2. Comparison of the outcome measures between the three groups of experience (n: novice, f: fellow, e: expert). (A) Comparison of the maximum value of the deformation signal during the first repetition between the three groups of experience. (B) Comparison of the mean value of the deformation signal during the second repetition between the three groups of experience.
Figure 2. Comparison of the outcome measures between the three groups of experience (n: novice, f: fellow, e: expert). (A) Comparison of the maximum value of the deformation signal during the first repetition between the three groups of experience. (B) Comparison of the mean value of the deformation signal during the second repetition between the three groups of experience.
Jcm 15 01054 g002
Figure 3. Specialties comparison of the outcome measure. The comparison between all four types of specialties (gen 2: general surgery 2, u: urology, g: gynecology, t: thoracic surgery) is reported on the top part of the figure together with its p-value. In the bottom part of the figure, the couple’s testing results having a statistically significant difference are reported.
Figure 3. Specialties comparison of the outcome measure. The comparison between all four types of specialties (gen 2: general surgery 2, u: urology, g: gynecology, t: thoracic surgery) is reported on the top part of the figure together with its p-value. In the bottom part of the figure, the couple’s testing results having a statistically significant difference are reported.
Jcm 15 01054 g003
Figure 4. (A) Example of deformation curve. The light-blue line is the deformation over time, while the filled space below is the area under the curve considered. (B) Boxplot showing the area under the curve distribution for each surgical specialty.
Figure 4. (A) Example of deformation curve. The light-blue line is the deformation over time, while the filled space below is the area under the curve considered. (B) Boxplot showing the area under the curve distribution for each surgical specialty.
Jcm 15 01054 g004
Figure 5. Comparison between the different levels of expertise in general surgery (n: novice, f: fellow, e: expert). (A) Comparison of the maximum value of the deformation signal between the three groups of experience during the first repetition. (B) Comparison of the mean value of the deformation signal between the three groups of experience during the second repetition.
Figure 5. Comparison between the different levels of expertise in general surgery (n: novice, f: fellow, e: expert). (A) Comparison of the maximum value of the deformation signal between the three groups of experience during the first repetition. (B) Comparison of the mean value of the deformation signal between the three groups of experience during the second repetition.
Jcm 15 01054 g005
Figure 6. Questionnaire results regarding the face- and content validities and the system usability scale. The questions left unanswered were counted as “neither unrealistic nor realistic” or “neither agree nor disagree”.
Figure 6. Questionnaire results regarding the face- and content validities and the system usability scale. The questions left unanswered were counted as “neither unrealistic nor realistic” or “neither agree nor disagree”.
Jcm 15 01054 g006
Table 1. Post-testing questionnaire.
Table 1. Post-testing questionnaire.
ItemsQuestion or Statement
FACE VALIDITY
Item 1Overall global impression of the simulator
Item 2Anatomical realism without cover with the adipose/connective tissue
Item 3Anatomical realism with cover with the adipose/connective tissue
Item 4Visual appearance of the vein
Item 5Visual appearance of the connective/adipose tissue
Item 6Haptic feedback of the vein
Item 7Haptic feedback of the connective/adipose tissue
Item 8Realism in instrument—vein interaction
Item 9Realism in instrument—tissue interaction
CONTENT VALIDITY
Item 10The simulator is useful in teaching vascular-structure isolation
Item 11The simulator is useful in teaching the stapling of vascular structure
Item 12The simulator is useful in teaching to minimize the forces applied to vascular structures
SYSTEM USABILITY SCALE
Item 13I think I would like to use this platform for training
Item 14I found the functioning of the simulator more complex than what I was thinking
Item 15I think the simulator is easy to use
Item 16I think I will need a technical person to support me to be able to use the simulator
Item 17I found the interface functionalities understandable and clear
Item 18I found difficulties in understanding the physical links to set up and use the simulator
Item 19I imagine that most of my colleagues will learn how to use the simulator very fast
Item 20I found the system intuitive and “plug and play”
Item 21I was comfortable with the use of the simulator
Item 22I think more time is needed to get familiar with the setup of the simulator
Table 2. Characteristics of the three groups. () = %, [] = age range, F = Female, M = Male, VR = Virtual Reality.
Table 2. Characteristics of the three groups. () = %, [] = age range, F = Female, M = Male, VR = Virtual Reality.
Novice (n = 13)Fellow (n = 8)Expert (n = 9)
Age (years)[29–51][26–55][26–64]
Sex M972
Sex F417
Right dominant hand1379
Specialty13 (100)8 (100)9 (100)
General surgery9 (69)5 (63)7 (77)
Thoracic surgery1 (50)1 (50)0 (0)
Gynecology1 (25)1 (25)2 (50)
Urology2 (66.6)1 (33.4)0 (0)
Simulation training9 (69)7 (88)9 (100)
VR simulation3 (33.3)3 (43)2 (22)
Physical simulation3 (33.3)1 (14)1 (11)
Both3 (33.3)3 (43)6 (67)
Table 3. Mean and standard deviation for each item from 1 to 9 of the face validity or evidence based on response processes validity.
Table 3. Mean and standard deviation for each item from 1 to 9 of the face validity or evidence based on response processes validity.
Item 1Item 2Item 3Item 4Item 5Item 6Item 7Item 8Item 9
Mean4.413.974.283.974.163.944.064.104.35
SD0.560.690.630.740.920.730.850.660.71
Table 4. Mean and standard deviation for each item from 10 to 12 of the content validity.
Table 4. Mean and standard deviation for each item from 10 to 12 of the content validity.
Item 10Item 11Item 12
Mean4.334.234.20
SD0.610.680.55
Table 5. Mean and standard deviation for each item from 13 to 22 of the system usability scale.
Table 5. Mean and standard deviation for each item from 13 to 22 of the system usability scale.
Item 13Item 14Item 15Item 16Item 17Item 18Item 19Item 20Item 21Item 22
Mean4.372.074.032.504.232.404.274.404.702.50
SD0.811.111.001.290.901.300.830.670.471.14
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gamberini, G.; Mazzotta, A.D.; Durante, A.; Tognarelli, S.; Petrucciani, N.; Mennini, G.; Silecchia, G.; Menciassi, A. Sensorized Vascular High-Fidelity Physical Simulator for Robot-Assisted Surgery Training: A Multisite Pilot Evaluation. J. Clin. Med. 2026, 15, 1054. https://doi.org/10.3390/jcm15031054

AMA Style

Gamberini G, Mazzotta AD, Durante A, Tognarelli S, Petrucciani N, Mennini G, Silecchia G, Menciassi A. Sensorized Vascular High-Fidelity Physical Simulator for Robot-Assisted Surgery Training: A Multisite Pilot Evaluation. Journal of Clinical Medicine. 2026; 15(3):1054. https://doi.org/10.3390/jcm15031054

Chicago/Turabian Style

Gamberini, Giulia, Alessandro Dario Mazzotta, Angela Durante, Selene Tognarelli, Niccolò Petrucciani, Gianluca Mennini, Gianfranco Silecchia, and Arianna Menciassi. 2026. "Sensorized Vascular High-Fidelity Physical Simulator for Robot-Assisted Surgery Training: A Multisite Pilot Evaluation" Journal of Clinical Medicine 15, no. 3: 1054. https://doi.org/10.3390/jcm15031054

APA Style

Gamberini, G., Mazzotta, A. D., Durante, A., Tognarelli, S., Petrucciani, N., Mennini, G., Silecchia, G., & Menciassi, A. (2026). Sensorized Vascular High-Fidelity Physical Simulator for Robot-Assisted Surgery Training: A Multisite Pilot Evaluation. Journal of Clinical Medicine, 15(3), 1054. https://doi.org/10.3390/jcm15031054

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop