You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

13 November 2025

Advancing Medical Training with Mixed Reality and Haptic Feedback Simulator for Acupuncture Needling

,
,
,
,
,
and
1
Industry 4.0 Convergence Bionics Engineering, Pukyong National University, Busan 48513, Republic of Korea
2
Digital Healthcare Research Center, Pukyong National University, Busan 48513, Republic of Korea
3
College of Korean Medicine, Kyung Hee University, Seoul 02453, Republic of Korea
4
College of Korean Medicine, Dongshin University, Jeonnam 58245, Republic of Korea
Sensors2025, 25(22), 6934;https://doi.org/10.3390/s25226934 
(registering DOI)
This article belongs to the Section Sensors and Robotics

Abstract

Traditional acupuncture training often lacks consistent, objective feedback, while current extended reality (XR) solutions rarely include quantitative assessment. This study developed and evaluated a feedback-enabled mixed reality (MR) acupuncture simulator to improve skill acquisition through depth-responsive guidance. The system, used on Microsoft HoloLens 2, combines a MetaHuman-based virtual patient with expert-designed acupoint geometries. It provides depth-dependent vibrotactile cues via a wearable haptic device and calculates a composite score from normalized metrics, including insertion depth, angular deviation, tip-to-center distance, and task duration. Ten participants (eight novices and two experts) performed needle tasks at LI4, LI11, and TE3 across two sessions. Mean depth error decreased from 6.41 mm to 3.58 mm, and task time from 9.29 s to 6.83 s. At LI11, beginners improved in achieved depth (16.24 ± 1.88 mm to 19.74 ± 1.23 mm), reduced angular deviation (27.83° to 15.34°), and shortened completion time (38.77 s to 13.28 s). Experts outperformed novices (69.25 ± 21.64 vs. 56.26 ± 23.37), confirming construct validity. Usability evaluation showed a mean overall score of 4.46 ± 0.51 and excellent reliability (McDonald’s ω = 0.93). These results demonstrate that expert-informed scoring and depth-responsive haptic feedback substantially enhance accuracy, efficiency, and learning confidence, validating the system’s technical robustness and educational readiness for clinical acupuncture training.

1. Introduction

Acupuncture, a key therapeutic modality in Traditional Korean Medicine (TKM), has demonstrated clinical efficacy in managing pain and a range of systemic disorders. However, mastering acupuncture requires precise localization of meridians and acupoints, controlled insertion depth, and accurate needle manipulation [,]. Even minor technical deviations can lead to complications, such as tissue damage or organ injury, underscoring the importance of safe and structured training environments. Conventional teaching methods, such as lectures, 2D anatomical atlases, and peer demonstrations, offer a limited three-dimensional understanding and inconsistent feedback. Silicone or gel-based phantoms partially mitigate safety concerns but lack anatomical fidelity and haptic realism, while assessment remains largely observational and subjective [,]. These limitations underscore the need for immersive, measurable, and feedback-driven approaches to procedural training in acupuncture.
Recent advances in XR, encompassing augmented reality (AR), virtual reality (VR), and MR, have transformed medical education by enabling realistic and interactive simulations, as illustrated in Figure 1. In this study, MR refers explicitly to the interactive holographic environment implemented on Microsoft HoloLens 2, where virtual content is spatially anchored to the real world. The term XR is used more broadly to denote the collective domain encompassing VR, AR, and MR technologies.
Figure 1. Overview of the existing XR Approaches for Medical Training.
Previous studies by Sung et al. [], Kim et al. [], Liu et al. [], Jongbloed et al. [], and Mitchell et al. [] consistently demonstrated that XR enhances knowledge retention, procedural accuracy, and user confidence across various healthcare disciplines, including nursing, anesthesiology, and surgery. According to Han et al. [], Fleet et al. [], and Rubin et al. [], XR facilitates deliberate practice, real-time feedback, and safe error correction, aligning with evidence-based learning principles. Among head-mounted displays, the Microsoft HoloLens 2 has been widely adopted for clinical simulation due to its spatial mapping capabilities, gesture tracking, and ability to integrate virtual content into real environments []. In surgical and anesthesia education, validated frameworks such as OSATS and GOALS have established the value of quantitative metrics, including task time, angular precision, and trajectory smoothness, in evaluating procedural competence [,,,,]. These successes demonstrate the potential of XR for complex, precision-dependent procedures such as acupuncture.
XR-based training for needle-guided interventions has expanded rapidly. MR lumbar-puncture simulators using force feedback have produced measurable gains in depth control, confidence, and learning efficiency [,]. Similarly, VR-based lumbar puncture and spinal anesthesia systems that combine tactile cues with immersive visualization have improved procedural accuracy and user engagement [,,,,,]. MR-enhanced neurosurgical and neuraxial anesthesia trainers, which integrate holographic overlays and trajectory guidance, have been shown to improve spatial orientation and reduce task time in studies by Agathe et al. [], Peng et al. [], and Tanwani et al. []. Furthermore, in [,,,,,,], the value of XR in reproducing realistic psychomotor demands was highlighted, particularly when paired with haptic feedback that reinforces depth awareness and tissue resistance. Despite such progress, most implementations focus on high-risk interventions (e.g., lumbar puncture, catheter placement) rather than on acupuncture, which demands subtler tactile perception and fine motor control.
By contrast, XR-based acupuncture systems remain in early development. Sun et al. [] introduced an MR visualization platform that enables acupoint exploration via HoloLens; however, it lacks interactivity and assessment capabilities. Several investigations have examined educational and cognitive dimensions, for example, EEG-integrated VR environments for acupoint selection by Liao et al. [], and AR-based mobile applications for anatomical recall by Zhang et al. [], but these approaches provide limited procedural realism. Likewise, according to Chiou et al. [], static AR acupoint tables and computer vision-based localization systems, as described by Yang et al. [], enhance visualization but fail to facilitate hands-on training. Notably, nearly all prior acupuncture–XR systems omit haptic feedback, a component essential for realistic depth calibration and for conveying tissue penetration. Likewise, few incorporate quantitative, multi-parameter scoring, instead relying on post hoc instructor evaluation. In contrast, contemporary surgical simulators employ automated, objective scoring to deliver immediate, data-driven feedback [,,]. This discrepancy highlights a critical gap between acupuncture education and other procedural domains.
Evidence from broader XR literature suggests that effective simulation platforms share three defining characteristics: (1) Task-specific, clinically grounded metrics, (2) Multidimensional performance assessment, and (3) Real-time, actionable feedback [,,,]. Translating these principles to acupuncture training requires a synchronized evaluation of insertion depth, angular alignment, lateral targeting, and temporal control elements that are rarely addressed in existing systems. Moreover, integrating tactile cues with immersive visual feedback could substantially enhance spatial memory and proprioceptive calibration during the insertion of needles.
Motivated by these gaps, this study proposes a MR-acupuncture training system that unifies expert-informed assessment with real-time haptic feedback to enhance procedural competence. Developed in Unreal Engine (UE) 5 and deployed on Microsoft HoloLens 2, the platform features a life-sized MetaHuman avatar, expert-derived acupoint models, and a wearable ESP32-based vibrotactile device providing depth-responsive vibration feedback. The system continuously captures key performance parameters, including needle insertion depth, angular deviation, tip-to-center distance, and task duration, and then synthesizes them into a composite, expert-weighted score. Unlike conventional models that rely solely on visual guidance, the proposed design enables multisensory feedback, reinforcing depth estimation through vibration intensity that is proportional to the depth of insertion.
The system incorporates two progressive training modes. The Beginner Mode displays visible acupoint markers to guide novices in spatial orientation, whereas the Advanced Mode conceals these cues, compelling reliance on anatomical recall and haptic sensation. Immediate visual and numerical feedback following each trial supports iterative self-correction, which aligns with the principles of deliberate practice pedagogy. The combination of visual immersion, tactile realism, and quantitative evaluation enables a closed-loop learning environment that mirrors the cognitive and sensory demands of authentic acupuncture procedures.
The novel contributions of this study are fourfold. First, it introduces a full-scale MR-compatible acupuncture simulator that integrates realistic virtual anatomy with natural hand interaction, providing lifelike procedural visualization that mirrors clinical needling contexts. Second, it incorporates expert-validated acupoint geometries and tolerance parameters to establish a robust foundation for quantitative, reproducible performance evaluation. Third, it presents a wearable haptic feedback interface that enhances depth perception through continuous, real-time vibration, enabling users to gauge insertion depth and prevent over-penetration intuitively. Fourth, it integrates a multi-parameter, expert-weighted scoring algorithm to objectively assess needling precision across spatial, angular, and temporal dimensions. Collectively, these features advance beyond prior XR-based acupuncture prototypes by combining high anatomical realism, quantitative feedback, and tactile guidance within a unified MR training framework.
By situating acupuncture within the broader framework of simulation-based medical education, this research advances the pedagogical shift toward quantitative, feedback-rich, and multisensory learning environments. The proposed system addresses not only the absence of tactile realism and objective scoring in previous XR acupuncture applications but also demonstrates the feasibility of combining expert knowledge, real-time analytics, and embodied interaction within a single, integrated framework. Consequently, it offers a promising foundation for standardized, scalable acupuncture training that aligns with contemporary trends in immersive healthcare education and competency-based assessment.

2. Methodology

This section outlines the system architecture, hardware–software integration, acupoint modeling strategy, data capture mechanism, expert-informed scoring algorithm, and user evaluation mechanism of the system. The entire application was developed in UE 5.0 and deployed on the Microsoft HoloLens 2 platform. Core components include a high-fidelity virtual patient model (MetaHuman), a real-scale virtual needle, anatomically grounded acupoint markers, a custom-designed wearable haptic feedback device, and a modular performance evaluation pipeline, as illustrated in Figure 2.
Figure 2. System architecture diagram showing the HoloLens 2 and Haptic device hardware pipeline, MR environment, MRTK integration, MetaHuman and interaction subsystems, and real-time communication with the wearable haptic device for depth-based vibration feedback.

2.1. Mixed Reality Application Design and Development

The system architecture comprises a MetaHuman-based virtual patient, a real-scale interactive needle, expert-defined acupoint models, and a wearable ESP-32 C3-driven vibrotactile device connected via TCP at a rate of 60 Hz. Together, these components create a synchronized visual-haptic training environment for procedural skill acquisition.
The HoloLens 2 offers inside-out spatial mapping and articulated hand tracking, allowing for natural, bare-hand interaction without the need for external controllers. The virtual clinical scene, consisting of a bed and side table, is spatially anchored to the real environment, allowing the MetaHuman avatar to appear at life size and preserving realistic ergonomics for the learner, as shown in Figure 3A.
Figure 3. The MR acupuncture training environment and interaction components. (A) Complete virtual clinical setup spatially anchored to the real world, (B) Virtual acupuncture needle positioned on the bedside table for natural reach and grasp, (C) Bare-hand interaction view showing how users see and manipulate the virtual needle, and (D) Floating user interface panels.
The virtual needle (see Figure 3B) was modeled as a slender 110 mm × 1 mm cylinder, approximating the dimensions of a clinical instrument. Gesture recognition via MRTK enables grasping, insertion, and release actions with the bare hand. As shown in Figure 3C, Users can freely approach and manipulate the virtual needle using hand gestures, replicating the posture of an authentic operator. To support usability, graphical interface panels (see Figure 3D) provide (1) Step-by-step task instructions, (2) Acupoint selection and session controls, and (3) Real-time performance feedback showing measured parameters and overall scores. These floating UI elements maintain visibility without obstructing the operative field.
Acupoint modeling was derived from structured questionnaires completed by three licensed practitioners, defining safe depth ranges, insertion angles, and tolerance margins for 18 clinically relevant acupoints. Expert-defined parameters, summarized in Table 1, establish clinically realistic bounds for evaluation. The marker’s central axis denotes the ideal insertion path, and deviations beyond the defined tolerances are automatically penalized during scoring. Each acupoint was represented as a cylindrical marker embedded beneath the MetaHuman skin surface (see Figure 4). The visible top view of the cylinder corresponds to the traditional surface marking, while the internal cylinder encodes permissible insertion depth and angular trajectory.
Table 1. Expert-derived marker dimensions and clinical parameters.
Figure 4. Cylindrical acupoint model representation. The top surface of the cylinder (top view) is aligned with the MetaHuman skin, while the remainder of the cylinder extends internally. The visible circle serves as surface guidance, and the central axis defines the ideal trajectory and optimal insertion depth.
A collision-based detection model in UE governs system interaction. When the user releases the virtual needle, the system classifies the event as (1) Unnecessary release (see Figure 5A), (2) Incorrect insertion, such as skin contact without acupoint intersection (see Figure 5B), or (3) Valid attempt intersecting an acupoint marker (see Figure 5C). According to Algorithm 1, only valid attempts trigger quantitative evaluation. For these cases, four parameters were automatically recorded: insertion depth, angular deviation, tip-to-center distance, and insertion time. These data served as inputs to the expert-informed scoring model as described in Section 2.3. The deliberate design choice to evaluate performance only upon release mirrors real-world practice, where assessment occurs after the needle is positioned.
Algorithm 1 Needle Release Event Handling and Data Capture
Require: Needle collider CN, release event R
Ensure:   Attempt classification C, performance metrics {D, A, T, IT}
 1: b1 ← CollideSurface (CN) ▷ collision with any surface?
 2: if ¬ b1 then
 3:  C ← “Unnecessary Release”
 4:  return (C, )
 5: b2 ← CollideSkin (CN) ▷ collision with skin?
 6: b3 ← CollideAcupoint (CN) ▷ intersects acupoint marker?
 7: if b2 ∧ ¬ b3 then
 8:  C ← “Wrong Needling Attempt”
 9:  return (C, )
 10: if b3 then
 11:  C ← “Proper Needling Attempt”
 12:  D ← ComputeDepth (Sskin, ptip)
 13:  A ← ComputeAngle ( a needle, a ideal)
 14:  T ← ComputeOffset (ptip, pap(D))
 15:  IT← ComputeDuration (tgrab, trelease)
 16:  RecordMetrics (D, A, T, IT)
 17:  TriggerScoring (D, A, T, IT)
 18:  return (C, {D, A, T, IT})
Figure 5. System classification of needle insertion outcomes and training mode visuals. (A) Unnecessary release, (B) Incorrect insertion, (C) Valid insertion, and (D) Advanced Mode visualization, which features hidden acupoint markers.
To support adaptive learning, the platform incorporates two training modes. In (1) Beginner Mode, Acupoint markers were visibly rendered on the MetaHuman surface to guide spatial localization and provide immediate visual feedback (see Figure 5C). In (2) Advanced Mode, markers were hidden (see Figure 5D), requiring reliance on anatomical memory and haptic cues from the wearable device to judge insertion depth. Regardless of mode, the system displays post-trial metrics and a composite performance score, enabling self-correction through iterative practice.
The wearable haptic device (discussed in Section 2.2) operates in tandem with the virtual system to provide depth-contingent vibration feedback during needle insertion. Accordingly, the system continuously calculates a real-time penetration depth ( d h a p ) whenever the virtual needle intersects the MetaHuman skin surface. This depth value represents the instantaneous tip-to-skin distance along the acupoint’s insertion axis and is used solely for generating haptic feedback, independent of the performance-scoring framework. When the needle is not overlapping with the MetaHuman or not in the grabbed state, d h a p is set to 0, ensuring that no vibration occurs during idle periods. The value is updated in real time at the HoloLens frame rate of 60 Hz. For each acupoint, the expert-defined maximum safe depth ( D m a x ) from Table 1 serves as a safety reference. The current d h a p value and its corresponding D m a x are combined into an HTTP string and transmitted to the ESP32-C3 microcontroller via TCP communication, with the microcontroller acting as a server that continuously listens for client updates. The VaREST plugin in Unreal Engine processes the URL request using the GET method. Upon receiving these values, the haptic module produces light vibration pulses for shallow insertion, progressively stronger pulses for deeper penetration, and a distinct buzz if d h a p exceeds D m a x . In contrast, the final insertion depth recorded at the needle-release event is separately used in the quantitative scoring model described in Section 2.3. This clear separation between real-time tactile feedback and post-trial scoring ensures consistent sensory guidance without interfering with objective evaluation.
Overall, the system establishes a closed-loop, sensor-based training framework in which visual, tactile, and quantitative feedback operate synergistically. By grounding the virtual geometry in expert-validated anatomical data and coupling it with real-time depth feedback, the design ensures that users can practice accurate needle placement under clinically meaningful constraints. The resultant platform supports consistent, measurable, and immersive acupuncture training aligned with contemporary simulation-based education principles.

2.2. Haptic Feedback System Development and Integration

In conventional MR interactions, users rely solely on visual feedback, as virtual objects provide no physical resistance or tactile sensation. During acupuncture simulation, the absence of haptic feedback makes it difficult to estimate insertion depth or perceive contact with the skin, especially when manipulating virtual instruments, such as needles. To address this limitation, a haptic feedback system was incorporated to recreate the sense of touch that was missing. When the virtual needle tip collides with the MetaHuman’s skin, the wearable device delivers a vibration pulse corresponding to the contact event, allowing users to sense the interaction physically. As the needle penetrates deeper, the vibration intensity increases progressively, providing depth-aware tactile cues that help users align their visual perception with the expected physical sensation of insertion.
The methodology for the virtual-needling haptic device centered on a compact, wearable implementation available in two interchangeable form factors, a finger ring and a wrist-band module, driven by an ESP32-C3 Super Mini microcontroller and a 3 V eccentric-rotating-mass (ERM) vibration motor as the tactile actuator. To achieve this, we designed a self-contained electronics stack that preserves access to the USB-C port and header pins, while arranging the controller, charging board, battery, and motor in a low-profile enclosure (see Figure 6A–C). The actuator is housed in a cylindrical sleeve and secured to the fingertip with an open-C, size-adjustable ring that maintains skin contact for efficient vibrotactile coupling during motion (see Figure 6D). The housing was designed in CAD to position the electronics, battery, and motor, and to provide reliable cable relief and motor seating. These components were fabricated using a FormLabs SLA printer. To balance stiffness, precision, and translucency, the base and circuit holder were printed in FormLabs Black resin, chosen for its dimensional stability, which is suitable for snap-fits and threaded inserts. Meanwhile, the protective enclosure was printed in Clear resin, allowing for visual inspection of the electronics and battery without disassembly. The ring itself was also printed in Black resin to maximize durability under bending and repeated donning.
Figure 6. Hardware design and integration of the wearable haptic feedback device. (A) Internal components include the ESP32-C3 Mini microcontroller, a 3.9 V lithium-polymer battery, and a 3 V micro ERM vibration motor. (B) USB-C charging module integrated into the compact PCB frame. (C) Fully assembled wrist-worn enclosure with transparent outer casing. (D) Haptic ring housing the vibration motor, worn on the index finger and connected to the wristband via a thin cable.
The electronics were selected to minimize footprint and simplify charging. A single-cell 3.7 V lithium-ion polymer battery powered a TP4057-based USB-C charging and protection module; the module’s OUT terminals provided a protected rail for the ESP32-C3 Super Mini and the motor driver stage. The ESP32-C3 was chosen for its integrated Wi-Fi and compact form factor, and it was used to synthesize the vibration drive signals from HTTP responses generated by the HoloLens.
During operation, a virtual simulation computed the instantaneous needle penetration depth (d) and the acupoint-specific maximum depth (Dmax). These quantities were transmitted from the virtual system to the wearable device; both systems were associated with the same access point to minimize routing overhead and reduce latency. Upon receipt of the HTTP payload, bounds were validated, and depth was mapped to a target vibration frequency via a linear law (Equation (1)):
f d = f m a x D m a x · d , w h e r e   0 d D m a x ,
In practice, the vibrotactile motor operated within an effective frequency range of 5–200 Hz, corresponding to low frequency for shallow insertion and progressively stronger vibration at greater depths. These parameters were determined empirically to ensure perceptible yet comfortable stimulation across all users while maintaining low power consumption. This scaling ensured that zero penetration yields no vibration, whereas full-scale penetration at a given acupoint produced a consistent maximum frequency fmax, independent of the absolute value of Dmax.
As shown in the logic pipelines of Algorithm 2, the firmware employed a rate-limited controller that filters noisy depth updates and adjusts the vibration frequency to prevent perceptible stepping. A moderate update cadence preserved natural haptic continuity while avoiding excessive network traffic. On the hardware side, the ESP32-C3 executed a command to drive an ERM motor. For the ring configuration, the motor was housed in a cylindrical cradle to maximize mechanical coupling to the finger pad.
Algorithm 2 Vibration Pulse Generation
Require: constants Dmax, fmax
Ensure:   vibration at f = (fmax/Dmax) · d when enabled; otherwise LOW
  1:   d ← 0; f ← 0; enabled ← false
  2:   outputHigh ← false; Δthalf ← 0; tnext ← 0
  3:   procedure Recalc
  4:  f ← max(0, (fmax/Dmax) · d)
  5:  if ¬ enabled ∨ f = 0 then
  6:    Δthalf ← 0; outputHigh ← false; DigitalWriteLow()
  7:    return
  8:  Δthalf ← 1/(2f); tnext ← Now() + Δthalf; return
  9:   procedure Start
  10:  enabled ← true; Recalc
  11: procedure Stop
  12:  enabled ← false; Recalc
  13: procedure SetDisplacement(dnew)
  14:  d ← min(max(0, dnew), Dmax); Recalc
  15: function MainLoop
  16:  HandleHTTPRequests()
  17:  if enabled ∧ Δthalf > 0 ∧ Now() ≥ tnext then
  18:   outputHigh ← ¬ outputHigh
  19:   if outputHigh then
  20:    DigitalWriteHigh()
  21:   else
  22:    DigitalWriteLow()
  23:   tnext ← tnext + Δthalf
  24:  return          ▷ loop MainLoop continuously
The final wearable designs comprise a finger-mounted vibration ring and a wireless wristband that enable unobstructed, freehand hologram interaction and the reliable delivery of depth-responsive tactile feedback, as illustrated in Figure 7.
Figure 7. Final design of a wearable haptic feedback device. The ring configuration features a vibration motor on the fingertip for high tactile precision. At the same time, the wristband variant positions the motor over the radial or ulnar side to enhance comfort and stability during needle insertion tasks.

2.3. Expert-Driven Scoring Model

The acupuncture training performance evaluation system was developed based on the expert-defined tolerances collected through the questionnaire study. For each acupoint, the experts specified a target insertion depth and an acceptable tolerance range, the ideal insertion angle with maximum allowable deviation, and the maximum lateral offset of the needle tip from the acupoint center. These values provided clinically grounded thresholds that guided the formulation of a multi-parameter scoring model (see Table 1).
The scoring framework comprises four normalized component scores, each scaled between 0 and 100, which together capture the multidimensional nature of acupuncture needling performance: depth accuracy, angular precision, spatial targeting, and temporal control.
1.
Depth Score: Depth accuracy was modeled as
S d =     100 ,     d u d t a r g e t     d t o l 100 · 1 d u d t a r g e t d t o l d m a x d m i n ,       W i t h i n   b o u n d s 0 ,     o t h e r w i s e ,
where du is the user’s insertion depth, dtarget is the ideal depth, dtol is the expert-defined tolerance, and dmin, dmax are the lower and upper bounds of safe insertion. Here, the expert-defined tolerance (dtol = ±2 mm) represents the clinically acceptable deviation from the target depth. The minimum (dmin) and maximum (dmax) depths define the absolute safe boundaries based on anatomical constraints. During scoring, depth errors within ±dtol are rewarded with full points, whereas insertions beyond this range but within [dmin, dmax] incur proportional penalties. Insertions exceeding dmax are automatically scored as 0 to reflect unsafe over-penetration.
2.
Angle Score: Angular deviation from the ideal trajectory was calculated as
S α =     100 ,       θ   θ t o l 100 · 1 θ θ t o l 45 θ t o l ,       θ   45 ° 0 ,     θ > 45 °
where θ is the deviation from the ideal angle, and θtol is the expert-defined angular tolerance.
3.
Weighted Accuracy: Since experts indicated that depth and angle should carry equal importance for every acupoint, a weighted accuracy score was computed as
A = ω d S d + ω α S α ,
with default weights ωd = ωα = 0.5.
4.
Tip Position Score: Spatial targeting was assessed based on the Euclidean distance of the needle tip from the acupoint center, compared to the permissible radius (R = 1.5 mm).
P t = 1.00 ,     r   R 0.75 ,     r   2 R 0.50 ,   r   3 R 0.25 ,   r > 3 R ,
5.
Time Score: The temporal performance was scored against expert-recommended insertion times, with an optimal range of 10–20 s and a maximum acceptable limit of 60 s.
S t = 1.00 ,     t < 20   s 0.75 ,     20   s   t < 30   s 0.50 ,     30   s   t < 40   s 0.25 ,     40 t < 60   s 0.0 ,     t   60   s ,
6.
Final Score. The overall performance score was computed as
A f i n a l = A · S t + P t 2
This formulation integrates weighted accuracy (depth and angle) with temporal and spatial control, producing a balanced measure of needling competence. The complete penalty functions for depth, angle, tip location, and timing are visualized in Figure 8.
Figure 8. Visualization of the schematic of composite score calculation parameters.

2.4. Usability Experiment Design

The primary objective of the user experiment was to demonstrate that the MR acupuncture needling trainer, integrated with a depth-responsive haptic device, is usable and capable of producing measurable learning effects within short, repeated sessions. This study focused on system-level feasibility, usability, and metric fidelity, rather than on evaluating clinical skill transfer to acupuncture trainees. Accordingly, the evaluation assessed user progress and interaction performance within the fully integrated MR–haptic environment, to validate the complete training system under realistic operating conditions.

2.4.1. Participants

A total of ten participants were recruited (Table 2). Eight were beginners without prior experience in acupuncture needling, but they were familiar with the theoretical background through books or demonstration videos. Two were experts, defined as medical students with prior exposure to acupuncture practice in clinical or educational settings. All participants were well-experienced in using MR headsets and HoloLens 2, which allowed the study to focus on interaction with the needling trainer rather than acclimatization to immersive technology. Recruiting MR-familiar but non-clinical users was appropriate for this early-phase evaluation, as the aim was to establish usability and data reliability. This approach aligns with prior VR/MR healthcare-related studies, which first assess feasibility using convenience samples before moving to medical learners. This initial usability study was therefore performed with MR-familiar participants to isolate system-interaction factors; however, future validation will target medical trainees and clinical learners to evaluate educational transfer and curricular integration.
Table 2. Participant Demographics.

2.4.2. Experimental Design and Tasks

The experimental protocol consisted of two sessions separated by 24 h. Each participant was provided 30–60 min of free practice to familiarize themselves with the virtual patient, acupoint markers, and haptic feedback. The haptic depth feedback was kept active in all sessions to maintain uniform experimental conditions and to evaluate user learning within the complete MR–haptic training environment rather than to isolate the independent effect of the haptic feedback module. In each session, participants performed ten consecutive needling attempts at three commonly used training acupoints: TE3, LI4, and LI11, which represent shallow, medium, and deep insertion depths, respectively. This produced 30 trials per session and 60 trials per participant across both sessions, yielding 20 repeated attempts per acupoint. An overview of the design, including participants, acupoints, and trial counts, is summarized in Table 3.
Table 3. Experiment Design Overview for all participants.
The system automatically logged each attempt. For valid insertions, the following metrics were recorded: insertion depth (mm), needle angle (°), tip-to-center distance (mm), insertion time (s), and the composite weighted scores derived from expert input. Errors, including needle drops and wrong insertions during a single trial, were also recorded. Additional measures included preparation time from the start of the session to the first needle grasp and total trial duration.

2.4.3. Usability Questionnaire

Upon completion of both training modes, all participants were administered a structured usability questionnaire designed to evaluate multiple dimensions of user experience with the MR acupuncture training system. The instrument was an adapted version of the USE questionnaire, comprising 15 items rated on a 5-point Likert scale (1 = Strongly Disagree, 5 = Strongly Agree). Items were grouped a priori into six theoretical factors: Usefulness (UF), Ease of Use (EOU), Ease of Learning (EOL), Satisfaction (SAT), Interaction Naturalism (INT), and Haptic Feedback (HAP). Each factor was represented by two to three items reflecting distinct but related constructs of system usability and interaction quality. In addition to the quantitative items, the questionnaire included open-ended questions to elicit qualitative feedback regarding perceived realism of the needling task, comfort of the haptic device, and suggestions for system improvement. This mixed-methods approach was intended to capture both measurable usability outcomes and nuanced user perceptions, particularly in relation to novel features such as wearable haptic feedback.

3. Results

A total of 600 valid trials were collected across all participants, comprising 480 beginner and 120 expert trials. The expert-weighted scoring model demonstrated robust construct validity, as experts consistently outperformed beginners in composite performance, angular precision, and consistency of depth control on all three acupoints (see Figure 9). While both groups achieved similarly high mean depth scores (experts: 94.34 ± 10.94; beginners: 89.09 ± 15.58), experts displayed notably lower variance, indicating more controlled and repeatable insertions. The angle score contributed most strongly to group differentiation, with experts attaining a mean of 75.49 ± 28.45 compared to 70.86 ± 34.65 for beginners. Notably, the composite final score was significantly higher among experts (69.25 ± 21.64) than beginners (56.26 ± 23.37), confirming the scoring model’s ability to detect experience-based performance distinctions (Table 4).
Figure 9. Comparison of performance metrics between beginners and experts. Boxplots illustrate differences in depth score, angle score, and composite final score across all trials for beginner (n = 8) and expert (n = 2) groups. Experts achieved higher final scores and exhibited reduced variability in depth accuracy, whereas angular deviation was the most distinguishing factor in group performance.
Table 4. Overall Mean (± SD) performance scores comparing beginners and experts.
As shown in Table 5, the procedural parameters, including insertion depth, angular deviation, and task duration, showed consistent improvement across the 20 trials for each acupoint, indicating clear learning progression among beginners.
Table 5. Early vs. Late Trial Performance among Beginners at each acupoint.
For LI11, mean insertion depth increased from 16.24 ± 1.88 mm in early trials (1–5) to 19.74 ± 1.23 mm in late trials (16–20), approaching the expert-defined target of 21.25 ± 2 mm. Although the mean insertion depth for LI11 (19.74 mm) appears slightly below the nominal target of 21.25 mm, it remains within the expert-defined acceptable range when considering the 2 mm tolerance used for performance evaluation. This outcome suggests that participants can approximate the clinically relevant depth despite having limited prior needling experience. The marginally shallower insertions may reflect cautious behavior, particularly for LI11, which requires deeper penetration than the other tested points, and are consistent with a natural tendency among novices to avoid over-insertion. As this study involved MR-familiar but acupuncture-naïve participants, the observed result supports the system’s capacity to facilitate depth control learning rather than representing a task failure. Simultaneously, angular deviation improved markedly, from 27.83 ± 16.62° to 15.34 ± 12.66°, converging toward the permissible ±15° range. Duration also declined sharply, from 38.77 ± 14.68 s to 13.28 ± 5.53 s, well within the 20-s expert threshold.
Similar trends were observed for LI4 and TE3. At LI4, the mean depth increased from 8.30 ± 1.90 mm to 11.72 ± 1.29 mm, approaching the clinical target of 12.5 ± 2 mm, while angle deviation dropped from 27.79 ± 15.73° to 14.84 ± 10.03°, and task time improved from 42.90 ± 18.17 s to 12.05 ± 4.67 s. At TE3, where the target depth was shallower (5.5 ± 2 mm), early trials showed systematic overshooting (8.25 ± 1.77 mm), which was corrected to 5.63 ± 1.63 mm by the end of the sequence. Angular deviation and time followed similar improvements, reducing from 28.07 ± 14.20° to 15.91 ± 12.01° and from 46.64 ± 19.47 s to 13.03 ± 5.06 s, respectively (see Table 5). These progressive refinements demonstrate the system’s efficacy in promoting short-term skill acquisition aligned with expert performance boundaries.
The learning curves in Figure 10 further illustrate this convergence, showing consistent reductions in depth error, angular deviation, and task duration across trials. Notably, all three metrics stabilized within the expert-defined acceptable ranges by approximately the 15th trial for most participants. The shaded threshold bands provided explicit training targets. The participants’ trajectories clearly trended toward and plateaued within these bands, indicating reliable internalization of procedural norms and stabilization of motor performance through the MR training system.
Figure 10. Learning curves of procedural performance over 20 trials in beginners with the shaded tolerance band and dashed safety limits. Trends for insertion depth, needle angle, and task duration are plotted across 20 consecutive trials at three acupoints (LI11, LI4, TE3). Shaded bands represent expert-defined acceptable thresholds for each parameter (depth: LI11 = 21.25 ± 2 mm, LI4 = 12.5 ± 2 mm, TE3 = 5.5 ± 2 mm; angle: 0–15°; duration ≤ 20 s).
Concomitantly, error rates and interface handling efficiency improved across sessions, further confirming enhanced procedural fluency. As summarized in Table 6, the average number of unnecessary needle droppings per five-trial block declined from 2.54 to 0.62, and anatomically incorrect insertions dropped from 3.44 to 1.08. Mean preparation time, defined as the interval from task initiation to first needle grasp, was reduced from 10.47 s to 3.11 s. These reductions indicate not only increased task familiarity but also a decrease in cognitive and operational load, which are essential for transitioning from early exploration to confident and efficient procedural execution.
Table 6. Reduction in error rates and preparation time across training progression.
While the scoring data captured accuracy and learning progression, the influence of real-time tactile cues provided by the haptic device on depth perception was further investigated. The analysis of vibration frequency patterns during trials provided insights into how haptic feedback assisted users in estimating needle depth. Figure 11 illustrates three representative cases of the trials performed on TE3: an under-insertion (shallow), an ideal insertion, and an over-insertion trial. The vibration frequency increased progressively as the needle advanced, but the relationship was not linear, reflecting the natural variability and discontinuity of hand motion during insertion. In the over-insertion trial, the continuous high-frequency vibration at the maximum depth served as a corrective cue, helping users recognize and limit excessive penetration. In contrast, the ideal trial exhibited a gradual and well-regulated frequency rise that stabilized near the target depth, indicating practical depth estimation through tactile cues. The shallow trial remained at lower frequencies, corresponding to insufficient penetration and limited haptic stimulation. Overall, these results suggest that depth-dependent vibration feedback not only conveys proximity to the target but also supports intuitive self-correction, thereby improving control over insertion depth during training.
Figure 11. Vibrator motor command frequency over time for three representative TE3 insertion trials. Each trial corresponds to a different outcome in terms of needling depth: an overshoot exceeding the maximum safe depth, an ideal insertion reaching the expert-defined target depth, and an undershoot remaining below the minimum acceptable depth. Shaded regions indicate time intervals where vibration frequency increased in response to needle progression.
To further assess the effectiveness of the wearable haptic device in aiding depth control during needle insertion, scatter plots of final depth versus depth error were generated for each acupoint and user group (see Figure 12). The haptic device was programmed to increase vibration intensity in proportion to the insertion depth and to emit a continuous buzz once the needle exceeded the maximum depth. The results show that both expert and beginner users were generally able to maintain their insertions within or close to the acceptable range, indicating effective utilization of haptic cues. For the LI11 and LI4 acupoints, depth estimates were well-distributed around the target region for both groups, although beginners exhibited slightly wider error margins. Notably, for the more challenging TE3 acupoint, which has a shallow depth range (target depth = 5.5 mm), a few beginner trials exceeded the maximum safe threshold. Still, the over-insertions remained minimal (within 1 mm). This suggests that the continuous vibration feedback effectively signaled excessive insertion to the users. However, the lack of distinct vibration cues at the Dmin and Dtarget thresholds may explain the broader spread of depth errors, particularly among beginners who lacked clear tactile markers to identify when the ideal insertion depth was reached. In interpreting Figure 10 and Figure 12, insertions within the ±2 mm tolerance band were considered accurate, whereas those outside this range but within the minimum–maximum boundaries were regarded as safe but suboptimal. Overall, the results demonstrate that the vibrotactile guidance system helped users estimate needle depth accurately and safely, with minimal deviation from the optimal range.
Figure 12. Scatter plots of final depth versus depth error for each acupoint and user group, with haptics enabled. Red dotted lines indicate expert-defined thresholds: minimum acceptable depth: D(Min), target depth: D(Target), and maximum safe depth: D(Max). Both expert and beginner users showed strong convergence toward the target region, guided by the haptic device.
Following completion of both training modes, participants rated the system using the 15-item usability questionnaire. The descriptive statistics for each usability factor are summarized in Table 7, and inter-factor relationships are visualized in the correlogram (see Figure 13). Overall, ratings were high across all dimensions, with mean scores above 4.4 on the 5-point scale, indicating consistently positive user evaluations of the MR training experience.
Table 7. Descriptive statistics for each usability factor.
Figure 13. Correlogram of factor-level Spearman correlations (ρ) for the six usability dimensions of the MR acupuncture training system.
Spearman correlation analysis revealed strong positive associations among most factors (ρ > 0.6), suggesting that perceptions of usefulness, ease of use, learning efficiency, and interaction realism were mutually reinforcing. The strongest relationship was observed between Usefulness and Haptic Feedback (ρ = 0.92), indicating that the haptic wrist-ring device was central to participants’ sense of the system’s practical training value. Similarly, correlations between Ease of Use and Haptic Feedback (ρ = 0.77) and between Ease of Use and Interaction Naturalism (ρ = 0.64) suggest that intuitive device control and realistic manipulation substantially contributed to user satisfaction. Satisfaction itself exhibited moderate associations with Interaction Naturalism (ρ = 0.61), but weaker links with other factors, implying that overall satisfaction may depend on experiential quality beyond operational simplicity.
Open-ended responses further reinforced these findings. Participants described the haptic cues as “realistic,” “informative,” and “helpful for judging depth,” though several noted mild discomfort from the wristband or HMD during extended use. Collectively, these results suggest that multisensory realism, particularly through tactile feedback, played a pivotal role in shaping the perceived effectiveness and acceptance of the training system.
These findings demonstrate that the MR acupuncture needling trainer is feasible, usable, and capable of producing measurable learning and procedural stability within short, repeated sessions. The results provide evidence for both the construct validity of the scoring model and the added value of haptic feedback for depth control. At the same time, user-reported measures confirm the system’s acceptability as a training platform.

4. Discussion

This study introduced and evaluated an MR acupuncture training platform incorporating a full-scale MetaHuman avatar, an expert-informed scoring algorithm, and a depth-responsive wearable haptic feedback device. The system demonstrated significant usability and early training effects among MR-familiar participants, offering a novel contribution to simulation-based acupuncture education. These findings highlight the simulator’s potential to improve procedural realism and skill acquisition by combining spatially anchored MR visualization with real-time haptic feedback and quantitative assessment.
This outcome aligns with prior XR-based medical training studies that have emphasized the importance of quantitative feedback and procedural accuracy improvement [,,,,]. However, unlike earlier XR or AR acupuncture prototypes that lacked haptic or performance assessment capabilities [,,,], the present system integrates both quantitative scoring and tactile guidance. The observed learning progression and usability scores support the findings from Sung et al. [] and Jongbloed et al. [], who reported that immersive feedback loops enhance skill retention and confidence. Moreover, by employing expert-defined acupoint tolerances, this study adapts task-specific evaluation frameworks previously validated for surgery and anesthesia simulators [,,,,] to the acupuncture domain. Thus, the results consolidate earlier XR evidence while addressing the literature-identified gaps in haptic realism and objective assessment for acupuncture training. These alignments with prior research further support the system’s potential as a validated training tool within broader XR-based medical education frameworks.
A key achievement of this work lies in the integration of expert-derived acupoint modeling with tolerance-based performance scoring. By using clinically informed depth, angle, and tip location thresholds, the system ensures that training reflects realistic procedural expectations. The inclusion of tolerance bands was instrumental in mitigating the inherent precision limitations of the HoloLens 2 while maintaining clinically meaningful evaluation fidelity. Furthermore, continuous expert involvement in the development of acupoint modeling, marker placement, and scoring logic helped preserve clinical validity while striking a balance between instructional rigor and technical feasibility.
Although the present system does not yet employ a formal adaptive learning algorithm, it promotes progressive skill refinement through iterative, feedback-driven practice. The combination of visual and haptic feedback allows users to self-correct and improve across sessions, functioning as an implicit adaptive mechanism. The modular design of the MR–haptic framework also enables future integration of algorithmic adaptation, such as difficulty scaling or personalized feedback based on performance trends.
The wearable haptic device played a pivotal role in enhancing depth awareness. Its real-time vibration pattern effectively guided users in regulating insertion depth, with continuous buzzing at over-insertion serving as a corrective cue. Despite differences between vibration-based feedback and the actual tactile sensations of skin penetration, participants reported that the cues were informative and natural, supporting intuitive depth regulation. Notably, the lightweight, wireless wrist-based design preserved freehand mobility and ergonomic comfort, avoiding the encumbrance typical of many haptic systems.
Additionally, using a highly realistic MetaHuman avatar improved psychological realism. Participants indicated that human-like visual fidelity increased engagement and could reduce anxiety about practicing on real patients, aligning with evidence that anthropomorphic virtual patients enhance emotional readiness and immersion in medical simulation. Collectively, these elements advance the pedagogical goals of immersive, quantitative, and multisensory training for acupuncture.
The usability study also revealed technical constraints associated with high-precision holographic simulation. The slender holographic needle exhibited visual instability and lag during rapid hand movements, primarily due to tracking latency and rendering delays of HoloLens 2. These limitations affected fine-motor alignment but were mitigated through the scoring logic, which evaluates the needle’s final position upon release—mirroring real-world assessment practices where final placement matters most.
Another factor influencing interaction quality was the participants’ familiarity with MR. Early pilot testing with MR-unfamiliar users yielded frequent interaction errors, including difficulty grasping the virtual needle and selecting UI elements. Consequently, the study intentionally used MR-experienced participants to focus on system feasibility rather than basic interface learning. However, this highlights the need for tutorial or onboarding modules to support novice users in future iterations.
Among measured parameters, angular deviation displayed greater variability than depth, particularly among beginners. This may stem from the lack of physical resistance in holographic simulations—unlike real tissue, which stabilizes the insertion angle. Addressing this issue may involve implementing virtual constraints or visual angle-maintenance guides to enhance angular consistency. Despite these technical challenges, the integration of tolerance-based evaluation successfully compensated for hardware imprecision, ensuring reliable scoring and maintaining educational validity within the operational limits of current MR hardware.
While the findings affirm the system’s feasibility and short-term training benefits, several study-specific limitations warrant discussion. The present evaluation involved MR-familiar participants rather than medical trainees, meaning that usability and interaction reliability were assessed in a technically competent but non-clinical population. Consequently, the educational transferability of the system to real training environments remains to be validated. Furthermore, the study focused on only three acupoints, which limits its generalizability across the full acupoint spectrum and anatomical diversity. The user interface and interaction design also require refinement to accommodate participants with varying levels of familiarity with head-mounted displays, as interaction complexity may pose barriers for first-time MR users. Additionally, the short-term study duration precludes conclusions about long-term skill retention or the transfer of learned motor patterns to real-world practice. Finally, the modest sample size of ten participants, including two experts, was not sufficient for inferential statistical testing or subgroup analysis. Nevertheless, this scale is typical for early-phase feasibility studies of mixed-reality training systems, which primarily aim to establish technical reliability and construct validity rather than educational generalizability.
Future work should extend validation to clinical cohorts of acupuncture trainees to evaluate educational transfer and curricular integration. Expanding the acupoint library, improving angular guidance features, and enhancing haptic fidelity to simulate tissue resistance gradients are critical next steps. Broader testing with MR-unfamiliar novices will ensure accessibility for diverse learner populations. Additionally, algorithmic personalization and adaptive feedback mechanisms may further optimize the learning curve. Once validated among medical students and integrated into curricula, the system could provide a standardized, scalable, and multisensory training platform for developing procedural skills in acupuncture.
Beyond acupuncture training, the haptic-MR framework developed here shares conceptual mechanisms with human–machine interaction systems, especially those applying impedance or admittance learning in contact tasks [,]. Thus, this wearable haptic-enabled MR method could be extended to domains such as tele-manipulation, robot-assisted rehabilitation, and human–robot contact learning. Future research could explore cross-domain applications of this lightweight haptic design in human–machine interaction and rehabilitation robotics, extending its impact beyond medical simulation.

5. Conclusions

This study presented an MR acupuncture training system that integrates a high-fidelity MetaHuman avatar, expert-informed acupoint modeling, a multi-parameter scoring algorithm, and a wearable haptic feedback device. The system demonstrated usability, procedural stability, and learning progression within a short training duration among MR-familiar participants. The incorporation of tolerance-based evaluation and vibration-guided depth feedback effectively mitigated hardware limitations of HoloLens 2, while the MetaHuman model enhanced the psychological realism of the training environment. The findings confirmed that the expert-weighted scoring framework reliably differentiated user proficiency and that repeated trials supported measurable improvement across depth, angle, and time metrics. However, challenges such as interaction complexity, angular instability, and limited haptic realism underscore areas for future enhancement. Broader validation with medical learners, expansion to diverse acupoints, and refinement of feedback modalities are necessary to confirm educational efficacy. Overall, this system represents a promising direction for immersive, quantitative, and feedback-driven acupuncture training within modern medical education frameworks.

Author Contributions

Conceptualization, K.G., H.-J.P. and C.-S.N.; methodology, K.G., H.M.K.K.M.B.H. and N.M.; software, K.G. and H.M.K.K.M.B.H.; validation, K.G.; formal analysis, K.G. and N.M.; investigation, K.G. and H.M.K.K.M.B.H.; resources, H.-J.P., C.-S.N., M.Y. and B.-i.L.; data curation, K.G. and H.M.K.K.M.B.H.; writing—original draft preparation, K.G. and H.M.K.K.M.B.H.; writing—review and editing, N.M., H.-J.P., C.-S.N., M.Y. and B.-i.L.; visualization, K.G. and H.M.K.K.M.B.H.; supervision, N.M., H.-J.P., C.-S.N., M.Y. and B.-i.L.; project administration, H.-J.P., C.-S.N. and B.-i.L.; funding acquisition, B.-i.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Research Foundation of Korea (NRF) and funded by the Ministry of Science and ICT (No. 2022M3A9B6082791).

Data Availability Statement

Data collected for this study are presented in the manuscript.

Acknowledgments

The authors would like to thank the acupuncture medical experts who contributed to the development of the scoring mechanism.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Zhang, M.; Chen, C.; Yarmand, M.; Rajeshkumar, A.; Weibel, N. AcuVR: Enhancing Acupuncture Training Workflow with Virtual Reality. arXiv 2024, arXiv:2407.02614. [Google Scholar] [CrossRef]
  2. Sun, Q.; Huang, J.; Zhang, H.; Craig, P.; Yu, L.; Lim, E.G. Design and Development of a Mixed Reality Acupuncture Training System. In Proceedings of the 2023 IEEE Conference on Virtual Reality and 3D User Interfaces, Shanghai, China, 25–29 March 2023; pp. 265–275. [Google Scholar]
  3. Jang, J.E.; Lee, Y.S.; Jang, W.S.; Sung, W.S.; Kim, E.-J.; Lee, S.D.; Kim, K.H.; Jung, C.Y. Trends in Acupuncture Training Research: Focus on Practical Phantom Models. J. Acupunct. Res. 2022, 39, 77–88. [Google Scholar] [CrossRef]
  4. Sung, H.; Kim, M.; Park, J.; Shin, N.; Han, Y. Effectiveness of Virtual Reality in Healthcare Education: Systematic Review and Meta-Analysis. Sustainability 2024, 16, 8520. [Google Scholar] [CrossRef]
  5. Kim, H.-Y.; Kim, E.-Y. Effects of Medical Education Program Using Virtual Reality: A Systematic Review and Meta-Analysis. Int. J. Environ. Res. Public Health 2023, 20, 3895. [Google Scholar] [CrossRef]
  6. Liu, K.; Zhang, W.; Li, W.; Wang, T.; Zheng, Y. Effectiveness of virtual reality in nursing education: A systematic review and meta-analysis. BMC Med. Educ. 2023, 23, 870. [Google Scholar] [CrossRef] [PubMed]
  7. Jongbloed, J.; Chaker, R.; Lavoué, E. Immersive procedural training in virtual reality: A systematic literature review. Comput. Educ. 2024, 221, 105124. [Google Scholar] [CrossRef]
  8. Mitchell, A.A.; Ivimey-Cook, E.R. Technology-enhanced simulation for healthcare professionals: A meta-analysis. Front. Med. 2023, 10, 1149048. [Google Scholar] [CrossRef] [PubMed]
  9. Han, S.H.; Kiroff, K.L.; Kinjo, S. Extended reality in anesthesia: A narrative review. Korean J. Anesthesiol. 2025, 78, 105–117. [Google Scholar] [CrossRef]
  10. Fleet, A.; Kaustov, L.; Belfiore, E.B.; Kapralos, B.; Matava, C.; Wiegelmann, J.; Giacobbe, P.; Alam, F. Current Clinical and Educational Uses of Immersive Reality in Anesthesia: Narrative Review. J. Med. Internet Res. 2025, 27, e62785. [Google Scholar] [CrossRef]
  11. Rubin, J.E.; Shanker, A.; Berman, A.B.; Pandian, B.; Jotwani, R. Utilisation of extended reality for preprocedural planning and education in anaesthesiology: A practical guide for spatial computing. Br. J. Anaesth. 2024, 132, 1342–1344. [Google Scholar] [CrossRef]
  12. Gsaxner, C.; Li, J.; Pepe, A.; Jin, Y.; Kleesiek, J.; Schmalstieg, D.; Egger, J. The HoloLens in medicine: A systematic review and taxonomy. Med. Image Anal. 2023, 85, 102757. [Google Scholar] [CrossRef]
  13. Asif, H.; McInnis, C.; Dang, F.; Ajzenberg, H.; Wang, P.L.; Mosa, A.; Ko, G.; Zevin, B.; Mann, S.; Winthrop, A. Objective Structured Assessment of technical skill in the Surgical Skills and Technology Elective Program: Comparison of peer and expert raters. Am. J. Surg. 2021, 223, 276–279. [Google Scholar] [CrossRef] [PubMed]
  14. Alibhai, K.; Fowler, A.; Gawad, N.; Wood, T.J.; Raîche, I. Assessment of laparoscopic skills: Comparing the reliability of global rating and entrustability tools. Can. Med. Educ. J. 2022, 13, 36–45. [Google Scholar] [CrossRef] [PubMed]
  15. Van Alexander, A.S.; Stefan, S.; Johannes, J.; Gisbertz, S.S.; Toorenvliet, B.R.; Tanis, P.J.; Bonjer, H.J.; Tuynman, J.B. Video-based tools for surgical quality assessment of technical skills in laparoscopic procedures: A systematic review. Surg. Endosc. 2023, 37, 4279–4297. [Google Scholar] [CrossRef]
  16. Wu, M.; Kit, C.Y.; Su, E.L.M.; Yeong, C.F.; Ahmmad, S.N.Z.; Holderbaum, W.; Yang, C. Quantitative metrics for evaluating surgical dexterity using virtual reality simulations. PLoS ONE. 2025, 20, e0318660. [Google Scholar] [CrossRef]
  17. Guerrero-Antolino, P.; Gutiérrez-Sánchez, C.; Millán-Scheiding, M. Assessment tools for minimally invasive surgery simulation programmes: A narrative review. Ann. Laparosc. Endosc. Surg. 2024, 9, 23. [Google Scholar] [CrossRef]
  18. Huang, X.; Yan, Z.; Gong, C.; Zhou, Z.; Xu, H.; Qin, C.; Wang, Z. A Mixed-Reality Stimulator for Lumbar Puncture Training: A Pilot Study. BMC Med. Educ. 2023, 23, 178. [Google Scholar] [CrossRef]
  19. Felten, R.; Bigaut, K.; Wirth, T.; Kremer, L.; Gauer, L.; Arnold, C.; Ollivier, I.; Godet, J.; Scherlinger, M.; Dubois, M.; et al. Advancing medical training with augmented reality and haptic feedback simulator: Outcomes of a randomized controlled trial on lumbar puncture. BMC Med. Educ. 2025, 25, 1231. [Google Scholar] [CrossRef]
  20. Gao, L.; Xu, Y.; Zhang, X.; Jiang, Z.; Wu, J.; Dong, Y.; Li, M.; Jin, L.; Qiu, J.; You, L.; et al. Comparison of Mixed Reality-Assisted Spinal Puncture with Landmark-Guided Spinal Puncture by Novice Practitioners: A Pilot Study. J. Pain Res. 2024, 17, 2701–2712. [Google Scholar] [CrossRef]
  21. Vrillon, A.; Gonzales-Marabal, L.; Ceccaldi, P.-F.; Plaisance, P.; Desrentes, E.; Paquet, C.; Dumurgier, J. Using Virtual Reality in Lumbar Puncture Training Improves Students Learning Experience. BMC Med. Educ. 2022, 22, 224. [Google Scholar] [CrossRef]
  22. Peng, Y.; Xie, Z.; Chen, S.; Wu, Y.; Dong, J.; Li, J.; He, J.; Chen, X.; Gao, H. Application effect of head-mounted mixed reality device combined with 3D printing model in neurosurgery ventricular and hematoma puncture training. BMC Med. Educ. 2023, 23, 670. [Google Scholar] [CrossRef]
  23. Tanwani, J.; Alam, F.; Matava, C.; Choi, S.; McHardy, P.; Singer, O.; Cheong, G.; Wiegelmann, J. Development of a Head-Mounted Holographic Needle Guidance System for Enhanced Ultrasound-Guided Neuraxial Anesthesia: System Development and Observational Evaluation. JMIR Form. Res. 2022, 6, e36931. [Google Scholar] [CrossRef]
  24. Hayasaka, T.; Kawano, K.; Onodera, Y.; Suzuki, H.; Nakane, M.; Kanoto, M.; Kawamae, K. Comparison of accuracy between augmented reality/mixed reality techniques and conventional techniques for epidural anesthesia using a practice phantom model kit. BMC Anesthesiol. 2023, 23, 171. [Google Scholar] [CrossRef] [PubMed]
  25. Zheng, T.; Xie, H.; Gao, F.; Gong, C.; Lin, W.; Ye, P.; Liu, Y.; He, B.; Zheng, X. Research and application of a teaching platform for combined spinal-epidural anesthesia based on virtual reality and haptic feedback technology. BMC Med. Educ. 2023, 23, 794. [Google Scholar] [CrossRef] [PubMed]
  26. Liao, S.-C.; Shao, S.-C.; Gao, S.-Y.; Lai, E.C.-C. Augmented reality visualization for ultrasound-guided interventions: A pilot randomized crossover trial to assess trainee performance and cognitive load. BMC Med. Educ. 2024, 24, 1058. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, Y.; Hua, W.; Zhou, Z.; Zhu, H.; Xiong, J.; Zhang, J.; Chen, D.; Guo, J. A novel acupuncture technique at the Zusanli point based on virtual reality and EEG: A pilot study. Front. Neurosci. 2024, 18, 1269903. [Google Scholar] [CrossRef]
  28. Chiou, S.-Y.; Chang, H.-H.; Chen, Y.-C.; Liu, G.-H. Augmented Reality Navigation for Acupuncture Procedures with Smart Glasses. Electronics 2025, 14, 3025. [Google Scholar] [CrossRef]
  29. Yang, S.; Ryu, C.; Kim, S.; Kim, J. A Development of an Acupoints Education Table using 3D Technology and Augmented Reality. Korean J. Acupunct. 2021, 38, 267–274. [Google Scholar] [CrossRef]
  30. Malekroodi, H.S.; Seo, S.-D.; Choi, J.; Na, C.-S.; Lee, B.; Yi, M. Real-Time Location of Acupuncture Points Based on Anatomical Landmarks and Pose Estimation Models. Front. Neurorobot. 2024, 18, 1484038. [Google Scholar] [CrossRef]
  31. Li, Y.; Chan, R.; Menon, M.R.G.; Ryan, J.F.; Mador, B.; Campbell, S.M.; Turner, S.R. Validity Evidence for Procedure-specific Competence Assessment Tools in Orthopaedic Surgery: A Scoping Review. JAAOS Glob. Res. Rev. 2024, 8, e23.00065. [Google Scholar] [CrossRef]
  32. Sekhavati, P.; Wild, T.; Ingrid, I.; Dion, P.; Woo, M.; Ramlogan, R.; Boet, S.; Shorr, R.; Gu, Y. Instructional Design Features in Ultrasound-Guided Regional Anaesthesia Simulation-Based Training: A Systematic Review. Anaesthesia 2025, 80, 572–581. [Google Scholar] [CrossRef]
  33. Li, H.-Y.; Dharmawan, A.G.; Paranawithana, I.; Yang, L.; Tan, U.-X. A Control Scheme for Physical Human-Robot Interaction Coupled with an Environment of Unknown Stiffness. J. Intell. Robot. Syst. 2020, 100, 165–182. [Google Scholar] [CrossRef]
  34. Xing, X.; Burdet, E.; Si, W.; Yang, C.; Li, Y. Impedance Learning for Human-Guided Robots in Contact with Unknown Environments. IEEE Trans. Robot. 2023, 39, 3705–3721. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.