Next Article in Journal
Study on the Co-Evolution Mechanism of Key Strata and Mining Fissure in Shallow Coal Seam Mining
Previous Article in Journal
Linear Optics Calibration in a Storage Ring Based on Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Experimental Evaluation of Extended Reality Technologies in the Development of Individualized Three-Dimensionally Printed Upper Limb Prostheses

1
Faculty of Mechanical Engineering, Poznan University of Technology, 60-965 Poznan, Poland
2
Department of Engineering Sciences, Faculty of Engineering and Science, University of Agder (UiA), Jon Lilletuns vei 9, NO-4879 Grimstad, Norway
3
Bizzcom s.r.o., Šľachtiteľská ulica 591/2, 919 28 Bučany, Slovakia
4
Department of Manufacturing Engineering, Faculty of Industrial Engineering and Robotics, University Politehnica of Bucharest, Blv. Splaiul Independenței, No. 313, Sector 6, 060042 Bucharest, Romania
5
Department of Manufacturing Engineering, Faculty of Industrial Engineering, Robotics and Production Management, Technical University of Cluj-Napoca, Blv. Muncii, No. 103-105, 400641 Cluj-Napoca, Romania
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(14), 8035; https://doi.org/10.3390/app13148035
Submission received: 12 June 2023 / Revised: 4 July 2023 / Accepted: 5 July 2023 / Published: 10 July 2023

Abstract

:

Featured Application

The work presented in this paper will be applied in the process of design and the production of customized, low-cost upper limb prostheses, using the AutoMedPrint technology. The developed extended reality (XR) applications will be used for training and facilitating communication with customers.

Abstract

This paper presents results from experimental studies that assess the utilization of virtual, augmented, and mixed reality (VR, AR, MR) at different stages of developing personalized 3D printed upper limb prostheses for adult patients. The prostheses are designed automatically using the AutoMedPrint system, leveraging 3D scans as described in various prior publications. Various stages of development of the prosthesis are made as applications of different extended reality technologies. An assembly instruction is implemented as an immersive VR application, a configurator is designed as AR application and a configurator and try-on application is prepared and deployed in MR. The applications are tested by an international group of experts during a scheduled experiment. The experts then participate to surveys and comparatively evaluate the potential of all the XR technologies. The paper presents the development of these applications, provides a detailed account of the experimental process, including the rankings of XR technologies for different applications throughout the lifecycle of a prosthetic device.

1. Introduction

The development of individualized three-dimensionally printed upper limb prostheses has been a significant area of research in the field of medical technology [1], with many advancements performed in recent years regarding additive manufacturing of such devices [1,2,3]. Prosthetic devices aim to improve the quality of life for individuals with limb loss by restoring lost functionality and enhancing their mobility and independence. However, designing and fabricating prosthetic devices that meet the unique needs and preferences of each patient remains a complex challenge. In recent years, the emergence of virtual reality (VR), augmented reality (AR), and mixed reality (MR) technologies has shown promising potential in revolutionizing the design of various medical devices, including prosthetics [4].
The field of upper limb prosthetics has witnessed remarkable advancements over the years. Traditional manufacturing processes for prostheses often involve time-consuming and labor-intensive techniques, resulting in standardized devices that may not fully meet the unique anatomical and functional requirements of individual patients [5,6]. This limitation has spurred the exploration of innovative technologies, such as three-dimensional printing, to address the customization challenges and to enable the production of personalized prostheses [1,5,6], although the modern process is not without its own issues to solve [3,7]. One of the examples of the innovative technologies for rapid production of individualized prostheses is the AutoMedPrint system [6]. It allows one to build inexpensive three-dimensionally printed prostheses for functional use, in a very short time, thanks to the automation of the data processing and design stages [6]. In the AutoMedPrint system, AR and VR technologies are part of the process, helping during work with the patient in customizing the product [8].
Virtual reality, an immersive technology that simulates a computer-generated environment, has gained significant attention in various domains, including medicine and rehabilitation [9]. Mixed reality and augmented reality, technologies that blend virtual elements with the real world, are also used in many contexts in engineering [10] and medicine [11]. MR devices, such as Microsoft HoloLens and Magic Leap, overlay digital content onto the user’s real-world environment, creating a seamless integration between the virtual and physical realms. This enables users to interact with virtual objects—often with their own hands by use of hand motion tracking—and to receive real-time feedback [12]. AR, on the other hand, can be utilized in everyday use devices, such as tablets and cellphones, requiring only markers (such as QR codes) to present the digital content overlaid on the real world. As such, AR applications can be easily widespread and remain accessible for many persons, facilitating education [13], as well as the improvement of many engineering processes [14]. Augmented reality and mixed reality operate on a similar principle—the difference lies in the spatial context (as defined in [15]), present only in AR, non-existent in MR. In other words, digital images in AR are linked spatially with specific real-world objects, while MR visualizations are mostly disconnected and able to freely move around and interact in the blended virtual-real scene.
Virtual, mixed, and augmented reality put together are considered as Extended Reality (XR) technologies by today’s nomenclature, filling the whole simulated reality continuum, as defined three decades prior by Milgram [16]. The current trend is integration of the three, both on the hardware and the software level, and, frequently, applications can blend between virtual and mixed and between mixed and augmented [17], even using a single device having all these capabilities (examples of such devices are Meta Quest Pro or Varjo XR-3 headsets, available on the market since 2022).
From the engineering viewpoint, XR technologies are considered to be an integral part of the Industry 4.0 concept and a vital part of the smart factory concept [18]. Numerous researchers investigated the use of Augmented and Virtual Reality in various stages of the product lifecycle, including design, assembly, quality control, and manufacturing simulation [19]. It is nowadays considered as general knowledge that all the XR technologies can be successfully implemented in industrial production and that they can greatly help in modern product development.
The medical domain, as well as medical production, is a prominent area of application of XR. For example, VR and MR technologies offer novel possibilities for preoperative planning and patient education. Virtual, augmented, and mixed reality, used together with haptics, can be also of a great help in remote patient examination [20].
As regards medical production, VR technology offers compelling opportunities for the design phase. By utilizing VR, designers and engineers can create virtual models of patient-specific devices and iterate upon them in a virtual space. This allows for rapid customization and modification, as well as the ability to incorporate patient-specific anatomical data [21]. For example, VR allows for the precise measurement and modeling of residual limbs, ensuring optimal fit and functionality of the prosthesis. By incorporating VR-based design processes, researchers and practitioners can develop individualized upper limb prostheses that are tailored to each patient’s unique requirements. VR-based design also enables collaboration among multidisciplinary teams, including clinicians, engineers, and patients, fostering a more comprehensive and patient-centered approach to prosthetic development [22].
For individuals who require upper limb amputation, VR can simulate the expected outcomes of different prosthetic designs and aid in the decision-making process. Furthermore, VR-based environments can provide a platform for patients to familiarize themselves with their future prostheses and practice activities of daily living, thus reducing anxiety and improving the overall acceptance and adoption of the devices [4].
Virtual and mixed reality technologies also offer immense potential in the training and rehabilitation of individuals using upper limb prostheses. With VR, users can engage in immersive training simulations that replicate real-world activities, allowing them to practice and refine their prosthetic control and manipulation skills [4,23]. These simulations provide a safe and controlled environment for patients to develop confidence and proficiency in using their prostheses, ultimately enhancing their overall rehabilitation experience. Furthermore, mixed reality (MR) devices, such as Microsoft HoloLens, overlay digital content onto the user’s real-world environment, enabling real-time feedback and guidance during training sessions [11].
In addition to design and training, virtual and mixed reality technologies have other promising applications in the medical domain. VR can be used to simulate the expected outcomes of different prosthetic designs, aiding in the decision-making process for individuals requiring upper limb amputation [24]. Patients can virtually visualize and interact with different prosthetic options, allowing them to make informed choices based on their preferences and functional requirements. Moreover, VR-based environments can also be utilized to reduce anxiety and to improve patient acceptance and adoption of prostheses. By providing a platform for patients to practice activities of daily living in a virtual space, VR fosters familiarity and confidence in using the prosthetic device [25].
In summary, the integration of virtual and mixed reality technologies in the development of individualized three-dimensionally printed upper limb prostheses has the potential to revolutionize the field. By leveraging these technologies, researchers and practitioners can create customized prosthetic designs, provide immersive training experiences, and enhance patient acceptance and rehabilitation outcomes. The state-of-the-art analysis presented in this section highlights the advancements and possibilities offered by VR and MR in the specific context of upper limb prosthetics. It establishes the foundation for the experimental evaluation that will be conducted in this research.
To date, there has been limited exploration of the combined potential of the entire spectrum of these technologies in the lifecycle and development of rapidly manufactured personalized upper limb prosthetics. The studies on the matter, presenting expert evaluation and comparison of various technologies from the XR continuum in such use, are practically non-existent in the current literature. As such, the use of these technologies is still quite limited and mostly experimental, and, in the authors’ opinion, this is in spite of their huge potential. Recent study examples include, e.g., comparison of AR versus VR in driver assistance [26] or general selection of XR methods for production based on case studies [27]. In these and other examples, academia experts were asked for their opinions to test and to validate various technologies in certain use cases. However, in the available literature, there are no attempts at testing and validating all three XR technologies in the case of a personalized medical product. As such, the authors believe that the proposed approach is novel and not directly comparable with other methods.
This scientific paper aims to explore and evaluate the applications of VR and MR technologies as a part of the lifecycle of individualized upper limb prostheses. By presenting the results of the performed experimental studies, the authors seek to provide valuable insights for scientists, researchers, and practitioners in the field, ultimately contributing to the advancement of personalized prosthetic solutions and incorporating XR technologies into the design and production processes.

2. Materials and Methods

2.1. Research Concept and Plan

The aim of the presented research is to evaluate and compare all three basic XR technologies—Virtual Reality, Augmented Reality and Mixed Reality—in the process of development of a customized, individually designed medical product, which is a modular prosthesis, manufactured with 3D printing technologies (Figure 1).
The research is realized using the following assumptions:
  • The prosthesis model is prepared in a CAD system, using available archive anatomical data as gathered in some previous studies [28]. It is exported to an appropriate filetype for visualization in a 3D engine. Two different versions of the model are created: one for the configuration process (AR and MR) and the other for the assembly process (VR).
  • The models are imported into the 3D engine, with Unity game engine being employed for all the applications, and subsequent preparation of the applications takes place. Each application is created using different frameworks, with their own algorithms and scripting for interaction implementation. However, only free-to-use plugins and software development kits are used to ensure enhanced reproducibility of the process, while duly considering their inherent limitations.
  • The applications are tested among the group of purposefully selected experts. The experts all have an engineering background and expertise in the design and manufacturing of customized medical products. However, the majority of experts lack extensive prior experience with XR technologies, leaving them uncertain about what to anticipate from the technology. The expert group is also diversified in terms of age, gender, and ethnicity.
  • After testing, experts fill out an anonymous survey for the evaluation of a given technology. After testing all three applications, the experts fill out another survey, aimed at comparing technologies and their potential. All the answers, except for the open questions, are performed using a 5-point Likert scale for easier comparison of the results. The experts are also free to exchange opinions among themselves and to ask any questions related to the technology, so as to be able to fully evaluate its potential in the product lifecycle.
  • In the final stage, results of testing, juxtaposed with researchers’ observations, are used to draw the final conclusions and recommendations of the use of XR technologies in the medical product development process.
The scheme of research as described above is presented in Figure 2. More details are given in the subsequent technical sections.

2.2. Background and Motivation—AutoMedPrint System

The initial concept of the modular prosthesis, which is a basis of all the applications, was made as part of the project “Automation of design and rapid production of individualized orthopedic and prosthetic products based on data from anthropometric measurements”, serving the development of the prototype AutoMedPrint system [28,29], continuing previous long-term studies by the authors. The prosthesis design was created and iteratively improved in a number of studies [30], resulting in successful implementations for various patients, demonstratively shown in Figure 3 (not part of this study).
The AutoMedPrint system (awarded as the Polish Product of the Future in the year 2022) itself is a complex solution, containing hardware and software for realization of a full cycle of design of customized orthopedic and prosthetic devices. Apart from 3D scanning, CAD, and 3D printing modules, the AutoMedPrint system also contains a visualization module, which is used for real-time communication with the patients using the system. The scheme in Figure 4 presents how the system actually operates, while Figure 5 presents how it looks in its current, prototype phase. Currently, the system is not yet directly available as a commercial solution, and it is treated as a research tool. One of the authors’ overarching goals in their studies, including this one, is the creation of a solution for commercial purposes to popularize the automated design and the production approach utilized in the system prototype.
It was noted by the authors that XR technologies could be successfully used in aiding the development of the customized products, helping the patients to familiarize with the designed customized devices, virtually configure and try them on, or even assembly them in virtual space. Therefore, the underlying motivation of the studies discussed in this paper is to assess, among a group of experts, how particular XR technologies could be effectively integrated into the design and development process. The objective is to potentially achieve improved outcomes and increased user satisfaction, ultimately leading to more efficient utilization of custom-designed medical products in therapy. This aspect remains undecided within the current literature. The obtained results will be used to refine the process of work with the system, both with child and adult patients, especially regarding the prosthetic devices.

2.3. XR Applications Development

2.3.1. Application Types and Data Preparation

The first stage of development was a selection of technologies and their matchup with basic application types used in the product lifecycle. The following decisions are made:
  • VR technology is assigned to one of the manufacturing process elements—the manual assembly process of the 3D printed prosthesis. In order to obtain more complex visualization and level of difficulty, it is decided to consider a mechatronic prosthesis, equipped with microcontrollers and sensors. It is decided to create a training application compatible with any PC VR headset, operated using a standard pair of goggles with controllers.
  • AR technology is assigned to the simplified configuration process, allowing one to visualize the prosthesis and change its basic modules and colors in the real environment. To ensure simplicity, it is decided to use a tablet device with the Android operating system, with interactions by use of touchscreen and with simple image markers.
  • MR technology is assigned to a mixed type of application, involving both configuration process and some interactions performed with the prosthesis, such as virtual try-on, explosive view, and other manipulations. It is decided to create the application using a popular VR headset with a passthrough image of the real world and hand tracking interactions, thus creating a MR experience.
The initial data preparation work involves use of the modular CAD model [31] to prepare a set of 3D models, with some demonstrative modularity capabilities for the configuration process. The same model is also used to develop a single variant of a prosthesis for assembly instruction. Both models are presented in Figure 6.
Apart from the 3D model data of the products, the following other data is used in building the applications:
  • UI icons, created by the authors,
  • 3D models and textures of environment (for the VR application), downloaded from multiple free sources and/or created by the authors in previous research,
  • scripts in the C# language, created in the previous research for scientific or didactic purposes.
When data packages are prepared, the actual work of development commences. All applications are created using the Unity engine (Unity Software Inc., San Francisco, CA, USA), with a free educational license granted to higher education institutions, in version 2019.4.37f1.

2.3.2. Virtual Reality—Assembly Training

The VR application was built using the aforementioned Unity Engine. The fundamental approach to incorporate VR functionalities involves transforming the basic Unity project into a VR-compatible version by leveraging the OpenVR library (integrated within Unity) and the SteamVR plugin for Unity, specifically version 2.7.3 (Valve Corporation, Bellevue, WA, USA). The SteamVR plugin is free for use, and that is why it was selected for further work. It allows one to create PC VR applications, compatible with all VR devices that work with the SteamVR desktop application (also produced by Valve). Therefore, most PC VR headsets are compatible with the created application (e.g., Meta Quest, HTC Vive, HP Reverb G2, and many others).
All the interactions are programmed either using SteamVR’s implemented code or the authors’ own scripts. Third party scripts are not modified. For visualization, a simulated texture of 3D printed “staircase” surface is applied as a combination of proper color and normal map settings.
The following assumptions are made for making the application:
  • The user starts inside a 3D visualization of a medical product development laboratory, which is created for this purpose by the authors.
  • In the beginning of the simulation, the product (prosthesis) is disassembled into pieces, and it is the user’s task to put them together—manually—to create a final product.
  • Movement in the working space is realized in two ways: using the user’s own movements and positional tracking (“on own legs”) or—for longer distances—via teleportation, operated by an analog stick of the VR controllers.
  • The user sees his simulated hands in the simulation. Interactions are realized using VR controllers—grasping the objects (parts) can be realized by touching them and holding either the “trigger” or the “grip” button of a controller representing a given hand.
  • Operations requiring use of tools in real life are simplified, and they happen automatically when the user puts a given part in a specific zone at the assembly basis.
  • It is possible to show helpers in the application—after interacting with a 3D cube in the work environment, assembly instructions are displayed, along with the names of all the parts.
  • It is impossible for the user to perform the assembly process in a wrong way due to imposed logical constraints. Additionally, it is not possible to disassemble the prosthesis after assembling it.
  • After finalizing the assembly, the user is congratulated, and the process ends. The user can operate the resulting prosthesis to try it on and to simulate movements.
Figure 7 shows the simulated environment, as well as the user’s starting location and initial state of the parts. Figure 8 shows the user while in the middle of the assembly procedure, aided with helpers. Figure 9 shows the completion of the assembly and a congratulation message for the user.
The application has been built as an executable, Windows application and tested that way during the experimental phase of research using Meta Quest/Quest 2 headsets.

2.3.3. Augmented Reality—Configurator

For the AR application, the Android operating system was selected as a target platform, with a tablet or a cellphone as the target devices. For basic AR functions, such as marker recognition, the free Vuforia package (PTC, Boston, MA, USA) was installed from inside the Unity package manager. Vuforia is a package that allows handling of target device’s built-in camera, as well as recognition of markers, as well as rendering 3D and content over them. It needs a license, but it can be obtained for free for educational purposes.
The application is built with the following assumptions:
  • The prosthesis is rendered over a single graphic marker, having the form of a randomized grayscale image (Figure 10).
  • When the prosthesis 3D model is visible, the user can operate a simple graphical user interface, with icons for prosthesis color change and additional icons for simple rotational animation in the elbow axis (Figure 10).
  • A second randomized graphic marker is used to render a 3D interface, with 3D icons enabling switching of modules of the prosthesis—once the icons are visible, the user can touch them to switch between different end effectors and sockets of the prosthesis (Figure 11).
  • The markers can be printed or displayed horizontally or vertically. Once a marker with the prosthesis is recognized, the prosthesis will stay in the camera’s field of view as long as environment does not change, so the prosthesis model can be rendered, even when the marker is not directly visible.
  • The markers can be also be moved, and the prosthesis and 3D UI models will move along. For virtual try-on, a marker can be placed on the user’s limb (or their stump, if an actual patient is using it).
The application has been deployed to some Android devices—7″ and 11″ tablets. For the actual experiments, the larger device is used for a better clarity of image.

2.3.4. Mixed Reality—Configurator and Try-On

As the previously described two applications, the MR application is built using the Unity engine. Just like the AR application, the Android platform is chosen as the preferred platform for development. However, this time, a headset is selected as the device of choice. The application is designed, deployed, and tested using the Meta Quest 2 goggles (Meta Platforms, Inc., Menlo Park, CA, USA), leveraging their unique feature of displaying the user’s real-world surroundings through the built-in cameras on the device’s screens, known as passthrough. The view is in grayscale, but it is enough for the user to fully notice and care about their surroundings—including their own body.
For building the application, a free Mixed Reality Toolkit package—MRTK (Microsoft Corporation, Redmond, WA, USA)—was used. The empty Unity project is converted to a Mixed Reality project by installing the MRTK package, along with hand interaction examples. Moreover, the Oculus Integration package by Meta was installed into the project, and the basic scene is prepared accordingly to enable passthrough in the Meta Quest device.
The application is built with the following assumptions:
  • The user starts in front of a 3D visualization of a product, surrounded by various user interface elements (both 3D and 2D icons)—as shown in Figure 12. The user can always see the passthrough layer (i.e., his real-world surroundings), but this image cannot be captured by means of screen or video capture—that is why, in all the Figures depicting MR, the background is black.
  • Using the user interface (UI), the user is able to make configuration changes—identical to the AR app (colors, modules)—but all the icons are 3D and operated through touch gestures.
  • Movement in the working space is realized using the user’s own movements and positional tracking (“on own legs”)—and, as the real world is all visible, no teleportation is possible. However, the user can “reset” his starting point to see the UI at the center of his location again (by using the Meta Quest context menu) or drag the whole scene contents with him. As such, there are no positional constraints, and the virtual prosthesis can be “taken” anywhere in the physical world.
  • The user sees his simulated hands in the simulation, superimposed on real hands (the rendered hands can be turned off). Interactions are realized using hand tracking and natural movements (grasping, pointing at objects from distance, tapping, etc.) VR controllers are also an option, if for some reason hand tracking does not work properly for a given user. If they are turned on and in the field of view, hand tracking is automatically disabled—and it is enabled once again when the controllers are put away.
  • The main 3D model—prosthesis—can be operated with use of two hands or one hand. One-hand interaction (Figure 13) allows one to drag around and rotate the object (the pivot depends on the grasping point), while two-hand interaction enables scaling (some limits were set to not allow the user to obtain extreme sizes of the prosthesis).
  • Using the UI, it is also possible to animate the rotating elbow of the prosthesis, as well as making an “exploded” view by using a 3D slider.
The application is deployed to a couple of Meta Quest 2 devices (set in developer’s mode to enable that) and tested only using that particular device type.

2.4. The Course of the Experimental Study

The experimental studies involved conducting organized tests on the capabilities of all three applications. These tests were carried out by an international group of experts who possess expertise, including engineering degrees, in areas relevant to design, mechatronics, and the development of medical products. The group was made of 14 people, aged between 22 and 65. The studies were organized as follows:
  • The first stage was the presentation of the application. The idea of the studies was presented to the experts, as well as the types of applications and their capabilities.
  • The experts could then familiarize themselves with all the XR technologies, playing with the devices and various interactive scenarios, if necessary (mostly for experts with no prior XR technology knowledge).
  • Following the experts’ familiarization with the technology, the actual testing phase was initiated. Each expert individually tested all the technologies sequentially, in the following order: VR → AR → MR. It was required for the entire group to complete testing of one application before moving on to the next.
  • In the VR application, Meta Quest goggles were primarily used for testing. The assembly scenario was realized once, until the final message, and the testers were able to stay a bit longer in VR space if they wanted to (time was not measured or otherwise constrained).
  • For AR application, an 11″ tablet was used for testing. The experts played with the application using all its options and testing various arrangements of printed and displayed markers. No time constraints were imposed.
  • Regarding the MR application, Meta Quest 2 goggles were used for testing, with full utilization of the hand-tracking technique. The experts played with the application until all the features were thoroughly tested (again, there were no time constraints).
  • After testing each technology, a dedicated survey was filled in. After finalizing all the tests, finally, a fourth survey was filled out (a comparison of different XR technology capabilities). The surveys contained mostly closed questions, using 5-point Likert scales for better visual comparison. Apart from that, there were open questions at the end of each survey.
  • During and after finalizing the whole testing procedure, various aspects were discussed, and the observations of experts were recorded.
Between testing particular technologies, certain pauses were made to allow for the possible effects of cyber sickness to wear off. The whole procedure of testing was realized during a single working day (8 h).
Figure 14 illustrates the particular phases of testing.
The first three surveys pertaining to particular technologies are divided into two parts:
  • general questions (four or three) to evaluate difficulty, effectiveness, usefulness, and simulator sickness symptoms (the last question is not asked in the case of AR, as no headset is used).
  • detailed technical questions, a bit different to each application, to evaluate, among other things, the quality of 3D and 2D graphics, fluency, intuitiveness, realism, and other interaction features.
Quantitatively, the surveys contain 11 (VR, MR) or 10 (AR) closed, 5-point Likert scale questions, with one open question (optional comments).
The fourth survey, for comparison of XR technologies, contains 5 questions. Each required an expert to make a ranking—writing 1, 2, and 3 numbers near the name of a technology. The 4 questions were to evaluate the potential in training, product design, and configuration, product use simulation, and the development of biomimetic mechatronic devices. The fifth question ranks the XR technologies by their general comfort and ease of use.
The surveys—full texts of questions as presented to the experts—are contained in Appendix A of this paper.

3. Results

3.1. XR Technologies Evaluation Results

The survey results were subjected to mathematical normalization, wherein a value of 1 represents the lowest (negative) assessment and a value of 5 represents the highest (positive) assessment. This normalization process was applied consistently across all questions, irrespective of their formulation.
The results of the first survey, pertaining to VR technology, are summarized in Table 1 (first part—general assessment) and Table 2 (second part—detailed assessment).
The following observations can be made regarding these results:
  • The difficulty of use has been assessed to be particularly low compared to all the other aspects of the VR technology. This might be caused by the relatively advanced assembly procedure being simulated, with the experts not having fluency in the use of VR goggles and controllers. A “stereotypical” correlation was observed here regarding age—younger testers assessed difficulty as lower than the older ones.
  • Almost no hard simulator sickness effects were observed—however, two persons had noticed some problems (such as headache, eye strain, and slight vertigo), and most of testers observed some slight effects. Anyhow, no one had to stop using the simulation due to symptoms, nor did they persist after the goggles were taken off. As such, simulation can be considered safe.
  • The effectiveness of learning was assessed positively, for the most part—most persons assessed it as “5”, a bit less “4”, and only some as “3”, with no lesser answers. It shows that, in the experts’ opinion, VR is actually as effective as traditional training, or better.
  • Usefulness in training was also assessed as high—a single answer of “3” was observed, with all other answers divided between “4” and “5”, meaning that, in the experts’ opinion, VR can partially or even fully replace real training procedures.
  • During the technical assessment, the quality of the 2D elements received the lowest rating among all the assessed features, although it still garnered a positive assessment as “rather good.” The only instances of “rather bad” feedback were provided by two individuals who expressed concerns about the intuitive and smooth interaction. It is worth noting that these same individuals had also rated the difficulty level in previous questions as “hard”. The other 12 experts assessed this aspect positively (no neutral opinions in this question). All the other technical aspects noted prevailing positive opinions among the experts, although the general average is closer to 4 than to a perfect 5.
The results of the second survey, pertaining to AR technology, are summarized in Table 3 (general assessment) and Table 4 (technical assessment).
The following observations can be made regarding these results:
  • Easiness of use of AR was assessed to be considerably high, which was expected, especially when comparing it to slightly problematic work with VR technology. However, it is still not as high, despite a very simplistic interface.
  • The effectiveness and the usefulness in the configuration process were assessed to be generally positive, but the feedback was not unequivocal. In general, AR was assessed as a useful product configuration tool, but, at this moment, it is likely not useful as a method to replace traditional means of communication with a customer. Two experts indicated that, in their opinion, it is rather useless and could only be used as a supplement or toy for the clients.
  • Technical assessment (the second part of survey) is positive, although in almost every question, there were some “rather bad” answers, and there were frequently more than one. In general, despite using exactly the same engine, product and rendering techniques (materials, lights, etc.), with a tablet of high-resolution screen, graphics quality was assessed to be lower than in the case of VR. Realism and blending with the real world were given the lowest score of all questions—still positive, but lots needed to improve in the experts’ opinion. In general, most experts evaluated the technical questions as “rather good” (4), with the average lowered by some having no opinion and some having a “rather bad” assessment.
The results of the third survey on MR technology are summarized in Table 5 (general assessment) and Table 6 (detailed technical assessment).
The following observations can be made regarding these results:
  • The easiness of operation of the MR equipment and application were generally assessed positively, although the score is still a bit below 4, with some persons answering that it was “hard” (2).
  • No persons reported considerable negative cyber sickness symptoms—they were generally low, and only one person scored it at 3 (neither small nor big symptoms), with the other experts evaluating it as small or very small.
  • In terms of the effectiveness in comparison with traditional product configuration and try-on, the experts also had a positive attitude—most answers were that it is “rather large” or “large” (4 or 5 score), with the occasional 3, meaning that no experts indicated MR as an ineffective tool—some just doubted that it could replace other methods.
  • Usefulness was also evaluated quite highly—half of experts agreed that it could even replace traditional methods of configuration and product try-on before buy. No one scored this aspect below 3.
  • In terms of technical evaluation, it is noteworthy that the comfort of use was scored as very good, mostly—no one selected a score below 4. All the other aspects were generally positively assessed—however, the quality of two-dimensional elements was the weakest aspect of the visualization, with some experts indicating poor readability of the text. Additionally, some persons mentioned that passthrough quality was not high, which is understandable due to its grayscale aspect. The fluency of visualization and animations was also scored high. In general, this part of the survey gathered positive results towards the technology.
In terms of the open questions, only a few experts filled in written answers (most preferred to vocalize their comments while testing). The comments were as follows:
  • The AR application looks nice, but the interface is not very responsive.
  • The MR application was difficult for interaction at first, but it became better with some practice; it also has readability problems (text elements).
  • The MR passthrough in Meta Quest 2 is blurry and not good—it would be better to utilize transparent screen glasses, such as the HoloLens device.

3.2. XR Technologies Comparison Results

Results of the fourth survey—the comparison of technologies—are presented in Table 7, as well as in Figure 15. When summarizing the results, the following assumptions were considered:
  • In each survey question, ranking was translated to points awarded to technologies in a category decided by the question—2 points were given if technology was ranked at no.1, 1 point when at no.2 and 0 points were given when the technology was ranked at no.3.
  • If all technologies were ranked equal in each question, they were all awarded 1 point.
  • If only one technology was marked, it was also awarded 1 point.
Table 7 and diagram in Figure 15 present rankings constructed on the total amount of points calculated this way, as awarded by the experts. Summarized ranking points are visualized in a pie chart in Figure 16.
The sum of the experts’ indications points out a clear result—that mixed reality technology was ranked highest, even in the categories that the presented application did not pertain to (training). Virtual Reality is a relatively close second, with Augmented Reality also awarded many points, but ultimately always ranked last. Although it is noteworthy that particular scores of experts were very different, no survey was the same, and each expert had a slightly different opinion from each other. The only categories with a slightly different results are general comfort and ease of use, where VR and MR were ranked equal—which is not surprising, as almost the same equipment was used (a headset).
In terms of the results of particular technologies compared between different application categories (points gathered in the first three questions of the survey), VR obtained the most points in its potential for design aiding, while AR and MR had the most points in the use simulation category.
Additionally, for the full picture, the answers to comparable questions of single technology surveys (first three surveys) have been presented in diagrams in Figure 17 and Figure 18. Technical question no. 6 of the detailed part of each survey was omitted from this comparison, as it is very specific and incomparable.
Of these results, it is apparent that AR technology was the one with the poorest score of all three in most categories. The MR technology seems to be slightly better than the VR, or comparable in most categories. VR prevails in three-dimensional and two-dimensional content quality and realism—which is due to better performance of PC VR systems (over the mobile Android used on Meta Quest headsets in standalone mode) and better quality and resolution of rendering. MR prevails in easiness, comfort, and fluency, with the same intuitiveness—and this is probably because of a lack of controllers (hand tracking) and generally more simplistic interaction methods. Efficiency and usefulness are very comparable—so potential, at least in numbers, is assessed as very similar. AR was scored considerably lower, despite evaluation still being positive. It prevails over other technologies only in easiness of use and lack of potential simulator sickness effect.

4. Discussion

In the performed experiments, the VR and MR technologies gathered more positive and enthusiastic feedback than the AR technology. This could be attained by their real efficiency and capabilities, as well as limitations of the simple AR technology implemented (tablet with simple image markers recognition). Part of that increased score, in the authors’ opinion, comes from the so-called “wow effect” of immersive technologies on inexperienced users. However, this does not mean that the expert evaluation is less valid, especially in the presented use case scenario (inexperienced factory workers—training in VR—or inexperienced clients—MR).
The XR technologies are still relatively unknown to the general public, and, as such, their use will be often met with the “wow” effect, bringing more positive evaluation of the technology itself, as well as a more enthusiastic approach to the presented product. That is very important when working with prosthetic patients, as many of them (especially children) do not accept their prostheses and have a hard time working with them and using them, which can further deteriorate their health and wellbeing [32]. That is why using simple applications, even in form of games or visualizations with the “fun factor”, combined with immersive equipment (VR/MR goggles), can potentially increase the acceptance and use rate of prosthetic devices among younger patients.
It was apparent from the technology comparison and ranking that Mixed Reality did the best impression on all the experts—and this might be caused by many factors—the standalone aspect of the application (mobility), maintaining contact with real world, intuitive and enjoyable manipulation by hand tracking, simplicity, etc.
A very important aspect of the tested applications is their ease of use. It was noted that all the experts were able to utilize all the features of the presented applications, if not with some minor difficulties. In terms of immersive technologies (VR/MR), it was noted that going into hand tracking, as opposed to use of controllers, had a positive effect on both perceived effectiveness and usefulness, as well as fluency and intuitiveness of interaction. It was also noted by the researchers that the time required to familiarize one’s self with the MR technology with hand tracking was considerably shorter than in the case of VR with controllers and their buttons. In the actual experiment, the difference was very apparent—but, in the scores resulting from the survey, it is not that significant (VR is slightly worse, but there is no huge difference in numbers).
However, it must be mentioned that some experts indicated that the quality of two-dimensional elements could be improved. Some of them also had troubles with slight or (rarely) moderate symptoms of cyber sickness while using immersive equipment. The surveys were anonymous, but from authors’ observations, correlation of age and “digital” experience is visible—older users, with less experience in digital technologies, tend to have more problems with perceiving user interface elements and were more prone to cyber sickness symptoms. However, the statistical sample was too small to actually prove that in the study—further, more detailed experiments are required here, with more numerous groups of users.
In the general evaluation of XR technology, based on the study results, it is quite important to notify that the experts were very positive about the prospect of XR technologies possibly replacing traditional ways many processes in a personalized product lifecycle are realized. More than half of the responders indicated that either VR or MR could replace, entirely, the traditional training or configuration with use of physical products. Such answers were also given (albeit more rarely) in the case of AR. That is a very promising direction in terms of development of XR technology, but the observed reality is far from that—although XR is present in the media and known by many, it is still relatively rarely encountered in the lifecycle of many products, especially considering prosthetics.
Results of the studies presented in the paper should be treated as an insight that this situation might be changing soon and that the technology is almost ready to be popularized and included as often as possible, especially as a means of communication—between designers, manufacturers, doctors, patients, and salesmen. Therefore, the experiments confirm that the XR in the view of experts is a technology that can be used for real work, not only as an attractor and a marketing device. According to the experts, and in authors’ opinion, the XR technologies presented in this paper will be used in the future in all medical centers, as well as in the educational field.

5. Conclusions

This paper presented an experimental evaluation of the applications of virtual, augmented and mixed reality (VR, AR, MR) technologies in the context of individualized three-dimensionally printed upper limb prostheses. Through complex and advanced applications, subjected to experimental verification, the authors aimed to provide a comprehensive understanding of the benefits, challenges, and future directions of these technologies in the field. It has been found that each XR technology has the potential to be used efficiently in the field of development of biomechatronic prosthetic devices. However, it is also quite personal, as experts had various opinions on each technology and their potentials (most Likert scale questions in the surveys gathered answers between 2 and 5). According to the authors’ observations, the most efficient technology is VR used for training purposes, but due to its many technical issues, it requires long familiarization. On the other hand, AR, and also MR, are quick to dive into and could be interesting in producer-customer relations, even more than VR, but their performance and effectiveness may be not that high, especially in their current states of development. It is, therefore, important to select the technology, each time, for a given case, and to fully consider use scenarios before deciding on the use of one of them—but it is important that they are all considered and should be thought as a standard part of the process of product development.
The presented studies might be some of the first steps in that direction, which will be further explored in future studies, also involving real patients. By disseminating this knowledge, the authors hope to stimulate further research and innovation, fostering advancements in the development of individualized prosthetic solutions and ultimately benefiting patients worldwide. Advances in artificial intelligence, augmented reality, and virtual reality mean fresh AR/VR design opportunities for designers. Recognizing the innovations that are occurring in the medicine and in the industry will help designers design smarter and make more informed creative decisions.
In future studies, the authors will focus on a more detailed approach regarding particular XR technologies. In the case of VR technology, its training capabilities in the case of manufacturing and assembly of personalized prosthetics and orthotics will be compared with real-life training scenario. In the case of AR technology, numerous more advanced object recognition methods will be explored, as opposed to the simple marker approach presented in the basic studies. In the case of MR technology, video see-through will be put against the optical see-through approach. In all technologies, it is also planned to explore the potential benefits of collaborative approach. Long-time studies are planned that will hopefully continue the development of XR technologies in medical production. The integration with a fully-immersive hapto-audio-visual framework may be also considered in the future [33,34].

Author Contributions

Conceptualization, F.G.; methodology, F.G.; software, F.G.; validation, R.P., F.S.; formal analysis, F.G.; investigation, F.G., M.Ż., R.Ł., M.O., F.S., D.-I.B. and R.P.; resources, F.G., M.Z.; data curation, R.Ł. and F.G.; writing—original draft preparation, F.G.; writing—review and editing, F.G., F.S., D.-I.B. and R.P.; visualization, F.G. and R.Ł.; supervision, F.G., R.P., F.S. and D.-I.B.; funding acquisition, F.G., R.P., D.-I.B., F.S. and M.Z. All authors have read and agreed to the published version of the manuscript.

Funding

The studies (whole paper, except of Section 2.2) were realized in scope of the grant entitled “European network for 3D printing of biomimetic mechatronic systems”, supported by EEA grants (Project No: 21-COP-0019, contract no. 541/15 February 2022). The system development, described in Section 2.2, was funded by the Polish National Center for Research and Development in the scope of the “LIDER” program (grant agreement no. LIDER/14/0078/L-8/16/NCBR/2017).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki. It did not involve intervention work on human subjects. Development of the basic system and three-dimensional models was a result of a process, which was approved by the Ethics Committee of Poznan University of Medical Sciences (protocol code no. 1200/17, 7 December 2017).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The detailed data presented in this study (i.e., survey results) are available on request from the corresponding author. The data are not publicly available due to privacy restrictions.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study, in the collection, analyses, or interpretation of data, in the writing of the manuscript, or in the decision to publish the results.

Appendix A. Full Texts of Realized Surveys

Survey no. 1—VR training application (prosthesis assembly)
1.
Determine the level of difficulty of using the presented VR training system during tests:
very easy (nothing was a problem for me)
easy (some activities were difficult for me)
neither easy nor difficult (I could proceed but I had many difficulties)
hard (I had a lot of problems the whole time I was using VR)
very difficult (I wasn’t able to use the VR app effectively)
2.
Determine the degree of discomfort associated with the appearance of simulator sickness symptoms while using the VR training system
very small (no negative symptoms)
small (some symptoms may have appeared but didn’t influence VR training)
neither small nor big (some symptoms appeared but influence was small)
big (negative symptoms made it hard to use the system)
very big (I was unable to finish the training)
3.
Evaluate the effectiveness and speed of learning in VR compared to traditional training:
large (VR allows you to gain knowledge much faster and more effectively than during traditional training)
rather large (VR allows you to gain knowledge faster and more effectively than during traditional training)
neither big nor small (VR allows you to gain knowledge in a similar time as during a traditional training)
rather small (traditional training is faster and more effective than in VR)
small (VR training is ineffective)
4.
Evaluate the usefulness of the VR system in the training process:
very useful (VR can fully replace real training on a real object, omitting some manual activities that cannot be learned in it)
useful (VR can partly replace real training, but it is necessary to complete it on a real object)
rather useful (VR can be an effective and helpful supplement to traditional training, but it will not replace it)
rather useless (VR can only be an introduction to traditional training)
useless (training in VR will not bring any benefits)
How do you rate (on a scale of 1–5) individual aspects of the presented VR application?
BadRather BadNo OpinionRather GoodVery Good
1.Comfort of using the helmet (adjustment to the head, eye spacing, image sharpness)12345
2.Quality of three-dimensional graphics12345
3.Quality and convenience of reading two-dimensional content (flat drawings, inscriptions)12345
4.Fluency while looking around12345
5.Smooth and intuitive interaction with objects12345
6.Navigation in the virtual space using the controller12345
7.The realism of the presented situations and objects, the impression of being present inside the virtual scene12345
Optional VR technology comments (helmet, controllers, movement, fluency, field of view, interactions, etc.):
Survey no. 2—AR application (product configurator)
1.
Determine the level of difficulty of using the presented AR application during tests:
very easy (nothing was a problem for me)
easy (some activities were difficult for me)
neither easy nor difficult (I could proceed but I had many difficulties)
hard (I had a lot of problems the whole time I was using AR)
very difficult (I wasn’t able to use the AR app effectively)
2.
Evaluate the effectiveness and speed of work in AR compared to traditional product configuration:
large (AR allows for very fast product configuration, no other methods are necessary)
rather large (AR is an efficient tool in configuration, but other methods also can be used)
neither big nor small (AR is a medium-level effective tool for product configuration)
rather small (in terms of effectiveness I’d prefer other methods for product configuration)
small (AR configuration is not effective at all)
3.
Evaluate the usefulness of the AR application in the product configuration process:
very useful (AR can replace traditional methods—such as product catalogues etc.)
useful (AR can partly replace traditional methods of configuration)
rather useful (AR can be an effective and helpful supplement to traditional methods of product configuration but cannot replace them)
rather useless (AR can be a supplement/toy for the clients only)
useless (AR does not bring any benefit to the configuration process)
How do you rate (on a scale of 1–5) individual aspects of the presented AR application?
BadRather BadNo OpinionRather GoodVery Good
1.Comfort of using the tablet12345
2.Quality of three-dimensional graphics12345
3.Quality of two-dimensional UI12345
4.Fluency of visualizations and animations12345
5.Smooth and intuitive interaction with objects12345
6.Markers recognition and proper placement of three-dimensional models over markers12345
7.The realism of the presented objects, impression of them smoothly mixing with the real world12345
Optional AR technology comments (tablet, interface, markers, etc.):
Survey no. 3—MR application (product configurator and try-on)
1.
Determine the level of difficulty of using the presented MR application during tests:
very easy (nothing was a problem for me)
easy (some activities were difficult for me)
neither easy nor difficult (I could proceed but I had many difficulties)
hard (I had a lot of problems the whole time I was using AR)
very difficult (I wasn’t able to use the AR app effectively)
2.
Determine the degree of discomfort associated with the appearance of simulator sickness symptoms while using the MR training system:
very small (no negative symptoms)
small (some symptoms may have appeared but didn’t influence the process)
neither small nor big (some symptoms appeared but influence was small)
big (negative symptoms made it hard to use the system)
very big (I was unable to work with the product)
3.
Evaluate the effectiveness and speed of work in MR compared to traditional product configuration and try-on:
large (MR allows for very fast product configuration, no other methods are necessary)
rather large (MR is an efficient tool in configuration, but other methods also can be used)
neither big nor small (MR is a medium-level effective tool for product configuration)
rather small (in terms of effectiveness I’d prefer other methods for product configuration)
small (MR configuration is not effective at all)
4.
Evaluate the usefulness of the AR application in the product configuration and try-on process:
very useful (MR can replace traditional methods of configuration and try-on entirely)
useful (MR can partly replace traditional methods of configuration and try-on)
rather useful (MR can be an effective and helpful supplement to traditional methods of product configuration but cannot replace them)
rather useless (MR can be a supplement/toy for the clients only)
useless (MR does not bring any benefit to the configuration and try-on process)
How do you rate (on a scale of 1–5) individual aspects of the presented MR application?
BadRather BadNo OpinionRather GoodVery Good
1.Comfort of using the MR goggles12345
2.Quality of three-dimensional graphics12345
3.Quality and readability of two-dimensional UI elements12345
4.Fluency of visualizations and animations12345
5.Smooth and intuitive interaction with objects using hand tracking12345
6.Passthrough quality (visibility of real world environment)12345
7.The realism of the presented objects, impression of them smoothly mixing with the real world12345
Optional MR technology comments (helmet, hand tracking, passthrough, etc.):
Survey no. 4—XR technology comparison
In the questions below, rank the technologies by your opinion on given subject, writing 1, 2 and 3 near the name of technology.
1.
Assess the potential of XR technologies in terms of training (such as operation of production machines, assembly processes etc.):
Virtual Reality
Augmented Reality
Mixed Reality
2.
Assess the potential of XR technologies in terms of product design and configuration (such as realistic visualization of the product, change in its design and material features to clients’ needs etc.)
Virtual Reality
Augmented Reality
Mixed Reality
3.
Assess the potential of XR technologies in terms of product use simulation (such as realistic visualization, look and feel of the product, try-on for ergonomic purposes, learning how to use the product etc.)
Virtual Reality
Augmented Reality
Mixed Reality
4.
Assess the general potential of XR technologies in terms of development of biomimetic mechatronic devices (such as 3D printed prostheses, orthoses etc.)
Virtual Reality
Augmented Reality
Mixed Reality
5.
Assess the general comfort and ease of use of XR technologies:
Virtual Reality
Augmented Reality
Mixed Reality

References

  1. Ten Kate, J.; Smit, G.; Breedveld, P. 3D-printed upper limb prostheses: A review. Disabil. Rehabil. Assist. Technol. 2017, 12, 300–314. [Google Scholar] [CrossRef]
  2. Barrios-Muriel, J.; Romero-Sánchez, F.; Alonso-Sánchez, F.J.; Salgado, D.R. Advances in Orthotic and Prosthetic Manufacturing: A Technology Review. Materials 2020, 13, 295. [Google Scholar] [CrossRef] [Green Version]
  3. Olsen, J.; Day, S.; Dupan, S.; Nazarpour, K.; Dyson, M. 3D-Printing and Upper-Limb Prosthetic Sockets: Promises and Pitfalls. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 527–535. [Google Scholar] [CrossRef]
  4. Phelan, I.; Arden, M.; Garcia, C.; Roast, C. Exploring virtual reality and prosthetic training. In Proceedings of the 2015 IEEE Virtual Reality (VR), Arles, France, 23–27 March 2015; pp. 353–354. [Google Scholar]
  5. Xu, K.; Qin, S. An Interdisciplinary Approach and Advanced Techniques for Enhanced 3D-Printed Upper Limb Prosthetic Socket Design: A Literature Review. Actuators 2023, 12, 223. [Google Scholar] [CrossRef]
  6. Górski, F.; Wichniarek, R.; Kuczko, W.; Żukowska, M. Study on Properties of Automatically Designed 3D-Printed Customized Prosthetic Sockets. Materials 2021, 14, 5240. [Google Scholar] [CrossRef]
  7. Diment, L.E.; Thompson, M.S.; Bergmann, J.H. Three-dimensional printed upper-limb prostheses lack random-ised controlled trials: A systematic review. Prosthet. Orthot. Int. 2018, 42, 7–13. [Google Scholar] [CrossRef] [Green Version]
  8. Gorski, F.; Bun, P.; Stefanska, K. Use of virtual mirror in design of individualized orthopedic supplies. In Innovations in Mechanical Engineering; Springer International Publishing: Cham, Switzerland, 2022; pp. 388–394. [Google Scholar]
  9. Kyaw, B.M.; Saxena, N.; Posadzki, P.; Vseteckova, J.; Nikolaou, C.K.; George, P.P.; Divakar, U.; Masiello, I.; Kononowicz, A.A.; Zary, N.; et al. Virtual Reality for Health Professions Education: Systematic Review and Meta-Analysis by the Digital Health Education Collaboration. J. Med. Internet Res. 2019, 21, e12959. [Google Scholar] [CrossRef] [Green Version]
  10. Minoufekr, M.; Plapper, P. Holo Study –Mixed Reality Framework for Industrial Engineering Education and Training. Trends Tech. Sci. Res. 2020, 4, 156–160. [Google Scholar] [CrossRef]
  11. Verhey, J.T.; Haglin, J.M.; Verhey, E.M.; Hartigan, D.E. Virtual, augmented, and mixed reality applications in orthopedic surgery. Int. J. Med. Robot. Comput. Assist. Surg. 2020, 16, e2067. [Google Scholar] [CrossRef]
  12. Khundam, C.; Vorachart, V.; Preeyawongsakul, P.; Hosap, W.; Noël, F. A comparative study of interaction time and usability of using controllers and hand tracking in virtual reality training. Informatics 2021, 8, 60. [Google Scholar] [CrossRef]
  13. Avila-Garzon, C.; Bacca-Acosta, J.; Duarte, J.; Betancourt, J. Augmented Reality in Education: An Overview of Twenty-Five Years of Research. Contemp. Educ. Technol. 2021, 13, ep302. [Google Scholar] [CrossRef]
  14. Egger, J.; Masood, T. Augmented reality in support of intelligent manufacturing—A systematic literature review. Comput. Ind. Eng. 2019, 140, 106195. [Google Scholar] [CrossRef]
  15. Azuma, R.; Baillot, Y.; Behringer, R.; Feiner, S.; Julier, S.; MacIntyre, B. Recent advances in augmented reality. IEEE Comput. Graph. Appl. 2001, 21, 34–47. [Google Scholar] [CrossRef] [Green Version]
  16. Milgram, P.; Kishino, F. A taxonomy of mixed reality visual displays. IEICE Trans. Inf. Syst. 1994, 77, 1321–1329. [Google Scholar]
  17. Viglialoro, R.M.; Condino, S.; Turini, G.; Carbone, M.; Ferrari, V.; Gesi, M. Augmented Reality, Mixed Reality, and Hybrid Approach in Healthcare Simulation: A Systematic Review. Appl. Sci. 2021, 11, 2338. [Google Scholar] [CrossRef]
  18. Zawadzki, P.; Żywicki, K. Smart Product Design and Production Control for Effective Mass Customization in the Industry 4.0 Concept. Manag. Prod. Eng. Rev. 2016, 7, 105–112. [Google Scholar] [CrossRef]
  19. Eswaran, M.; Bahubalendruni, M.R. Challenges and opportunities on AR/VR technologies for manufacturing systems in the context of industry 4.0: A state of the art review. J. Manuf. Syst. 2022, 65, 260–278. [Google Scholar] [CrossRef]
  20. Minopoulos, G.M.; Memos, V.A.; Stergiou, K.D.; Stergiou, C.L.; Psannis, K.E. A Medical Image Visualization Technique Assisted with AI-Based Haptic Feedback for Robotic Surgery and Healthcare. Appl. Sci. 2023, 13, 3592. [Google Scholar] [CrossRef]
  21. Bates, T.J.; Fergason, J.R.; Pierrie, S.N. Technological Advances in Prosthesis Design and Rehabilitation Following Upper Extremity Limb Loss. Curr. Rev. Musculoskelet. Med. 2020, 13, 485–493. [Google Scholar] [CrossRef]
  22. Sahija, D. Critical review of mixed reality integration with medical devices for patientcare. Int. J. Innov. Res. Multidiscip. Field 2022, 10, 100–105. [Google Scholar]
  23. Sharma, A.; Hunt, C.L.; Maheshwari, A.; Osborn, L.; Levay, G.; Kaliki, R.R.; Soares, A.B.; Thakor, N. A mixed-reality training environment for upper limb prosthesis control. In Proceedings of the 2018 IEEE Biomedical Circuits and Systems Conference (BioCAS), Cleveland, OH, USA, 17–19 October 2018; pp. 1–4. [Google Scholar]
  24. Hauschild, M.; Davoodi, R.; Loeb, G.E. A Virtual Reality Environment for Designing and Fitting Neural Prosthetic Limbs. IEEE Trans. Neural Syst. Rehabil. Eng. 2007, 15, 9–15. [Google Scholar] [CrossRef] [PubMed]
  25. Putrino, D.; Wong, Y.T.; Weiss, A.; Pesaran, B. A training platform for many-dimensional prosthetic devices using a virtual reality environment. J. Neurosci. Methods 2015, 244, 68–77. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Mukhopadhyay, A.; Sharma, V.K.; Tatyarao, P.G.; Shah, A.K.; Rao, A.M.C.; Subin, P.R.; Biswas, P. A comparison study between XR interfaces for driver assistance in take over request. Transp. Eng. 2023, 11, 100159. [Google Scholar] [CrossRef]
  27. Doolani, S.; Wessels, C.; Kanal, V.; Sevastopoulos, C.; Jaiswal, A.; Nambiappan, H.; Makedon, F. A review of ex-tended reality (xr) technologies for manufacturing training. Technologies 2020, 8, 77. [Google Scholar] [CrossRef]
  28. Górski, F.; Sahaj, N.; Kuczko, W.; Hamrol, A.; Żukowska, M. Risk Assessment of Individualized 3D Printed Pros-theses Using Failure Mode and Effect Analysis. Adv. Sci. Technol. Res. J. 2022, 16, 189–200. [Google Scholar] [CrossRef]
  29. Górski, F.; Wichniarek, R.; Kuczko, W.; Żukowska, M.; Lulkiewicz, M.; Zawadzki, P. Experimental Studies on 3D Printing of Automatically Designed Customized Wrist-Hand Orthoses. Materials 2020, 13, 4091. [Google Scholar] [CrossRef]
  30. Zawadzki, P.; Wichniarek, R.; Kuczko, W.; Slupińska, S.; Żukowska, M. Automated Design and Rapid Manu-facturing of Low-Cost Customized Upper Limb Prostheses. J. Phys. Conf. Ser. 2022, 2198, 012040. [Google Scholar]
  31. Komorowska, O. Automation of Design of Modular Upper Limb Prosthesis. Master’s Thesis, Poznan University of Technology, Poznań, Poland, 2022. [Google Scholar]
  32. Biddiss, E.A.; Chau, T.T. Multivariate prediction of upper limb prosthesis acceptance or rejection. Disabil. Rehabil. Assist. Technol. 2008, 3, 181–192. [Google Scholar] [CrossRef]
  33. Sanfilippo, F.; Hatledal, L.I.; Pettersen, K.Y. A Fully-Immersive Hapto-Audio-Visual Framework for Remote Touch. In Proceedings of the 11th IEEE International Conference on Innovations in Information Technology (IIT’15), Dubai, United Arab Emirates, 1–3 November 2015. [Google Scholar]
  34. Sanfilippo, F.; Pacchierotti, C. A Wearable Haptic System for the Health Monitoring of Elderly People in Smart Cities. Int. J. Online Eng. (iJOE) 2018, 14, 52–66. [Google Scholar] [CrossRef]
Figure 1. Modular hand prosthesis—basic product considered in the studies, manufactured using Fused Filament Fabrication 3D printing technology, out of PETG material [28].
Figure 1. Modular hand prosthesis—basic product considered in the studies, manufactured using Fused Filament Fabrication 3D printing technology, out of PETG material [28].
Applsci 13 08035 g001
Figure 2. Plan of the research.
Figure 2. Plan of the research.
Applsci 13 08035 g002
Figure 3. Implementation of the prosthesis in real-life use for child and adult patients [28].
Figure 3. Implementation of the prosthesis in real-life use for child and adult patients [28].
Applsci 13 08035 g003
Figure 4. AutoMedPrint—scheme of the work [30].
Figure 4. AutoMedPrint—scheme of the work [30].
Applsci 13 08035 g004
Figure 5. AutoMedPrint prototype—real life use (demonstrated by authors), (a) 3D scanning of limb, (b) use of a 3D configurator tool.
Figure 5. AutoMedPrint prototype—real life use (demonstrated by authors), (a) 3D scanning of limb, (b) use of a 3D configurator tool.
Applsci 13 08035 g005
Figure 6. Basic 3D data set: (a) modular prosthesis for configuration; (b) single variant for assembly instructions.
Figure 6. Basic 3D data set: (a) modular prosthesis for configuration; (b) single variant for assembly instructions.
Applsci 13 08035 g006
Figure 7. Initial scene view, visible environment and all prosthesis parts.
Figure 7. Initial scene view, visible environment and all prosthesis parts.
Applsci 13 08035 g007
Figure 8. In-game view during the assembly procedure, helpers visible.
Figure 8. In-game view during the assembly procedure, helpers visible.
Applsci 13 08035 g008
Figure 9. Finishing of the assembly procedure.
Figure 9. Finishing of the assembly procedure.
Applsci 13 08035 g009
Figure 10. Basic AR view of the application—3D model rendered over a graphic marker, with simple configuration, and the animation UI is visible.
Figure 10. Basic AR view of the application—3D model rendered over a graphic marker, with simple configuration, and the animation UI is visible.
Applsci 13 08035 g010
Figure 11. 3D configuration interface visible over the second marker, with different module configuration.
Figure 11. 3D configuration interface visible over the second marker, with different module configuration.
Applsci 13 08035 g011
Figure 12. MR visualization—basic starting interface.
Figure 12. MR visualization—basic starting interface.
Applsci 13 08035 g012
Figure 13. Manipulation of the prosthesis model using hand tracking capabilities, (a) one-hand rotation and movement, (b) two-hand scaling.
Figure 13. Manipulation of the prosthesis model using hand tracking capabilities, (a) one-hand rotation and movement, (b) two-hand scaling.
Applsci 13 08035 g013
Figure 14. Testing three applications: (a) VR, interaction with controllers; (b) AR, touch interaction with tablet; (c) MR, hand tracking interaction.
Figure 14. Testing three applications: (a) VR, interaction with controllers; (b) AR, touch interaction with tablet; (c) MR, hand tracking interaction.
Applsci 13 08035 g014
Figure 15. The XR technologies ranking in various categories.
Figure 15. The XR technologies ranking in various categories.
Applsci 13 08035 g015
Figure 16. The XR technologies’ general ranking points distribution.
Figure 16. The XR technologies’ general ranking points distribution.
Applsci 13 08035 g016
Figure 17. Comparison of the general assessment of three surveys of XR technologies.
Figure 17. Comparison of the general assessment of three surveys of XR technologies.
Applsci 13 08035 g017
Figure 18. Comparison of the detailed assessment of three surveys of XR technologies.
Figure 18. Comparison of the detailed assessment of three surveys of XR technologies.
Applsci 13 08035 g018
Table 1. Results of the VR technology evaluation—general assessment.
Table 1. Results of the VR technology evaluation—general assessment.
No.QuestionMeanMinMax
1Difficulty (easiness)3.6425
2Simulator sickness4.4325
3Effectiveness of learning4.1435
4Usefulness in training4.5035
Average:4.18
Table 2. Results of the VR technology evaluation—detailed assessment.
Table 2. Results of the VR technology evaluation—detailed assessment.
No.QuestionMeanMinMax
1Comfort of use4.2945
2Quality of 3D elements4.3645
3Quality of 2D elements4.0753
4Fluency—look around4.3635
5Intuitive interaction4.2125
6Navigation with controller4.6435
7Realism of situations and objects4.2145
Average:4.31
Table 3. Results of the AR technology evaluation—general assessment.
Table 3. Results of the AR technology evaluation—general assessment.
No.QuestionMeanMinMax
1Difficulty (easiness)4.00
2Effectiveness of work3.86
3Usefulness in configuration4.36
Average:4.07
Table 4. Results of the AR technology evaluation—detailed assessment.
Table 4. Results of the AR technology evaluation—detailed assessment.
No.QuestionMeanMinMax
1Comfort of use3.7925
2Quality of 3D elements3.6425
3Quality of 2D elements3.8635
4Fluency—visualization3.9325
5Intuitive interaction3.8625
6Marker recognition quality3.6425
7Realism of objects and blending with real world3.5725
Average:3.76
Table 5. Results of the MR technology evaluation—general assessment.
Table 5. Results of the MR technology evaluation—general assessment.
No.QuestionMeanMinMax
1Difficulty (easiness)3.9325
2Simulator sickness4.5035
3Effectiveness of work4.1435
4Usefulness in configuration and try-on4.4335
Average:4.25
Table 6. Results of the MR technology evaluation—detailed assessment.
Table 6. Results of the MR technology evaluation—detailed assessment.
No.QuestionMeanMinMax
1Comfort of use4.7145
2Quality of 3D elements4.2135
3Quality of 2D elements3.9325
4Fluency—visualization4.5735
5Intuitive interaction4.2135
6Passthrough quality4.2125
7Realism of objects and blending with real world4.1435
Average:4.29
Table 7. Results of the XR technology comparison—ranking points.
Table 7. Results of the XR technology comparison—ranking points.
Question/Technology1—Training2—Design3—Use Simulation4—Product Develop.5—Ease, ComfortTotal
VR12 (2)13 (2)11 (2)12 (2)15 (1)63 (2)
AR8 (3)7 (3)9 (3)7 (3)8 (2)39 (3)
MR18 (1)17 (1)19 (1)19 (1)15 (1)88 (1)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Górski, F.; Łabudzki, R.; Żukowska, M.; Sanfilippo, F.; Ottestad, M.; Zelenay, M.; Băilă, D.-I.; Pacurar, R. Experimental Evaluation of Extended Reality Technologies in the Development of Individualized Three-Dimensionally Printed Upper Limb Prostheses. Appl. Sci. 2023, 13, 8035. https://doi.org/10.3390/app13148035

AMA Style

Górski F, Łabudzki R, Żukowska M, Sanfilippo F, Ottestad M, Zelenay M, Băilă D-I, Pacurar R. Experimental Evaluation of Extended Reality Technologies in the Development of Individualized Three-Dimensionally Printed Upper Limb Prostheses. Applied Sciences. 2023; 13(14):8035. https://doi.org/10.3390/app13148035

Chicago/Turabian Style

Górski, Filip, Remigiusz Łabudzki, Magdalena Żukowska, Filippo Sanfilippo, Morten Ottestad, Martin Zelenay, Diana-Irinel Băilă, and Razvan Pacurar. 2023. "Experimental Evaluation of Extended Reality Technologies in the Development of Individualized Three-Dimensionally Printed Upper Limb Prostheses" Applied Sciences 13, no. 14: 8035. https://doi.org/10.3390/app13148035

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop