Next Article in Journal
Characterization of Ancient Marquetry Using Different Non-Destructive Testing Techniques
Next Article in Special Issue
FlowHaptics: Mid-Air Haptic Representation of Liquid Flow
Previous Article in Journal
Multi-Modal Emotion Recognition Using Speech Features and Text-Embedding
Previous Article in Special Issue
A Subjective and Behavioral Assessment of Affordances in Virtual Architectural Walkthroughs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Needs-Based Augmented Reality System

School of Engineering Sciences, Lappeenranta-Lahti University of Technology, Yliopistonkatu 34, 53850 Lappeenranta, Finland
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(17), 7978; https://doi.org/10.3390/app11177978
Submission received: 18 July 2021 / Revised: 13 August 2021 / Accepted: 17 August 2021 / Published: 28 August 2021
(This article belongs to the Special Issue Extended Reality: From Theory to Applications)

Abstract

:
Augmented Reality aims to enhance the real world with computer-generated information. AR technology is both attractive and promising. Current AR experiences depend on external elements to launch, such as markers, images, and location. For an AR experience to be more personalized, this research proposes a scheme to trigger AR experiences based on human needs. This approach should enable human needs to be captured, and analyze them to select the most suited experiences that fulfill or aids in fulfilling needs. The contribution of this paper includes (1) a study of current AR technologies and triggers, (2) an analysis of human needs into measurable elements, and (3) a description of a needs-based AR application process with a demonstration of the process guidelines. The research presents a proof of concept prototype of a restaurant that satisfies the subsistence need for hunger. The results show the effectiveness of the guidelines in detecting human needs and recommending AR experiences; however, producing correct predictions and recommendations requires a well-established dataset.

Graphical Abstract

1. Introduction

A classical survey on augmented reality (AR) describes it as “AR supplements reality rather than completely replacing it” [1].
Augmented Reality (AR) is a technology that enhances a real environment with computer-generated information using different sensory modalities, including visual, auditory, haptic, and olfactory. AR aims to simplify users’ lives by bringing virtual information to their attention [2]. An Augmented Reality (AR) system embodies the following properties [3]:
  • Enhances real environments by adding virtual objects.
  • Works in real-time and provides interactivity.
  • Provides the correct placement of virtual objects within the environment.
AR has many application areas such as education and learning [4], entertainment and gaming [5], food and beverage industry [6], health care [7], manufacturing [8], museums [9], space exploration [10], and tourism [11].
The study of augmented reality incorporates many computing areas such as tracking, interaction, display, mobile AR, authoring, visualization, calibration, and rendering [12]. AR literature focuses on the development of the technology components; however, little attention is devoted to examining personalized experiences. This research aims to enrich augmented reality paradigms by developing a novel AR experience trigger: a needs trigger. Despite the importance of needs and their satisfaction in human life, there is still a shortage of incorporating human needs in information systems and tools [13]. Augmented reality is a growing field that may benefit from the use of human needs. Similar to the present location, markers, and image recognition triggers, a basic human need may be used to activate a more personalized experience.
This research makes several contributions: (1) study augmented reality classifications and triggers; (2) analyze human needs and transfer them into sensible data elements; and (3) present the human needs trigger framework and guidelines for augmented reality.
The following sections discuss the state-of-the-art in the different types of AR, provide a discussion of the proposed framework and guidelines, and finally, demonstrate and discuss the use of this approach for creating needs-based AR experiences.

2. Background

Augmentation is described in previous literature according to human senses [14]; there is visual, auditory, and tactile augmentation. Each of these is used for a particular goal and requires different hardware. This section presents a literature study on AR.

2.1. Technologies in Augmented Reality

It is essential to understand the general process in an AR system, to realize where each of the technologies fits and how they function together. This section provides definitions for the various topics and technologies supporting the AR process [15] (Figure 1).
The components in a vision-based AR process include:
  • Image acquisition device: this device, usually a camera, captures the marker or object to be recognized.
  • Computing and storage device: the component responsible for processing the gathered input and producing the correct output.
  • Tracking and location technology: tracks the user position, view direction, and motion to decide on the virtual object to display.
  • User interaction technology: the interface that captures the user input.
  • Virtual object database: a database that includes relevant virtual objects.
  • Virtual-real fusion technology: combines the virtual objects and real environment to ensure correct occlusion, shadows, and illumination.
  • System display technology: the display technology presents the augmented information integrated with the real environment. The display technology depends on the type of AR system, visual, auditory, haptic, etc.
The general process starts by acquiring an image from the real world and sending it to computing and storage devices. This image and user interaction gestures are sent to the tracking technology, which decides what virtual information to display based on the user’s position, viewing direction, and state of motion. The virtual information is retrieved from the virtual objects database and sent to the fusion technology that ensures proper object registration in the real world. The virtual information is then displayed using the display technology [15]. Table 1 provides definitions for the core AR technologies that remain popular research topics: tracking, interaction and user interfaces, calibration and registration, display techniques, and AR applications [16].
While Table 1 describes the core AR technologies in a general manner, these technologies differ with the type of AR used. For example, the tracking techniques for vision-based AR use a camera to identify markers, while the tracking techniques for auditory AR uses acoustic tracking systems that analyze sound waves [17].

2.2. Classification of Augmented Reality

There are different types of AR systems and apps depending on the concepts and technology. The classification of these types is not unified; different research papers classify the types differently. Edwards-Stewart, Hoyt, and Reger [18] describe six types of AR under two main categories: triggered and view-based augmentation. The triggered AR technologies include marker-based AR, location-based AR, dynamic augmentation, and complex augmentation. The view-based augmentation includes Indirect and non-specific digital augmentations. Table 2 describes the classification mentioned above.
Other classifications divide AR into marker-based, marker-less, outlining AR, and superimposition AR. The marker-based type of AR depends on a marker that, when scanned by a camera, triggers an AR experience. The marker may be an image, a fiducial marker (Figure 2), or a physical object.
Markerless AR does not require markers; it uses other triggers such as location-based AR. The projection-based AR is also markerless and works by projecting digital information on objects in the user’s environment. Outlining AR uses the special abilities of cameras in recognizing objects in conditions that may be difficult to recognize for the eyes and provide guidance in the form of outlines on objects. Superimposition AR recognizes objects and superimposes virtual information on them. It includes face filters as done by Instagram and Snapchat.
AR spans beyond the visual augmentations to include the natural senses of hearing, touch, taste, and smell forming the auditory, haptic, gustatory, and olfaction AR systems respectively. This variety in AR systems constitutes another classification based on the senses.
Auditory AR is the second most common class after visual AR. It embeds auditory content into a user’s acoustic environment. In many cases, the sound is integrated with visual AR to enhance it, while in others it is used as the main augmentation to support visually impaired individuals [19].
Haptic AR is an underexplored type of AR, it “enables the user to feel a real environment augmented with synthetic haptic stimuli” [20]. Jeon and Choi [20] provide a comparison between visual and haptic AR as shown in Table 3.
On the other hand, gustatory AR is related to the sense of taste and is concerned with embedding gustatory information into human-computer interaction [21]. This type of AR research is limited because of its complexity since the sense of taste depends on different factors including vision, olfaction, and memories [21].
An olfaction AR system focuses on the sense of smell and has an olfactory display to produce odor in a gaseous or liquid state to the user [22].

2.3. Review of Exiting Platforms

This section provides a summary of prominent AR platforms. These platforms offer complete AR-experience creation capabilities. The focus is on how the different experiences are triggered. Table 4 summarizes the types of experiences and triggers of the prominent AR platforms, AWE (Augmented Web Experiences), Zap Works, BlippAR, Spark AR, Wikitude, and Unity AR.
As is clear from the summary above, marker-based AR is the most used type with image and object recognition. While most platforms enable generic experience creation, some offer face tracking, location-based, and context-aware capabilities allowing for customized experiences.

2.4. Triggers

From the previous sections, we observe the variety of AR triggers. Triggers in AR are “stimuli or characteristics that initiate or trigger the augmentation” [18]. Currently, the top starters of an AR experience are markers, images, physical objects, scene recognition, movement, location, and sometimes the choice to load an experience. In some instances, an experience may be initiated by multiple separate triggers [23].
In the future, it is expected that more advanced forms of triggers will be used in augmented reality, such as sound or voice recognition, temperature, smell, and gesture [23].

3. Theoretical Framework

This research proposes the hypothesis, “With advancing sensor technology, basic human needs are viable triggers for augmented reality experiences”. The goal is to create a novel paradigm and provide developers with more options to personalize AR experiences. This section describes the conceptual basis leading to and supporting the hypothesis.

3.1. Personalization in Augmented Reality

In many cases, AR is used to enhance the user experience in museums and tourism [24,25], education [26,27], entertainment, and medicine by adding digital information to the real environment. These approaches usually focus on the experience and the object requiring enhancement. Another way to conceive experiences is to concentrate on the user/users viewing them. This is the concept of personalization, defined as “a process that changes the functionality, interface, information access, and content, or distinctiveness of a system to increase its relevance to an individual or a category of individuals” [28]. The core elements of personalization definitions are:
  • “a purpose or goal of personalization”.
  • “What is personalized”. Four aspects of information systems may be personalized: the information (content), the presentation of information (user interface), the delivery method (channel), and the action.
  • “The target of personalization”. The target can be a group of individuals or a specific individual.
Personalization enables users to acquire information specific to their “needs, goals, knowledge, interests or other characteristics” [29]. Studying the types of AR experiences in literature, three primary levels of personalization are evident. The first level is generic experiences in which no personalization is implemented. The AR experience displays the same information to all viewers at any time, such as in museum displays. The second personalization level includes experiences receptive to external factors such as location and context [30,31]. The third level displays information about a specific user. An example of this is SentiAR (https://sentiar.com/) (accessed on 18 August 2021) an AR experience about a patient in a surgery room or an AccuVein (https://www.accuvein.com/) (accessed on 18 August 2021) device enabling the view of a patient vein in blood sampling.
This research aims to formulate a roadmap that embeds human needs in AR experiences to enhance experience personalization. In this sense, personalization elements are defined as:
  • Purpose: to trigger AR experiences based on user needs
  • What is personalized: the display of experience according to need
  • Target: the user viewing the AR experience
  • The following sections explain how these elements are applied in the proposed system.

3.2. Human Needs in Pervasive Environments

The satisfaction of human needs is a core value in pervasive environments. The study of needs in computing has been around for some time now. This topic is addressed from many viewpoints:
  • Needs representation: Some scholars attempted to represent human needs using ontologies [13,32], others used directed graphs [33].
  • Human needs are identified in several methods: interviews, questionnaires, signal processing on brain scans, and prediction methods that depend on sentiment analysis.
  • From the literature study, it is clear that technology may provide need satisfaction by different means, including providing services, social media use [34,35], internet and mobile use, online relationships, video games, and gamification.

3.3. Proposed Concept

As there are different types of AR experiences with different triggers, this research proposes to add a new type, which is a Needs-based AR. For a Needs-Based AR system, a basic human need triggers and starts the AR experience. Manfred Max-Neef’s model [36] provides a guide for the categorization of needs in his research. A detailed analysis will enable knowledge about the required sensors for the different needs and the logic necessary to transform sensor data into definite needs. Figure 3 depicts the layers of functionality in the proposed system. The first step is to collect sensor data, then applying the correct logic, this data is transformed into an identifiable need. Based on the identified need, an AR experience is selected to satisfy it or aid in its satisfaction.
The focus in the present research is on how needs trigger AR rather than how AR satisfies a need. Nevertheless, it is essential to indicate that the augmented information in a user’s environment is the possible satisfiers for the specific need that triggered the AR experience. For example, a satisfier for a hunger need may be the directions to the nearest restaurant to the user, or instructions on appropriate meal suggestions for a diabetic user. The satisfaction of needs in developing AR could be the responsibility of experience developers and marketers, and the selection of the relevant satisfier depends greatly on the user profile. Further explanation on needs and satisfiers is present in Section 3.4.

3.4. Need Analysis

Capturing human needs in a measurable form is a complex endeavor. Therefore, it requires extensive study and analysis. In this research, Max-Neef’s human scale development theory and the proposed human needs matrix [36] (Table 5) are used as a guideline to explore the various needs and dissect them into measurable elements. The human needs matrix below presents examples of satisfiers for each need category, it is not exhaustive, further explanation and guidance on satisfiers are present in the human scale development book [36]. Furthermore, the Need Context Technology (NCT) framework (Table 6) [37] is used to map relations and tools in the analysis. The aim is to study the technology that supports need detection, starting with the most basic measurable needs (usually related to health readings, hunger, stress) and evolving into the most complex needs.
The emphasis in this work is on human needs as triggers for an AR experience. Hence the goal is to provide the logic to respond to situations such as this example:
  • If a user has a subsistence need on a Being axis, trigger experience A.
This statement leads to the following questions:
  • What elements define a need?
  • How to measure a need?
Figure 4 describes the relationships between the main elements of the proposed model as follows:
  • A person has a need.
  • A need has one of three triggers: homeostasis imbalance, incentive, or stimulation.
  • A need has a satisfier, which has an AR experience.
  • A context describes the state of a person and can be an external state (environmental) or the internal state of a user.
  • An external environmental state can be an incentive that triggers a need.
  • An internal state can reflect a homeostasis imbalance that triggers a need.
A trigger in psychology is a factor that activates a need. There are three types of needs triggers [40]:
  • Homeostasis Imbalance: is the internal state that reflects a malfunction in the body processes resulting in a rise of a need. (internal)
  • Incentive: is an external positive or negative environmental stimulus that motivates a person. (external)
  • Stimulation: is an activity that causes excitement or pleasure.
Connecting these triggers (Figure 5) with the basic needs and the related technology (Table 6) forms the foundation for detecting a human need and establishing a need as a trigger for an AR experience.

3.5. Needs to Sensor Analysis

This section connects the various needs with currently available sensors. To establish the connection, first, the identification of the major categories of needs triggers: internal and external is presented (Figure 6). The internal type relates to a user’s body readings, including homeostasis, emotions, and stimulation. While the external relates to the surroundings and the environment, incentives, context, habits, and stimulation, these categories also apply to the sensor types. Specific body and wearable sensors can be used for internal triggers, such as Wearable Health Systems (WHS) or Wearable Biomedical Systems (WBS) [41]. These are termed non-invasive measuring devices that provide continuous monitoring. The external triggers relate to context and environmental sensors, including location, temperature, lighting conditions, and sound.

4. Materials and Methods

This section introduces the developed methodology to present the concept of a Needs-Based Trigger for Augmented Reality. It details an example scenario to explain the needs-based AR experience, followed by an application process and a proof of concept prototype.

4.1. Scenario Analysis

A scenario is a “hypothetical story used to help a person think through a complex problem or system” [42]. The intention is to use the scenario as an example to explain the use of the model.
For this scenario, the decision is to use the subsistence need on the being axis. Basic subsistence needs include the need for food, water, and accommodation. Therefore, the AR experience developer decides that: when a user is hungry, trigger experience A, an advertisement for the nearby restaurant.
Analyzing this scenario, it is possible to have one of these cases:
  • There is a homeostatic need for food.
  • There is a regular habit of eating at a specific time.
  • There is an external incentive that excites the user to have food.
The homeostatic need requires special sensors to detect the production of the ghrelin hormone or a low blood glucose level that indicates energy scarcity and hunger and triggers the experience. The habit may be predicted from previous user activity; therefore, at a particular time every day, the user needs to eat, and the system predicts that and presents the AR experience. Detecting context information leads to identifying possible incentives that might cause hunger, offering an AR experience in this situation, and asking for feedback can enhance the process of detection and trigger initiation.

4.2. Application Process

Based on the previous scenario, the following application process is developed (Figure 7). The process starts with selecting the sensor data, analyzing the data, followed by predictions that lead to triggering the AR experience. Afterward, collecting user feedback enhances the analysis and prediction stages. To test this application process, a prototype architecture is suggested (Figure 8). The architecture is composed of the frontend user interface and the backend experience developer interface. The developer decides the need for which he/she develops the AR experience. Then collects relevant content in the form of audio, image, 3D objects, or animation. Then, the designer designs and produces the experience. These experiences are then saved in the platform database. The user frontend initiates by creating a user profile and collecting all basic user information. The application then collects continuous sensor data and analyzes possible needs. The analysis can result in a need-based recognition or prediction. Based on the identified need, an AR experience is recommended and displayed to the user.

4.3. A Proof of Concept Prototype

Concurrent research provides an extensive explanation of the triggers, sensors, and data requirements for the needs-based AR concept. As a result, it presents suggested guidelines to develop a needs-based AR system matching the concepts in Figure 7 and Figure 8. The nine guidelines may be generalized to all needs in the fundamental human needs matrix beyond the basic needs discussed in this study. While the prior research focused on data analysis and need prediction, this study has a goal to connect all the elements and present a connected concept demonstrating the complete application process. These guidelines are seen from the developer/designer’s point of view. Following are the guidelines listed and exhibited for a prototype study.
  • Decide on the need you plan to satisfy with the AR experience.
The case study is a restaurant business that satisfies the subsistence needs of hunger and thirst. The goal is to trigger the AR experience that interests the individual to visit the restaurant at the detection of the subsistence need.
2.
Determine needs specific elements and requirements (context, personas, type)
The subsistence need is dependent on internal and external contexts. Internal context represents the homeostasis imbalance of low blood glucose or dehydration. External context includes time, day, location, activity, and settings. The personas considered for this case are healthy individuals with no food allergies, and the type is needs-based triggered AR experience.
3.
Select the trigger (homeostasis imbalance, habits, or incentives).
Developing a collective experience requires the use of all triggers; however, to simplify the process, this case selects the use of habits trigger. The habits trigger requires the collection of several records for accurate need prediction. The records must include basic profile information in addition to context information, activity, and current need.
4.
Develop the AR experience with multiple options and preferences for that particular need.
This step aims to create an AR experience that targets particular needs and matches the business goals of the experience creator. Examples of such experiences (Figure 9) are retrieved from the menu AR application [43], and the SeeFood Augmented Reality Menu [44].
In the database, the AR experience should be paired with restaurant information such as restaurant name, location, cuisine, description, and such to create a point of interest and map directions for the individual.
5.
Based on the need and trigger, select the appropriate sensors required to collect data.
Since subsistence is the selected need and habits constitute the trigger, the required sensors must capture the following data: time, day, activity, location, and proximity with others. The use of a mobile device allows for capturing the time, day, GPS location (identified by the user as home, restaurant, or other), and proximity with others (identified by the user as being with family, friend, or no one).
6.
Collect data and build a user profile and database.
A sample of the database records representing the required data fields is presented in Figure 10. These records are extracted from the Context-Aware Personalization for Augmented Reality (CAP-AR) dataset [37].
7.
Decide and develop the analysis mechanism (a rule engine, predictive algorithms).
The analysis mechanism of choice for the subsistence need based on habits are predictive algorithms. This choice is because habits are usually repetitive, and it is possible to train an algorithm to recognize them having enough records in the dataset. The analysis mechanism should detect the occurrence of a basic human need (trigger) and initiates the recommendation process, to present the individual with related AR experience.
8.
Apply a recommendation process to select the AR experience specific to the individual’s preferences.
In the case of needs-based AR applications, embodying context awareness is essential in the recommendation process. Therefore, the choice of a context-aware collaborative filtering algorithm [45] fits the criteria of such an application. This algorithm enables recommendations based on other similar users’ feedback in addition to considering the current context of the user.
9.
Design a feedback collection method to improve recommendations.
After displaying the AR experience to the user, a user rating of the experience and satisfaction level should be recorded. User feedback aids in the creation of a ranking system that supports collaborative filtering recommendations.

5. Results

Section 4.3 presents a prototype case study that demonstrates the Needs-based AR application. The principal findings of the case study are that the transfer from sensor data to meaningful information requires appropriate convergence methods since constantly requesting information from the individual may decrease pervasiveness. Since human behavior is changing, following a clear pattern on certain days and changing the behavior in others, it is essential to have sufficient records to apply machine learning algorithms and produce correct predictions. Applying the predictive algorithms to predict needs for a certain persona (User Id: P11) in the CAP-AR dataset produced the results in Table 7.

6. Discussion

The present research explores the technologies and classification of augmented reality. To provide personalized AR experiences, it connects the concept of AR with that of human needs. In this case, the human need is considered to be the trigger of the AR experience.
The needs prediction results from the CAP-AR dataset show low accuracy for the Eat/Drink subsistence need, this is due to the varying context information associate with one certain need. Basic human needs may occur at different times and contexts, correct prediction of needs requires an ample number of records in the dataset
Despite the low accuracy of the prediction results, the guidelines demonstrate a coherent means to develop a needs-based AR application that should be further tested.
While a need is a primary controller of the experience trigger, it might stir an argument: if a user has deficiencies in basic human needs, what makes the user capable of possessing the technology to operate the experience? This question might be answered in two folds:
  • Technology is now reaching many people in various living conditions.
  • People undergo changing levels of needs daily regardless of capabilities.
  • AR experiences are custom-made for specific reasons; however, these reasons are usually generic and not user-related, even in the case of personalized AR, the experiences are product, location, or service related more than user related. This research questions that from a plethora of AR experiences that shall be available in the future, how do we make AR more user-specific and find the best experience for a user at a certain time? This study answers this question by incorporating basic human needs into the equation. Therefore, the novelty is in connecting the concepts of augmented reality, needs detection, and satisfier recommendation.
The goal of this research is to create the means to trigger the experience based on needs. However, the actual design and the AR experience’s purpose are left to the creators and designers of the experience.

7. Conclusions

To realize needs-based augmented reality, it is essential to capture the different needs using computing methods. In this paper, the focus is on the automatic detection of basic needs. A theoretical framework is proposed that combines augmented reality with human needs to realize the process of needs-based AR.
First, a summary of technologies and classifications of AR is provided. A review of existing platforms is also provided.
Second, a discussion on a theoretical framework leading to the foundation of the proposed concept. This is followed by a need and sensor analysis.
Third, a scenario is presented to demonstrate the proposed concept, followed by the application process based on the scenario and a proof of concept prototype.
To use human needs as triggers in augmented reality applications is significant and necessary in various domains. For pervasive computing, the combination expands the variety of research and enables the personalization of experiences. This research is also practical in marketing, entertainment, and ambient assisted living. The proposed framework established based on previous research, focuses on needs analysis, while the guidelines provide a roadmap to create a needs-based AR application. Future work will concentrate on system implementation with a vision of achieving needs-based AR.

Author Contributions

Conceptualization, M.A.Y., and A.D.; methodology, M.A.Y.; formal analysis, M.A.Y.; data curation, M.A.Y.; writing—original draft preparation, M.A.Y.; writing—review and editing, A.D.; visualization, M.A.Y.; supervision, A.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request due to privacy restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ronald, T.; Azuma, A. Survey of augmented reality. Presence Teleoper. Virtual Environ. 1997, 6, 355–385. [Google Scholar] [CrossRef]
  2. Carmigniani, J.; Furht, B.; Anisetti, M.; Ceravolo, P.; Damiani, E.; Ivkovic, M. Augmented reality technologies, systems and applications. Multimed. Tools Appl. 2011, 51, 341–377. [Google Scholar] [CrossRef]
  3. Azuma, R.; Baillot, Y.; Behringer, R.; Feiner, S.; Julier, S.; MacIntyre, B. Recent advances in augmented reality. IEEE Comput. Graph. Appl. 2001, 21, 34–47. [Google Scholar] [CrossRef] [Green Version]
  4. Dunleavy, M.; Dede, C. Augmented reality teaching and learning. In Handbook of Research on Educational Communications and Technology; Springer: Cham, Switzerland, 2014; pp. 735–745. [Google Scholar]
  5. Patricio, J.M.; Costa, M.C.; Carranca, J.A.; Farropo, B. SolarSystemGO—An augmented reality based game with astronomical concepts. In Proceedings of the 2018 13th Iberian Conference on Information Systems and Technologies (CISTI), Caceres, Spain, 13–16 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–3. [Google Scholar]
  6. Waltner, G.; Schwarz, M.; Ladstätter, S.; Weber, A.; Luley, P.; Bischof, H.; Lindschinger, M.; Schmid, I.; Paletta, L. Mango-mobile augmented reality with functional eating guidance and food awareness. In Lecture Notes in Computer Science, Proceedings of the International Conference on Image Analysis and Processing, Genoa, Italy, 7–8 September 2015; Springer: Cham, Switzerland, 2015; pp. 425–432. [Google Scholar]
  7. Khor, W.S.; Baker, B.; Amin, K.; Chan, A.; Patel, K.; Wong, J. Augmented and virtual reality in surgery—the digital surgical environment: Applications, limitations and legal pitfalls. Ann. Translat. Med. 2016, 4, 454. [Google Scholar] [CrossRef] [Green Version]
  8. Frigo, M.A.; da Silva, E.C.; Barbosa, G.F. Augmented reality in aerospace manufacturing: A review. J. Ind. Intell. Inform. 2016, 4, 125–130. [Google Scholar] [CrossRef]
  9. Hammady, R.; Ma, M.; Temple, N. Augmented reality and gamification in heritage museums. In Lecture Notes in Computer Science, Proceedings of the Joint International Conference on Serious Games, Brisbane, QLD, Australia, 26–27 September 2016; Springer: Cham, Switzerland, 2016; pp. 181–187. [Google Scholar]
  10. Boy, G.A.; Platt, D. A situation awareness assistant for human deep space exploration. In Lecture Notes in Computer Science, Proceedings of the International Conference on Human-Computer Interaction, Las Vegas, NV, USA, 21–26 July 2013; Springer: Cham, Switzerland, 2013; pp. 629–636. [Google Scholar]
  11. Han, D.; tom Dieck, M.C.; Jung, T. User experience model for augmented reality applications in urban heritage tourism. J. Herit. Tour. 2018, 13, 46–61. [Google Scholar] [CrossRef]
  12. Zhou, F.; Duh, H.B.; Billinghurst, M. Trends in augmented reality tracking, interaction and display: A review of ten years of ISMAR. In Proceedings of the 2008 7th IEEE/ACM International Symposium on Mixed and Augmented Reality, Cambridge, UK, 15–18 September 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 193–202. [Google Scholar]
  13. Human, S.; Fahrenbach, F.; Kragulj, F.; Savenkov, V. Ontology for Representing Human Needs, Knowledge Engineering and Semantic Web; Różewski, P., Lange, C., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 195–210. [Google Scholar]
  14. Livingston, M.A. Evaluating human factors in augmented reality systems. IEEE Comput. Graph. Appl. 2005, 25, 6–9. [Google Scholar] [CrossRef]
  15. Juan, C.; YuLin, W.; Wei, S. Construction of interactive teaching system for course of mechanical drawing based on mobile augmented reality technology. Int. J. Emerg. Technol. Learn. 2018, 13, 126. [Google Scholar]
  16. Kim, K.; Billinghurst, M.; Bruder, G.; Duh, H.B.; Welch, G.F. Revisiting trends in augmented reality research: A review of the 2nd decade of ISMAR (2008–2017). IEEE Trans. Vis. Comput. Graph. 2018, 24, 2947–2962. [Google Scholar] [CrossRef]
  17. Gamper, H. Enabling Technologies for Audio Augmented Reality Systems. Ph.D. Thesis, Aalto University, Espoo, Finland, 2014. [Google Scholar]
  18. Edwards-Stewart, A.; Hoyt, T.; Reger, G. Classifying different types of augmented reality technology. Ann. Rev. CyberTher. Telemed. 2016, 14, 199–202. [Google Scholar]
  19. Ribeiro, F.; Florêncio, D.; Chou, P.A.; Zhang, Z. Auditory augmented reality: Object sonification for the visually impaired. In Proceedings of the IEEE 14th International Workshop on Multimedia Signal Processing (MMSP), Banff, AB, Canada, 17–19 September 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 319–324. [Google Scholar]
  20. Jeon, S.; Choi, S. Haptic augmented reality: Taxonomy and an example of stiffness modulation. Presence Teleoper. Virtual Environ. 2009, 18, 387–408. [Google Scholar] [CrossRef]
  21. Narumi, T.; Nishizaka, S.; Kajinami, T.; Tanikawa, T.; Hirose, M. Augmented Reality Flavors: Gustatory Display Based on Edible Marker and Cross-Modal Interaction. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Vancouver, BC, Canada, 7–12 May 2011; Association for Computing Machinery: New York, NY, USA, 2011; pp. 93–102. [Google Scholar]
  22. Yamada, T.; Yokoyama, S.; Tanikawa, T.; Hirota, K.; Hirose, M. Wearable Olfactory Display: Using Odor in Outdoor Environment. In Proceedings of the IEEE Virtual Reality Conference (VR 2006), Alexandria, VA, USA, 25–29 March 2006; pp. 199–206. [Google Scholar]
  23. Bower, M.; Howe, C.; McCredie, N.; Robinson, A.; Grover, D. Augmented Reality in education—Cases, places and potentials. Educ. Media Int. 2014, 51, 1–15. [Google Scholar] [CrossRef]
  24. He, Z.; Wu, L.; Li, X.R. When art meets tech: The role of augmented reality in enhancing museum experiences and purchase intentions. Tour. Manag. 2018, 68, 127–139. [Google Scholar] [CrossRef]
  25. Serravalle, F.; Ferraris, A.; Vrontis, D.; Thrassou, A.; Christofi, M. Augmented reality in the tourism industry: A multi-stakeholder analysis of museums. Tour. Manag. Perspect. 2019, 32, 100549. [Google Scholar] [CrossRef]
  26. Billinghurst, M. Augmented reality in education. New Horiz. Learn. 2002, 12, 1–5. [Google Scholar]
  27. Chen, P.; Liu, X.; Cheng, W.; Huang, R. A review of using Augmented Reality in Education from 2011 to 2016. In Innovations in Smart Learning; Springer: Cham, Switzerland, 2017; pp. 13–18. [Google Scholar]
  28. Fan, H.; Poole, M.S. What is personalization? Perspectives on the design and implementation of personalization in information systems. J. Organ. Comput. Electr. Commer. 2006, 16, 179–202. [Google Scholar]
  29. Zimmermann, A.; Specht, M.; Lorenz, A. Personalization and context management. User Model. User Adapt. Interact. 2005, 15, 275–302. [Google Scholar] [CrossRef]
  30. Paavilainen, J.; Korhonen, H.; Alha, K.; Stenros, J.; Koskinen, E.; Mayra, F. The Pokémon GO experience: A location-based augmented reality mobile game goes mainstream. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems, Denver, CO, USA, 6–11 May 2017; pp. 2493–2498. [Google Scholar]
  31. Tahara, T.; Seno, T.; Narita, G.; Ishikawa, T. Retargetable AR: Context-aware Augmented Reality in Indoor Scenes based on 3D Scene Graph. In Proceedings of the 2020 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Recife, Brazil, 9–13 November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 249–255. [Google Scholar]
  32. Ghadiri, N.; Nematbakhsh, M.A.; Baraani-Dastjerdi, A.; Ghasem-Aghaee, N. A Context-Aware Service Discovery Framework Based on Human Needs Model, Service-Oriented Computing–ICSOC 20-07; Krämer, B.J., Lin, K., Narasimhan, P., Eds.; Springer: Berlin/Heidelberg, Germany, 2007; pp. 404–409. [Google Scholar]
  33. Abramovich, A. Theory of needs and problems. Socio Econ. Sci. 2020, 4, 411. [Google Scholar]
  34. Zhu, Y.; Chen, H. Social media and human need satisfaction: Implications for social media marketing. Bus. Horiz. 2015, 58, 335–345. [Google Scholar] [CrossRef]
  35. Houghton, D.; Pressey, A.; Istanbulluoglu, D. Who needs social networking? An empirical enquiry into the capability of Facebook to meet human needs and satisfaction with life. Comput. Hum. Behav. 2020, 104, 106153. [Google Scholar] [CrossRef]
  36. Max-Neef, M.A. Human Scale Development; Apex Press: New York, NY, USA, 1991. [Google Scholar]
  37. Yahya, M.; Dahanayake, A. A Needs-based personalization model for context aware applications. Front. Artif. Intell. Appl. 2016, 292, 63–82. [Google Scholar]
  38. Max-Neef, M.; Elizalde, A.; Hopenhayn, M. Development and human needs. In Real-Life Economics: Understanding Wealth Creation; Routledge: London, UK, 1992; pp. 197–213. [Google Scholar]
  39. Yahya, M.A.; Dahanayake, A. Augmented reality for human needs: An ontology. Front. Artif. Intell. Appl. 2020, 333, 275–294. [Google Scholar]
  40. Myers, D.G. Psychology, 10th ed.; Worth Publishers: New York, NY, USA, 2013. [Google Scholar]
  41. Andreoni, G.; Standoli, C.E.; Perego, P. Defining requirements and related methods for designing sensorized garments. Sensors 2016, 16, 769. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Cem Kaner, J.D. An Introduction to Scenario Testing; Florida Institute of Technology: Melbourne, Australia, 2013; pp. 1–13. [Google Scholar]
  43. Menu, A.R. Look at Food in a New Way. Available online: http://menuar.ru/en.php (accessed on 18 July 2021).
  44. See Food Augmented Reality Menu. Available online: https://sketchfab.com/seefood (accessed on 18 July 2021).
  45. Chen, A. Context-Aware Collaborative Filtering System: Predicting the User’s Preference in the Ubiquitous Computing Environment, Location-and Context-Awareness; Strang, T., Linnhoff-Popien, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; pp. 244–253. [Google Scholar]
Figure 1. General process in an augmented reality system.
Figure 1. General process in an augmented reality system.
Applsci 11 07978 g001
Figure 2. Fiducial marker.
Figure 2. Fiducial marker.
Applsci 11 07978 g002
Figure 3. Layers of functionality in the proposed application.
Figure 3. Layers of functionality in the proposed application.
Applsci 11 07978 g003
Figure 4. Conceptual model of augmented reality for human needs [39].
Figure 4. Conceptual model of augmented reality for human needs [39].
Applsci 11 07978 g004
Figure 5. Components for human needs detection.
Figure 5. Components for human needs detection.
Applsci 11 07978 g005
Figure 6. External and Internal Needs.
Figure 6. External and Internal Needs.
Applsci 11 07978 g006
Figure 7. Application Process.
Figure 7. Application Process.
Applsci 11 07978 g007
Figure 8. Testing prototype architecture.
Figure 8. Testing prototype architecture.
Applsci 11 07978 g008
Figure 9. 3D models for AR subsistence need hunger experiences. Source: (a,b,e) [43], (c,d,f) [44].
Figure 9. 3D models for AR subsistence need hunger experiences. Source: (a,b,e) [43], (c,d,f) [44].
Applsci 11 07978 g009aApplsci 11 07978 g009b
Figure 10. Records of the dataset portraying required fields.
Figure 10. Records of the dataset portraying required fields.
Applsci 11 07978 g010
Table 1. Core augmented reality technologies and techniques [16].
Table 1. Core augmented reality technologies and techniques [16].
Core AR TechnologyDefinition
Tracking techniques“Methods of tracking a target object/environment via cameras and sensors, and estimating viewpoint poses”.
Interaction techniques and user interfaces“Techniques and interfaces for interacting with virtual content”.
Calibration and registration“Geometric or photometric calibration methods, and method to align multiple coordinate frames”.
Display techniques“Display hardware to present virtual content in AR, including head-worn, handheld, and projected displays”.
AR applications“AR systems in application domains such as medicine, manufacturing, or military, among others”.
Table 2. Classification of augmented reality types [18].
Table 2. Classification of augmented reality types [18].
CategoryTypeBriefExample
TriggeredMarker-BasedTriggered by a marker:
Paper (image)
Physical object
Museum displays
Location-BasedTriggered by GPS locationMonocle Restaurant information
Dynamic AugmentationResponsive to object changesDigitally trying clothes and accessories with shopping apps
Complex AugmentationA combination of the aboveDynamic view with digital information from the Internet
View-BasedIndirect AugmentationIntelligent augmentation of a static viewTaking a picture and changing the wall color
Non-Specific digital AugmentationAugmentation of a dynamic viewAugmentation in
mobile games
Table 3. Computational procedures of visual and haptic AR [20].
Table 3. Computational procedures of visual and haptic AR [20].
ProcedureVisual ARHaptic AR
Sensing a real
environment
Captures real information needed for visual augmentationSenses real information needed for haptic augmentation
Sensor examplesCamera, range finder, trackerPosition encoder, accelerometer, force sensor, thermometer
Constructing stimuli for augmentation
  • Real-virtual registration
  • Overlay of the virtual object on the real scene
  • Contact detection between the tool and a real object
  • Modulation of real haptic stimuli
Displaying
augmented stimuli
Uses a visual displayUses a haptic display
Display examplesHead-mounted display, projector, mobile phoneForce-feedback interface, tactile display, thermal display
Table 4. Comparison of AR platforms with types and triggers.
Table 4. Comparison of AR platforms with types and triggers.
PlatformPublish Application
or Web-Based
SDKTypes of
Experiences
Experience Trigger, AR-type
AWE Media Studio
https://awe.media/
(accessed on 18 August 2021).
WebNoImage, spatial, face tracking, GPS location, 360°Weblink,
Non-specific digital augmentation
Zap Works
https://zap.works/
(accessed on 18 August 2021).
EitherYesImage, face tracking, 360°Marker-based (special marker)
BlippAR Builder
https://www.blippar.com/build-ar
(accessed on 18 August 2021).
EitherYesImageMarker-based (image scan)
Spark AR Studio
https://sparkar.facebook.com/ar-studio/
(accessed on 18 August 2021).
On Facebook or InstagramNoFace tracking, image-basedMaker-based, dynamic augmentation
Wikitude AR
https://www.wikitude.com/
(accessed on 18 August 2021).
ApplicationYesImage, object, scene recognition, instant tracking, Geo ARMarker-based, location-based, dynamic augmentation
Unity MARS
https://unity.com/products/unity-mars
(accessed on 18 August 2021).
ApplicationNoLocation-aware, context-awareMarker-based, complex augmentation
Table 5. Human needs matrix [38].
Table 5. Human needs matrix [38].
Needs According to Axiological CategoriesBeingHavingDoingInteracting
Subsistence1/Physical health, mental health,2/Food, shelter, work3/Feed, procreate, rest, work4/Living environment, social setting
Protection5/Care, adaptability, autonomy6/insurance systems, savings, work,7/prevent, plan, take care of, cure, help8/Living space, social environment,
Affection9/Self-esteem, solidarity, respect10/Friendship, family, partnerships, a sense of humor11/Caress, express emotions12/Privacy, intimacy, home
Understanding13/Critical conscience, curiosity, receptiveness14/Literature, teachers, method, educational policies15/Investigate, study, experiment, educate, analyze, meditate16/Settings of formative interaction, schools, universities
Participation17/Adaptability, receptiveness, solidarity18/Rights, responsibilities, duties, privilege, work, a sense of humor19/Become affiliated, cooperate, propose20/Settings of participative interaction, parties, associations
Leisure21/Curiosity, receptiveness, imagination22/Games, spectacles, clubs, parties, peace of mind, a sense of humor23/Day-dream, brood, dream24/Privacy, intimacy, spaces of closeness
Creation25/Passion, determination, intuition26/Abilities, skills, method, work27/Work, invent, build, design, compose, interpret28/Productive and feedback settings, workshops
Identity29/Sense of belonging, consistency30/Symbols, language, religions, habits, customs,31/Commit oneself, integrate oneself, confront32/Social rhythms, everyday settings,
Freedom33/Autonomy, self-esteem, determination34/Equal rights35/Dissent, choose, be different from36/Temporal/Spatial plasticity
Table 6. Need–context–technology framework [37].
Table 6. Need–context–technology framework [37].
Fundamental Human Needs Existential CategoriesBeing
(Qualities)
Having
(Things)
Doing
(Actions)
Interacting
(Settings)
Context-Aware
Categorization
User, Who
(Identity)
ThingsWhat (Activity)Where (Location), Weather, Social, Networking
When (Time)
Sensors and TechnologyEmotion Sensors
Body Sensors
IoT systems and SensorsActivity recognition through motion sensorsLocation awareness, nearby user device (for proximity with other users)
Table 7. Need prediction results.
Table 7. Need prediction results.
AlgorithmAccuracy of all Needs PredictionEat/Drink Need Class PrecisionEat/Drink Need Class Recall
Decision Tree49.32%44.44%42.11%
Random Forest32.88%40.00%21.05%
Naive Bayes38.36%42.11%42.11%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yahya, M.A.; Dahanayake, A. A Needs-Based Augmented Reality System. Appl. Sci. 2021, 11, 7978. https://doi.org/10.3390/app11177978

AMA Style

Yahya MA, Dahanayake A. A Needs-Based Augmented Reality System. Applied Sciences. 2021; 11(17):7978. https://doi.org/10.3390/app11177978

Chicago/Turabian Style

Yahya, Manal A., and Ajantha Dahanayake. 2021. "A Needs-Based Augmented Reality System" Applied Sciences 11, no. 17: 7978. https://doi.org/10.3390/app11177978

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop