Next Article in Journal
Fractal Aspects in Classical Parallel Computing
Next Article in Special Issue
Ambient Technology to Assist Elderly People in Indoor Risks
Previous Article in Journal
Color Reduction in an Authenticate Live 3D Point Cloud Video Streaming System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A University-Based Smart and Context Aware Solution for People with Disabilities (USCAS-PWD)

1
Riyadh Techno Valley, King Saud University, P.O. Box 3966, Riyadh 12373-8383, Saudi Arabia
2
FARCAMT CHAIR, Advanced Manufacturing Institute, King Saud University, Riyadh 11421, Saudi Arabia
3
College of Computer and Information Sciences, King Saud University, Riyadh 11421, Saudi Arabia
4
Department of Electrical Engineering, University of Dubai, Dubai 14143, United Arab Emirates
*
Author to whom correspondence should be addressed.
Computers 2016, 5(3), 18; https://doi.org/10.3390/computers5030018
Submission received: 23 May 2016 / Revised: 10 July 2016 / Accepted: 22 July 2016 / Published: 29 July 2016
(This article belongs to the Special Issue Theory, Design and Prototyping of Wearable Electronics and Computing)

Abstract

:
(1) Background: A disabled student or employee in a certain university faces a large number of obstacles in achieving his/her ordinary duties. An interactive smart search and communication application can support the people at the university campus and Science Park in a number of ways. Primarily, it can strengthen their professional network and establish a responsive eco-system. Therefore, the objective of this research work is to design and implement a unified flexible and adaptable interface. This interface supports an intensive search and communication tool across the university. It would benefit everybody on campus, especially the People with Disabilities (PWDs). (2) Methods: In this project, three main contributions are presented: (A) Assistive Technology (AT) software design and implementation (based on user- and technology-centered design); (B) A wireless sensor network employed to track and determine user’s location; and (C) A novel event behavior algorithm and movement direction algorithm used to monitor and predict users’ behavior and intervene with them and their caregivers when required. (3) Results: This work has developed a comprehensive and universal application with a unified, flexible, and adaptable interface to support the different conditions of PWDs. It has employed an interactive smart based-location service for establishing a smart university Geographic Information System (GIS) solution. This GIS solution has been based on tracking location service, mobility, and wireless sensor network technologies. (4) Conclusion: The proposed system empowered inter-disciplinary interaction between management, staff, researchers, and students, including the PWDs. Identifying the needs of the PWDs has led to the determination of the relevant requirements for designing and implementing a unified flexible and adaptable interface suitable for PWDs on the university campus.

1. Introduction

Presently, over a billion people, including children (or about 15% of the world’s population) have been estimated to be living with disabilities [1]. The absence of or the presence of ineffective support services or tools can make people with disabilities (PWDs) overly dependent on their families, friends, and caregivers. As a result, they are prohibited from being economically active and socially included. In a university, to make PWDs economically independent and to provide them with a sense of inclusion, a PWD-friendly campus/environment is needed. It is very important for the designers to take into account all the possible users, their interactions among themselves as well as with the environment before developing any products or services. The user involvement at each and every phase of the development process helps the designers to accomplish the optimal requirements for the PWDs [2]. A proper solution for disabled people on the university campus can be defined as the one which is universal, adaptable, heterogeneous, and user-centered. In fact, it should be able to make PWDs life easy and productive instead of being idle and frustrated. Moreover, increased information sharing within the university campus has a positive impact on the evolution of the knowledge as well as the innovative capabilities of the academic environment. It would also give rise to opportunities and prospects for new collaborative and enriched basic and applied research.
The advances in mobile and wireless technology, and the new concept of location-based services (LBS) together with Geographic Information Systems (GIS), have led to the introduction of the smart adaptive professional network that is required for the establishment of a successful eco-system. The accelerated rapid development of the wireless network and mobile computing technologies that are the driving forces behind the mobile information services for PWDs has become an appealing research topic. In particular, the LBSs on mobile devices can convey location-based information to individual users, thus helping the users (including the PWDs) to acquire helpful and adaptive information [3]. Furthermore, the interactive LBSs enhance personalized and context-aware information communication, and maximize the gain from such dynamic information sharing by empowering the user with extended functionalities.
In this project, a unified interface has been designed to support the PWDs during their movement within the campus. This new approach uses the concept of GIS within the platform of wireless technologies and develops an interactive smart context-aware y to enhance information sharing, communication, and dissemination within the King Saud University campus. The interface has been designed to suit everybody on campus, including PWDs. The size of font, font color, window color, as well as the voice recognition used for controlling the users’ interface have been considered in the design and development according to the needs of the PWDs. Consideration of all of the above aspects led to the development of an information-intensive eco-system to support knowledge, creativity, and innovation within the multi-disciplinary environment.
The paper has been divided in six sections as follows. A literature review pertaining to GIS and sensors technologies has been conducted to determine the right solution for this project (as shown in Geographic Information System and Sensors Technologies). The interface has been designed in accordance with the specified criteria applicable to the impairment conditions of the majority of PWDs (as described in Design Consideration and Methodology). A review of the existing commercialized solutions has been performed to verify their compliance with the specified criteria, as well as to establish the direction of improvement, in order to maximize the PWD users’ satisfaction (as shown in the Interface Design consideration to Support PWD Needs). The complete design of the University-Based Smart and Context Aware Solution for People with Disabilities (USCAS-PWD) can be realized in the Group Design and Implementation Consideration. A description of the group selection for the PWDs to support the different impairment conditions has been presented in the implementation of the USCAS-PWD. An implementation of the Unified Interface has been described for the indoor (inside the building) and outdoor at the university campus within the section Implementation of the Unified Interface for Indoor Network. The event behavior algorithm and the movement direction algorithm are discussed within the algorithm section.

2. Geographic Information System and Sensors Technologies

With the continuous increase in online spatial data, the adaptation of spatial content according to the user’s context has become crucial. In addition, the introduction of GIS [4] has changed the way geographical and spatial information is manipulated, accessed, and understood. For example, Li [5] developed a conceptual model based on LBS to assist the urban pedestrian. The results of his study revealed the importance of user preferences on the information requirements. Similarly, Zipf [6] and Yuxia [7] presented several approaches to realize adaptive mobile GIS services in the domain of pedestrian navigation and tourist information. The Intelligent Map Agents (IMA) architecture developed by Gervais [8] for GIS was aimed at superseding the monolithic approach through a new dynamic, lean, and customizable system supporting spatially-oriented applications. Relevant to the concept of the adaptive GIS, personalization, and context-awareness, Aoidh [9] introduced an approach that implicitly monitored the user’s activity and generated a user profile reflecting her/his information preferences based on the interactions of the user with the system, and her/his physical location and movements. Brossard et al. [10] have also proposed a solution allowing the context to be managed inside an application’s conceptual model, in order to provide more flexible web applications from the user’s point of view.
Recently, considerable interest has been shown in interactive LBS, specifically in the fields of mobile communications and wireless networking. For instance, Jung [11] proposed an interactive approach by developing social networks between mobile users through collection of a datasets from about two million users. Chang et al. [12], developed an adaptive context-aware recommendation system. The system was built using a slow intelligence approach. The system was implemented as an application on the smart phone. A case study in Pittsburg, USA was conducted, and the results obtained by the experimental campaign were satisfactory and showed good perspective of this kind of approach. Similarly, Rebollo-Monedero [13] presented a novel protocol based on user collaboration to privately retrieve location-based information from an LBS provider. In their approach, the user queries, comprising accurate locations, remained unchanged, and the collaborative protocol did not impose any special requirements on the query–response function of the LBS. The application devised by Kbar [14] enabled the user to conduct a search using a set of three modes: keyword, location, and targeted profile. The GIS was utilized within a platform of wireless technologies to develop an interactive smart context-aware GIS for a University. Until now, many technologies have been developed based on Global Positioning System (GPS) receivers and General Packet Radio Service GPRS transmitters, where the positions are stored and accessed via MySQL or web servers through internet connectivity [15,16,17].
Different technologies, including WiFi tag through Access Point (AP) [18], sensor technology based on wireless personal area network (WPAN) IEE 802.15.4 [19], passive and active radio frequency identification (RFID) sensors, GPS technologies [20], etc., can be employed in tracking user location. For example, Ning et al. [21] introduced a tree-like code structure to develop a unified modeling scheme of the physical objects (in the RFID network) without an identification code. A hybrid platform which is based on the combination of RFID, WiFi, and GPS technologies showed better accuracy in comparison to the individual technologies [22]. In the system developed by Dao et al. [22], the user (with a smartphone) was detected by the RFID readers and the WiFi APs. Similarly, Li et al. [23] developed a system based on radio-over-fiber (RoF) network involving the simultaneous transmission of RFID, Wi-Fi, and ZigBee services. They integrated all the wireless signals in the electrical domain, and then modulated them onto the optical carrier through an optical modulator. The system was tested over a 2 km RoF link where RFID, WiFi, and ZigBee master nodes could communicate effectively. Meanwhile, S. Szewcyzk [24] analyzed the individual impairment behavior using machine learning techniques. They used inhabitant feedback to decrease the annotation time and improve their performance. In fact, the Wireless Body/Personal Area Network (WBPAN) developed by G. Yella Reddy [25] included low-cost, light, and small sensors for continuous health monitoring and sending instantaneous feedback to the user.
Assistive technologies (ATs) can also be used to track and monitor the activity and vital signs of elder and motor-impaired individuals through different types of wearable sensors, implantable sensors, and microsystems that can be swallowed, such as microcapsule devices [26]. Five basic features, including the automation, multi-functionality, adaptability, interactivity, and efficiency play a significant role in the design of smart homes [27]. In order to help blind people to walk on the streets, Chen and Zhi [28] proposed a solution based on sensors that detected the blind individual’s movement through an RFID tag installed on a cane. The walking path of the blind man was designed to incorporate the electronic tags underneath it. An indoor localization aid, called portable position and orientation estimation (POSE) was developed by Hesch and Roumeliotis [29] for visually-impaired persons, in order to increase their safety and independence. The 3D orientation of the cane was tracked in the first layer through measurements using a three-axis gyroscope and laser scanner. The 2D position of the person was detected in the second layer using corner features extracted from the laser-scan data, linear velocity measurements from the pedometer, and a filtered version of the cane’s yaw. Kammoun et al. [30] utilized micro-navigation (sensing immediate environments) and macro-navigation (reaching remote destinations) functions to design a system for blind people in indoor and outdoor environments. This system started searching for any object requested by the user in the captured images. After the object was detected, the system would direct the user to move their hand for grasping via voice command.
The multi-sensory (audio and force feedback) learning environment developed by Darrah [31] was used to teach visually-impaired or blind children. It consisted of a PC, a low-cost force feedback stylus-based haptic device, auditory cues, and high contrast graphics. Similarly, the electronic navigation system [32] could detect obstacles around the subject (visually-impaired and blind people) up to 500 cm in the front, left, and right directions through a network of ultrasonic sensors. Gallaghera et al. [33] also designed and tested an indoor navigation system for visually-impaired persons. This system was based on the internal hardware of the smartphone (such as the accelerometer) to measure the user’s position and direction. A navigation and way-finding system [34] for visually-impaired persons used RFID technology to estimate the position of the user. The system was designed for indoor use, and tags were installed under the floor. The tag reader was attached to a shoe or a stick that sent a query to tags. The orientation guiding system built by Ghiani et al. [35] for blind persons actually used vibrotactile feedback. The prototype used the RFID network and tags to determine the person’s surroundings, and was installed inside the museum to guide blind persons. Similarly, Colace et al. [36] developed a context-aware application based on a slow intelligence approach for cultural heritage application. The developed system was implemented as an application for the Android System. A case study with 50 users was conducted in which the users were made to visit in some rooms of the Capodimonte museum (Salerno, Italy ). The results showed that the users were satisfied with the performance of the application. Another RFID-based system [37] helped blind users to identify objects in the home/classroom. The system had four major components: RFID Reader, FM Transmitter, Database Server, and RFID Tags. The system developed by El-Alamy et al. [38] allowed visually-impaired persons to catch buses safely. The framework incorporated an auditory device, a tactile interface, and a wireless communication system. It also included a sensor, a radio frequency reader (RF) that was employed in the transport station. A wearable device was designed by Jain [39] to aid blind persons in indoor navigation. The system had two major constituents: wall modules were installed in the building, and the user end included a waist-worn gadget linked to a smartphone. All instructions were communicated to the user via the Text-to-Speech (TTS) engine of the smartphone application. The “Drishti” project [40] was aimed at designing a standalone-integrated system to help blind and low-vision individuals during outdoor navigation. It consisted of a wearable computer, GPS receiver, headset, and access to a wireless network for position information and way-guidance in the form of text-to-speech (TTS) voice commands. Gomez et al. [41] also developed a system to provide visually-impaired persons with a more practical and functional device based on the sensory substitution methods (SSDs). The device was called the See ColOr, and consisted of SSD, a 3D camera, bone-phones, an iPad for tactile feedback, and a 14-inch laptop. The recognition module based on computer vision allowed the users to reach targets and avoid obstacles. Similarly, Iannizzotto et al. [42] developed a system to help blind people to discover objects in the indoor environment around them. The system read the barcode attached to all interested objects. This system used an ultrasonic sensor to detect obstacles (i.e., objects with no tags). The ultrasonic system was attached to the user’s belt. This system used a head-mounted camera to read barcode tags, and a headphone to get the voice feedback of the query about any interested object passed to the system. Ando et al. [43] developed a device based on an IR multisensory array that could help blind people in environment orientation and movement. It was based on smart signal processing to provide the user with suitable information about the position of obstacles found in his/her path.
The emergence and evolution of the concept and technologies of Ambient Intelligent (AmI) and ubiquitous computing have paved the way towards building smart environments. These smart environments exploit the key enabling technologies, which are necessary to improve the quality of life and performance and provide assistance in different kinds of environments. AmI has specific features, attributes, and aspects, including context-awareness. A context-aware system provides a predictive behavior based on knowledge of the environment. A few researchers, including Cooper and Cooper [44], have addressed the issue of quality of life for people who suffered spinal cord injuries. They stated that technology plays a critical role in promoting the well-being, activity, and participation of individuals with spinal cord injury (SCI). School, work, travel, and leisure activities can all be facilitated by the technology. They pointed out that software has made computer interfaces adaptive and intelligent, through learning of the user’s behavior and optimizing its structure. Another important concept that can be employed in an assistive workplace for PWDs is the “Internet of Things (IoT)” [45]. The IoT can offer PWDs the assistance and support, they need to achieve a good quality of life and allows them to participate in social and economic life. The IoT depicts a world of networked smart devices, where everything is interconnected and has a digital entity [46]. According to Giannetsos et al. [47], user-based technologies can never succeed without appropriate provisions addressing security and privacy. As a result of the ubiquitous nature of data emerging from the sensors carried by people, the highly dynamic and mobile setting presents new challenges for information security, data privacy, and ethics. To overcome security issues in IoT m-Health devices, Doukas et al. [48] employed a system based on digital certificates and Public Key Infrastructure (PKI) data encryption. An adaptive approach, as proposed by Yau et al. [49], can also be used to prevent communication systems from various security attacks. Similarly, Templeton [50] addressed the issues of the security challenges in the increased use of cyber-physical devices in assistive environments. In fact, there have been many instances where the AT system was not properly secure, which means they were susceptible to fraud and harm as a result of unexpected interaction with other systems.
Most of the research works presented in this section have focused on home or outdoor ATs and neglected the social inclusion of PWDs in universities. In addition, most of the developed smart solutions and ATs were not comprehensive, as they addressed only one or two impairments. Furthermore, few of these papers attempted to analyze the behavior of PWD and provide relevant intervention methods to guide them at home or in the workplace.
Furthermore, it is clear from this section that the use of technologies based on GIS and LBS in mobile applications can help PWDs and other students on the university campus. This would also help in monitoring user activities and help PWDs in accessing and sharing of information flexibly. However, there have been issues, such as network security, interactive application and social networks and the use of IoT, etc., that require further attention. Moreover, a smart context-aware based design would help in analyzing PWDs users’ needs and to perform their activities efficiently. It also goes beyond meeting the users’ requirements by designing and implementing a smart AT solution that is adaptable to different conditions according to the specific context, which helps to attain better user satisfaction and adoption. A USCAS-PWD would assist PWDs in accessing relevant information anywhere on the campus freely, flexibly, securely, and at all times. It would also help them to communicate with each other and their caregivers efficiently and effectively. The design of such system on the university campus and within the buildings would require the identification of the user (PWDs) needs through a proper framework and design methodology. This methodology should be able to assess their needs and match them to relevant technologies that can satisfy them, as described in next section. Moreover, in order to support behavior analysis and intervention, there is a need of a wireless network that can be used to track user movements within the building. Real time tracking location system (RTLS) can be useful for behavior analysis of PWDs, as it helps to identify their location within the home, building, and work environments. The RFID and sensor networks play an essential role in tracking assets within a defined environment, where it transmits the identity associated with these assets wirelessly using radio waves [51].

3. Design Methodology and Consideration

The following sub-sections describe the interface design considerations to support PWD needs, assess existing software applications for the PWDs to identify improvements, identify the design features of USCAS-PWD that covers indoor and outdoor applications, and evaluate the proposed solution through a comparison between the existing software applications and the proposed design.

3.1. AT Design Methedology to Support PWD Needs

A user-driven process that focused on technologies as well as quality of life has been proposed by Federici and Scherer [52]. This solution, which relied on AT and addressed the needs of the PWDs was achieved through the Assistive Technology Assessment (ATA) process. In the ATA process, the improvement in individual functioning can be realized in three phases: phase 1—users seeking a solution, where a professional team would interact with these users to get their requirements and then match their needs to a suitable solution; phase 2—users validate the solution provided by the professional team; phase 3—users adopt the solution and get appropriate training. Three important considerations, including accessibility, universality, and sustainability have been proposed by the authors. The criterion “accessibility” is actually used to measure the level of environmental characteristics [53,54]. Similarly, the universality consideration leads to a common solution that satisfies the majority of stakeholders. A universal system is necessary because it can support a wide range of users [55]. A universal AT with multi-modal input and multimedia output interface as proposed by Karpov and Ronzhin [56] was based on four assistive systems: automatic audio-speech recognition (relevant for blind persons), text-to-audiovisual speech synthesis (relevant for blind persons), sign language visual synthesis (relevant for deaf persons), and multimodal hand-free computer control (relevant for motor disabled people). In the same manner, sustainability can be associated with how technology can adapt over time to a person’s changing needs.
In this project, an AT proper based solution (AT-PBS) has been used to (a) improve the functional capabilities of the PWDs; (b) improve the information and communication means of the PWDs; and (c) improve the service support accessibilities of the PWDs. Two design models, including the ATA model [25] and the human activity assistive technology (HAAT) model [57], have been adopted, as shown in Figure 1. In phase 1, the objective was to define the requirements and needs of the PWDs. This was done by directly taking feedback from the PWD users, obtaining expert opinions, and analyzing the existing solutions in order to identify gaps. In phase 2, the professional team of the project identified the solution satisfying the needs of the PWDs and interacted with them to get their feedback. This was managed through environment assessment process, matching assessment and suggesting relevant solutions as in ATA. Finally, in phase 3, the proposed solution was tested by PWDs for ease of use and desired function. This has been aligned to Phase 3 of the ATA model and the assistive technologies assessment of the HAAT model.

3.2. AT Design Consideration for PWDs on the University Campus

To support a universal solution, there have been different ATs that can be used to support speech impairment (SI), hearing impairment (HI), visual impairment (VI), and motor impairment (MI). Table 1 presents the types of interfaces and ATs that can assist PWD with a particular impairment condition, where (x) signifies a technology that is relevant for that particular impairment.
A Smart system must have at least one input and one output to allow PWD users to interact with the system effectively. In addition, it should have mobile technology to support them while moving. For input, it should have Voice Recognition (1), Voice to text (2), and Keyboard and mouse (3), while gesture control can be additional. For output, it should include Text-To-Voice (5) and Display Screen (6), while Vibration and Flashing (7) is supplementary. For mobility support, it should consist of Mobile Platform devices (8) and Tracking Location and Behavior (9). The needs of PWDs (phase 1) at the university campus were identified using the following three steps: surveying PWDs to know their requirements for accessibility solutions, obtaining expert opinions, conducting a literature review, and through surveying existing technological solutions to identify solutions’ strengths as well as areas of improvement. The implementation of the interface that supports the campus network and the unified interface that takes care of internal network coverage inside the building, as well as choosing the right technology for the voice recognition engine have been described in the section “Implementation of the USCAS-PWD”.

3.2.1. Identify the Requirements and Needs of PWDs (Phase-1)

In order to identify the needs of the PWDs, a user-centric design approach involving the surveying of PWD stakeholders is useful. Since PWDs might not be aware of the different AT solutions, there is a need to verify the relevant technologies, based on expert opinion and according to literature review. Moreover, there is a need to assess existing research and commercial technological solutions in order to identify criteria associated with the best practices and identify the right tools that can be used for voice recognition.
Step 1:
Survey PWDs to know their needs
Conventionally, the majority of customer product design processes do not explicitly pay attention to the needs of older or disabled people, which is in fact a wrong practice. There exists a number of approaches to surveying PWD stakeholders, such as observation, surveying, and interviewing, etc. As mentioned by Green et al. [58], the user needs vary depending on the product usage context. A very common practice used to identify the needs of PWDs involves discussion with the actual user or their caretakers [59].
Similar to the work done in [60,61], in this project also, a survey targeting PWD students has been conducted using a questionnaire form which identified their needs (as shown in Appendix-A1), and the results of the survey are presented in Appendix-A2. The results revealed the following main requirements for designing an AT-based solution:
  • Accessible system: To access any information related to the curriculum while walking on campus.
  • Mobile searchable system: Ability to search for location and use of the LBS to reach their destination.
  • Supported communication system: Ability to communicate with others and their teachers as well as their caregiver from anywhere on the campus, while moving.
  • Auto-searchable system: Ability to find and talk to any person that matches their profile.
  • Universal Interface system: To support different impairment conditions of the PWDs.
  • Comprehensive solution: For information access, communicating with others, support of editing tools, and tracking users for intervention and guidance.
Step 2:
Identify the needs of PWDs through Experts and Literature Review
Another relevant technique that can be used to obtain the needs of PWD stakeholders is through expert opinion, and previous research work in that particular field. Butterfield and Ramseur [62] conducted a comprehensive literature review of the use of ATs at the workplace. They concluded that several needs of the disabled employee, such as accommodation, should be near to the workplace; the path used for the daily movement should be safe and easy; and proper navigation facilities should be provided. Yeager et al. [63] conducted research to study the barriers at the workplace for PWDs. It was reported by the PWDs that AT helped them to overcome some of the barriers; however, some issues still needed to be addressed. Galvin and Scherer [64] wrote a book on the evaluation, selection, and use of various ATs. Out of 15 chapters, ten were written by PWDs. The use of AT in various fields, such as the workplace, schools, playgrounds, etc. was described. The barriers, solutions, selection criteria, and future technologies were reported in the book. Based on the literature review and the issues highlighted in the introduction section, the following key needs for PWD have been identified:
  • AT system or application should be user friendly
  • AT solution should be mobile
  • AT solution should be accurate and flexible to meet user needs
  • AT solution should provide good help and support documentation
  • AT solution should be comprehensive in nature
  • AT solution should be interoperable
Step 3:
Survey of Existing Technological Solutions for PWDs Gap Analysis
Smartphone(s) have become very popular these days, and they can be easily owned by people to support mobility in the university campus. Therefore, many research and commercially available assistive solutions work on smartphone(s) platforms. A total of ten research/commercial products have been evaluated (as shown in Table 2) according to design criteria defined in Table 1. This helped the designers to identify the relevant criteria that support AT based on best practices. In addition, a brief assessment has been done based on the criteria of usability, support, and maintenance, as shown in Table 3 and Table 4 [65,66,67,68,69,70,71,72,73,74].
Table 3 provides the assessment of 10 research and commercial assistive applications. This assessment has been performed by scoring each technology (from 0 to 10) based on the evidence of child criteria of usability and maintainability, as identified in Table 2. It is clear that 3 out of 10 of these products have a good score for the majority of these criteria, and therefore they meet the specific conditions of PWD users in terms of usability and maintainability.
Innovative technologies can assist researchers and developers in providing intelligent solutions for PWDs which efficiently interact with the environment. These technologies employ different input methods, such as voice recognition, voice-to-text, and gesture control using sensors. They use different output methods, such as text-to-voice, vibration and haptic control, in addition to controlling devices, such as wheelchair and smart phone guidance systems. The current research lacks a complete unified smart solution relevant to the PWDs at the workplace or in the university environment. Therefore, this work has addressed this issue, as presented in Table 4. The research and commercially available solutions have mainly focused on a particular group of PWD. This means that the target group is not diverse. These types of systems are incomplete if a person suffers from more than one type of disability. Therefore, there is a need of a comprehensive solution that offers services to a wide range of PWD groups, rather than focusing on one group. In this work, a comprehensive and smart solution has been developed and implemented in an academic work environment. In order to design a supportive solution for PWDs, it should exhibit the following characteristics: usability in terms of flexibility, easy to use, adaptive to user conditions, and good maintainability after release.
  • Conclusions from Phase 1
An efficient smart and context aware system, on the university campus can be achieved through a universal interface that supports the needs for occupational therapy and rehabilitation (Medical model) [76], and empowers PWDs through a user-centered solution (Social model) [77]. The following features of AT should be considered:
For an AT to suit all conditions of PWDs for better functional capabilities, and mandatory requirements for an AT-PBS (a) it should include the following features:
  • Speech-to-text
  • Text-to-speech and synthesis
  • Speech/Voice command control
  • Display enhancement through magnification and zooming
  • Speech volume control
  • Keyboard, mouse, and touch screen
For AT to support full Information accessibility, and enhancements to an AT-PBS (b) it should include the following features:
  • Automatic emergency response
  • Comprehensive support of activities at the university campus, including smart help including an auto-searchable system, profile setup, noting, and communication.
For AT to have better service support accessibility, and enhancements to an AT-PBS (c) it should include the following features:
  • Mobility support
  • Tracking, behavior analysis, and intervention
  • Adaptable interface according to user profile
  • Design for all by supporting universality
  • Overcome existing environmental and social barriers by supporting usability solutions that are adaptive to the user’s conditions.
There can be more enhancements to support extra functional capabilities, such as gesture control and video recognition, video synthesis, sign language and avatar, Braille keyboard, and controlling the movement of a wheelchair through smart interface.

3.2.2. Identifying the Matching Solution Including USCAS-PWD System Design (Phase 2)

USCAS-PWD should incorporate the features identified in phase 1. As a result, a universal prototype solution that is secure, comprehensive, and which supports the needs of the majority of PWDs at the university campus have been developed. This proposed solution would be adaptive in nature; i.e., it would allow its users—including PWD students—to conduct a smart and regular search on the university campus, as well as inside the building. It also allows the users to interact with the selected party using different effective mechanisms, such as SMS, email, chatting, notification, and events booking. In addition, these solutions support a smart help feature for the PWDs to smartly identify their user profile and adopt the interface display parameters (color, fonts, etc.) and environments (speaker volume, mic volume, virtual mouse, and virtual keyboard) depending on the requirements as well as the predefined and stored profile. There are two program interfaces that have been designed and developed to support the PWDs at the campus network and inside the building. The first interface, which has been developed using the USCAS-PWD program (based on a web-based SQL database, and written in .net language), covers the university campus. This interface allows users to securely search for the relevant data using Wi-Fi and GPS technology anywhere on the campus while they are mobile or stationary. It also allows them to interact with the selected party effectively, as shown in Figure 2. The second program interface has been designed to cover the internal network within a building using a RFID/WiFi based-network. It covers the Riyadh Techno Valley (RTV) building and is written in the Java language, which interacts with a local SQL database and tracking location database. This interface allows the system to monitor PWD movement and location within the building, as shown in Figure 3. It also allows the PWDs to execute a smart helping program and smart editor program, as will be explained in the “Algorithm” section. The ATs used in the USCAS-PWD and the relevant criteria explained in the section “Design Consideration and Methodology” (Table 1) are illustrated in Table 4, along with other existing solutions.
  • Comparison of Technology Design Features
Comparison of the design considerations of the USCAS-PWD interface program with the 10 existing software applications revealed that USCAS-PWD satisfied the input/output conditions for different impairment PWD groups. However, the other products lacked a comprehensive solution to support PWDs with multiple impairment conditions, as shown in Table 4. All existing software products for PWDs listed in Table 4 lacked the support of a flexible adaptable interface, and therefore could not be used by different PWDs with multiple impairment conditions. The USCAS-PWD can support 11 groups with different impairment conditions, as discussed in the following section. The other important criteria “Tracking Location and Behavior” have been partially supported in two existing solutions, but have been fully supported in USCAS-PWD. In addition, the USCAS-PWD supports speech recognition, along with five other existing solutions. Therefore, only USCAS-PWD satisfies the needs of the PWDs, in terms of supporting all necessary input/output technologies that are needed by all PWD groups. It also supports tracking location and behavior to assist PWDs at the work environment and on the university campus.
  • Impairment Groups (Design and Implementation Consideration)
In order to cover the different PWD categories according to impairment conditions, it was realized that PWD who can see, or partially see, hear or partially hear, can move their hand or cannot, and can speak or are mute can be legitimate users of the system, and the groups which have been devised are shown in Table A1 of Appendix A3.
For a person with full vision and partial vision requirements that can co-exist with three other requirements (hearing, speaking, and moving hand), there would be 25 = 32 possible combination groups. Since there were constraints for this combination (by having at least one input and one output), the total number of groups were reduced to eight. These constraints restricted the number of groups as follows: for output from the system, a person must see or hear; for input to the system, a person must speak or move their hand. For a person with partial hearing, he/she must see but can/cannot speak or move hand, and therefore three other groups that were applicable (group 9, group 10, and group 11) are shown in Table 5. Hence, Table 5 identifies the different combinations according to different PWD impairments. It consists of the 11 groups that have been defined for different PWDs. The flexible interface would be designed to accommodate all conditions of these groups by adjusting the microphone volume, the speaker volume, and the font color, the font size, and the display window size and color.

4. Implementation of the USCAS-PWD

One of the main design considerations was to choose the right technological tools for voice recognition, which could support the features of the proposed solution at the PWD interface. Hence, a comparison between different technologies to support voice recognition, text-to-voice, and voice-to-text has been carried out as follows.

4.1. Technology Consideration and Comparison to Choose Voice Recognition Software

Voice recognition software plays an essential role in providing various services for PWDs in the proposed system. The choice of this software was made after the assessment of 11 popular software applications, which simplified the development process. This assessment was done in order to select a reliable software tool from existing solutions satisfying the relevant criteria. The assessment criteria consisted of two main parameters—namely, usability and sustainability and maintainability. These two parameters were named as the “Parent” criteria used to assess the software on certain “Child” criteria, as listed in Table 6 and described as follows:

4.1.1. Usability Criteria

Usability included the following “Child” parameters for assessing software applications:
  • Ability to understand, learn, build, and install with the right documentation: These were assessed to determine the flexibility of software—i.e., if they could be easily understood for their functionality and use. They were also assessed to determine their building process and installation procedure (simple or difficult). The scoring was done on a scale of 0 to 10, with 10 assigned to the most flexible software.
  • Speech to Text (STT) and Text to Speech (TTS) facility: 5 points were given to only one feature and 10 to both features (STT and TTS).
  • Web Based: Software accessible on the web was scored at 10, otherwise it was given 5.
  • Downloads: Scoring from 0 to 10 was given based on the number of downloads in a specific period. For example, the software that was downloaded between 500 and 1000 times got 5 points, whereas the one with more than 2000 downloads received 10 points.

4.1.2. Sustainability and Maintainability Criteria

Sustainability and maintainability included the following “Child” parameters for assessing the software:
  • Identity, copyright, community usage, and accessibility: Assessment was based on clear and unique identity of the software. In addition, the evidence of current/future community of the software and the ability to download were evaluated on a scale of 0 to 10.
  • Supportability, testability, and changeability: The software was assessed based on the evidence of existing and future developer support. In addition, the level of flexibility in modifying the software was scored on a scale of 0 to 10.
  • Cost: Free software were assigned a score of 10, while reasonable price was given a score of 8. Expensive price software got a score of 6, and the ones that charged per subscription got only 0 to 5 points.
  • Portability on operating systems: The software were assessed for their suitability on various operating systems, including Windows, Android, and iOS. The scoring was performed on a scale of 0 to 20. The software that was supported across multiple platforms scored 20. The score of 15 was given to one, compatible with Windows, Android, and iOS. The Windows and Android based software scored 10, and a Windows-only package scored 5.
  • Available languages: The package that supported both English and Arabic was given a score of 10, whereas software with only one language scored 5.
  • Available voices: Higher scores were given to software with a greater number of voices for converting sound from text. Software offering more than 15 voices scored 10.
  • Programming languages: Support of the two languages C++ and Java attracted high score, with preference to Java. A score of 5 was given to only C++ supported one, while only Java supported software got 10 points, and support for both languages received a score of 15.
The scores given to 11 software applications were evaluated by two Engineering Programmers who developed the smart solution for the project in order to select a relevant software tool for voice recognition that can be integrated with the smart unified interface. These assessment results are based on the software application features, with evidence as shown in Table 6 [78,79,80,81,82,83,84,85,86,87], where the evidence of C3.1, C4.1, and C4.2 were given directly by the Engineers according to their test assessment. The numerical assessment is presented in Table B1 and Table B2 of Appendix B.
The software applications were assessed, and were based on two main Parent criteria—i.e., Usability and Sustainability and Maintainability. The output results are shown in Figure 4 and Figure 5. As can be seen in Figure 4, the J.A.R.V.I.S speech API scored the highest (38 points) in criterion Usability, whereas VOCE ranks as the fourth top speech API in regards to Usability. However, by considering the highest scores of Sustainability and Maintainability features of voice recognition software, VOCE attained the highest score with 75 points, whereas Sphinx came second at 70 and J.A.R.V.I.S speech API and iSpeech’s got scores of around 65 points, as shown in Figure 5. Since Sustainability and Maintainability features are considered critical to our application, VOCE became a clear choice for a speech recognition system, as it also satisfied other requirements, such as supporting both Java and C++. The software was freely available and it worked on all of the operating systems, including Windows, Android, and iOS. It also enabled both text-to-speech (TTS) and speech-to-text (STT). It also offers several speaking voices in the package and works in the English language. Although VOCE is not web-based like J.A.R.V.I.S, VOCE was the most suitable, as this feature was not very necessary.

4.2. Implementation of the Unified Interface for the Indoor Network

According to the design considerations in previous sections, the implemented solution would address the following important features:
  • Universal interface that suits the majority of PWDs,
  • Comprehensive solution that supports all the services needed at the workplace and on the university campus,
  • Support of the mandatory requirements of AT listed in the section “Identify the requirements and Needs for PWDs”,
  • Support of the enhancement of AT to support full Information accessibility and automatic emergency response,
  • Support of service capabilities, including mobility, tracking and intervention, behaviour analysis, adaptable user interface, and overcoming environmental and social barriers.
Therefore, a unified interface has been developed to allow PWD users/students to log into the system and adjust the display for different font size relevant to them, even at the first login window (as shown in Figure 6 and Figure 7), and as presented in movie-1 of the supplementary file “CommonUnified.mp4”. The user can login at the interface either using a keyboard with the relevant font size, depending on his/her condition, as well as using RFID or voice recognition. According to the user profile, the different interface with relevant font size and window color and size can be displayed. The user profile is set up either according to the default group he/she belongs to, or by customization through user preference. The first indoor interface for PWD after login allows the user to select between different applications that are set up to customize user profile, environment setup, smart editor, and smart help, as shown in Figure 8. Note that the privacy issue has also been addressed in the developed solution through encryption and authentication. The user would be authenticated during login using user name and password, and it will be based on RFID login in a future prototype version, while their messages would be encrypted. Three main program features, including setting up user profile and environment profile, smart editor and smart helping (as shown in Figure 8) have been supported for PWDs to provide a comprehensive solution on campus. Users can drive the interface program through keyboard, mouse, or voice-to-text by using the short-cut that is displayed in red in Figure 8, Figure 9, Figure 10 and Figure 11. Smart helping allows the PWD users to retrieve relevant guiding information from the help server using voice recognition, which would translate the voice into text and then into a command to get information from the server, as shown in Figure 9 and Figure 10, and as presented in movie-2 of the supplementary file “SmartHelpDesktop.mp4”. As shown in Figure 9 and Figure 10, the USCAS-PWD program allows users to select among multiple search options: Main Search includes search list, keyword search, search employee by location, and search history; Building info; Name of Employee, to get full information about any employee working in the building; Auto Search, which includes search by location and search by profile, which allows search results to be matched to user properties such as interest and specialties, where a list of users that match the user properties (interest or specialty) will be displayed on the auto search interface; and My Location, which determines the current user location within the building. In the search List, PWD users are given the option to select an item from the list, where these items are related to amenities in the building, as shown in Figure 10.
In addition, the users would be able to setup events and communicate with the selected party through different mechanisms, such as sending SMS or email, or call other party, as shown in Figure 11 and as presented in movie-3 of the supplementary file “CommInter-BehAnala-ProfileSetup.mp4”. User location tracking within a building can be done using the RFID–WiFi system, which has been integrated with the unified interface to assist users in determining his/her location as well as tracking movement on map to reach a particular destination. The event behavior algorithm (Section 4.4.2) has been integrated to the track server, which determines if the user has executed the event according to schedule and sends an intervention message to him/her or her/his caregiver to advise them about a missing behavior or about the date and time of the events. Event setup allows the user to setup a new event by determining the event’s location, date, and time, as well as sending a reminder by using email or SMS, as shown in Figure 11. The intervention shown in Figure 11 allows the PWD user to setup and view the intervention messages that were sent to him/her as part of executing the event behavior algorithm, as described in section “Event and Noting Behavior Algorithm” and as presented in movie-3 of the supplementary file “CommInter-BehAnala-ProfileSetup.mp4”. The location tracking feature shown in Figure 11 allows the PWD user to track location while moving toward their destination in real time, and an alert message would be sent to him/her if moving in wrong direction according to movement algorithm, described in Section 4.4.1. The users would have the choice to disable tracking by not sending his/her identity through the interface if he/she is not willing to be tracked. Personal reminder allows the users to view and control the current and previous events’ status.
Furthermore, the solution would provide help or mergency response (as displayed in Figure 9, Figure 10 and Figure 11) to handle difficult situations and assist PWDs in getting the right help and guide them to solve their problems. The emergency call would be handled by the Auto Emergency Response (AER) Program installed on the server, as illustrated in Figure 12. The client terminal attached to the PWD user would be tracked continuously by the AER program through a WiFi AP that has been connected to the AER server through a wired or wireless connection, and the location data would be stored in the MSQL server (sequence 1). If a PWD user issues an emergency call through speech or by pressing a button, a signal would be sent to AER server (sequence 2), where the AER program would then request the location and map from the MSQL server to identify the user ID and its location (sequence 3). Then, the AER server would send a reply to the PWD user terminal with a few options, indicating the type of help they require for further assistance. The message received at the requested PWD terminal would be played to the PWD and displayed on the terminal to choose one of the help options he/she wants (sequence 5). The PWD user can then speak the chosen option, which would be sent to AER server (sequence 6) for analysis by the AER for further action (sequence 7). If the response coming from the PWD is according to the type of help that was sent to him/her, then a guidance message would be sent to the PWD again (sequence 8 with new message).
A light universal interface program has also been developed to run on a smart mobile (iPhone). It would allow the user to interact through the system to get information stored in MySql server. This has been achieved through an application running on the iPhone, which interfaces through a Web service based on Web Services Description Language (WSDL). This web service converts the received messages in C# format, so that it can be understood by the iPhone terminals, as shown in Figure 13 and Figure 14. In Figure 13, five options for the search have been supported. In Figure 14, the execution of search “Name of Employee” displayed three names. When character “g” is entered, two names have been filtered out, and more information about the selected name is displayed. Movie-4 of the supplementary file “MobileAddDataToDB.mp4” demonstrates how to add user data to the database using an iPhone.

4.3. Implementation of the System for Outdoor at the Campus Network

The USCAS-PWD program has been implemented using the asp.net language and SQL database to support PWDs while moving at the campus, where the client can access the information using the web interface, as shown in Figure 3. The main relational tables represent building, researchers, cluster, events, keyword-search, specialist, researcher’s specialist, booking, and sending SMS and email. The fields associated with the building table are building ID, map, photo, IP address, location, building description, and image size. User would be able to book for an event according to chosen location found in the building. The researcher table is linked to the building table through building ID, cluster table through cluster ID, researcher specialist table through its ID, keyword-search ID, and sending SMS and email table Section ID. It also includes many fields regarding the researchers. This would allow the users to get all the information regarding the building and its researchers, as well as information regarding the researchers that are relevant to their domain cluster and specialist. Furthermore, users would be able to communicate with a selected found party using SMS, chat, or email.
The building IP address field included in the building table would be used by the USCAS-PWD program to make a comparison with the reading IP cluster of the user after it has been associated with the access point. Then, the USCAS-PWD program would determine the location of the users if the IP match was determined, and subsequently display the search result according to the search option that has been selected by users. The GIS mechanism presented in this paper provides a flexible cheap solution for users who don’t have GPS in their terminals. The location-based search is a smart adaptive search solution that can dynamically determine the user’s location on campus using WiFi IP subnetting-based location technology or GPS technology, and then display its location, information about the location’s department and its researchers, as well as any event associated with this department. Satisfactory feedback has been obtained from PWDs, students, and the developers who tested the prototype USCAS-PWD program. King Saud University’s (KSU) students—including the PWDs as well as researchers and RTV staff—would benefit from the USCAS-PWD program, since it allows them to display the matching research profile as well as relevant guidance, where the mobile devices would receive information regarding a building or a group of buildings in the surrounding areas and display it on the user’s terminal.

4.3.1. Search by Keyword and Booking Event

Users are able to search for any keyword associated with other researchers, clusters, or specialists anywhere on the university campus, whether the users are stationary or mobile. As shown in Figure 15, a user searches for a particular keyword in the cluster option “scientific valley”, and the USCAS-PWD program displayed the list of all matching researchers, gave the option to select a particular researcher displayed in the list, and displayed the relevant information about the researcher and the building he/she works in, as well as any event planned for this building. As shown in Figure 15, there are many researchers that have been found on the matching list, and the selected researcher information for “Saeed” and his RTV information are displayed. The user has also executed the “Find Direction” option of the Engineering department, where he is located at the science department, as shown in Figure 15. The arrow on the bottom right side in Figure 15 indicates the direction from the current location of the user to the destination that has been selected by this user.

4.3.2. Auto-Search by Location and Find Direction

While user is moving on the university campus, the USCAS-PWD program would determine the user’s location and display information about the department existing in that location, as well as a list of all researchers in this department. If the user selects a researcher from the list, the relevant information of the selected researcher (e.g., “Dr. Abdulmuhseen” from Science) is displayed on left bottom of the window, and the information about Science department and its future events are displayed on the right window, as shown in Figure 16. The user interacts with the USCAS-PWD interface by booking the event. In addition, the user executed the Find Direction from the science building to the Engineering building, where the arrow on Figure 16 displays the direction.

4.3.3. Auto-Search by User Profile and Sending Email and SMS

A user profile can be set explicitly by the user according to its predefined preferences or implicitly by observing user actions while applying the keyword search, which can update the interest field. This would allow the user profile to be updated dynamically according to the collection of search entries executed by the user. The auto-search by the user profile would first determine the user location and then identify the researcher’s profiles that match with the user’s profile and display them on the most suitable format of the matching result, as shown in Figure 17. Two options associated with user profile have been displayed for the user, which are specialist and interest. The interest list will be added by the user to indicate his/her interests. The interest list can also be updated dynamically based on previous keyword (KW) search, and previous Loc searches that have been added to the interest list if they have been selected by the same user with a threshold of n = 5 times. In Figure 17, the user has selected the previous Loc search, where the Loc Scientific valley has been chosen, and user chooses Dr Abdulrahman from the list of researchers. On the right hand side of the USCAS-PWD interface, the user is sending an email to Dr Abdulrahman, as shown in Figure 17. If the user chooses the specialist option, and since there are multiple keywords in his/her specialist field or in the interest field/Keyword search field, the USCAS-PWD program would keep checking the database to determine if it has a matching word in the specialist or interest fields that exists at the same department location and display them on the user interface. Figure 18 displays the results of the user’s profile-based search and previous KW option of the interest list, where “Network Management” is chosen, and Ghassan Kbar is selected from the displayed list. The user can also send an SMS to the same researcher, as shown in Figure 18.

4.4. Algorithms

The following subsections describe the algorithms that have been implemented to determine the movement direction of the PWD user toward a destination. It also describes the event and noting behavior algorithm, which determines the right actions to be taken according to scheduled events and the conditions of the PWD user, including date and location.

4.4.1. Algorithm for Determining the Direction of PWD Movement within the Building

To check the movement of the current behavior and predict if it leads to correct expected movement direction, we need to check the following:
Assume that PWD is moving to target location (meeting room according to behavior a1 with room coordinate a1(x3) = (Tx, Ty), the current PWD user coordinate location is (Cx, Cy), and the updated PWD user coordinate after period of T is ( C x ,   C y ).
There are four possible relative locations of PWD in relation to target room a1 as follows:
  • Cx > Tx and Cy > Ty (Group location A, east–west)
  • Cx > Tx and Cy < Ty (Group location B, east–south)
  • Cx < Tx and Cy > Ty (Group location C, west–north)
  • Cx < Tx and Cy < Ty (Group location D, west–south)
The following presents the possible movement direction to determine if it is heading to target location or not:
For Group A: if C x ↓ decreases compared to Cx and C x ( T ) , and C y ↓ decreases compared to Cy and C y ( T ) , then the object is moving in the right direction, or else is moving in a false direction, where C x is the current coordinate location of the user, C x ( T ) is the previous coordinate location of the user, and Cx is the original coordinate location of the user.
For Group B: if C x ↓ decreases compare to Cx and C x ( T ) , and C y ↑ increases compared to Cy and C y ( T ) , then the object is moving in the right direction, or else is moving in a false direction.
For Group C: if C x ↑ increases compared to Cx and C x ( T ) , and C y ↓ decreases compared to Cy and C y ( T ) , then the object is moving in the right direction, or else is moving in a false direction.
For Group D: if C x ↑ increases compared to Cx and C x ( T ) , and C y ↑ increases compared to Cy and C y ( T ) , then the object is moving in the right direction, or else is moving in a false direction.
The movement algorithm assists in predicting the movement direction of PWD users toward a destination and alerts them when they are moving in the wrong direction.

4.4.2. Event and Noting Behavior Algorithm

In this section, the noting and event algorithm that periodically (every 4 h) monitors the events set up by PWD users, as well as their location to determine how to intervene with them and their caregivers, has been described. In addition, messages would be logged in the database to be analyzed by the personal reminder interface. Figure 19 displays the event algorithm and how it tracks the user and updates event status to “Done” or “Missed”, according to the user’s location and scheduled date. Figure 20 illustrates the algorithm for monitoring personal noting according to four cases, which are break, medication, pickup, and birthday. Based on different conditions associated with scheduled date, alert date, and time, the user’s location, and noting status, different actions would be undertaken for these noting events. These actions can be sending an alert message to the user to remind him/her about taking medication on time, about pick up date and time, about birthday time, or break time. In addition, actions would include changing the noting event status to “Done” or “Missed”, depending on the location of the user, as well as getting his/her response on time. These algorithms would be executed on the server to monitor the execution of scheduled events of all PWD users and to remind them about future events or inform them about event inconsistency.

5. Discussion and Conclusions

A universal interface that supports AT solution has been designed by following the ATA and HAAT design methodologies to satisfy the needs of PWDs. The proposed solution is based on a Smart and Context-Aware at University Campus for PWDs. The USCAS-PWD represents a smart solution that is universal, adaptive, mobility supportive, flexible, and provides an effective comprehensive solution, including search for implementing a professional networking tool, which can be used by the researchers and students and PWDs with different impairment conditions. This application interface supports PWDs on the university campus and inside a building that is equipped with a RFID–WiFi network. The USCAS-PWD program uses asp.net programming languages along with SQL database for the campus network, and RFID–WiFi technology indoors to determine the user’s location. The user location-based search, along with the adaptive user profile and keyword-search, together make the GIS solution—smart and very effective in determining the relevant information for parties as well as providing the ability to communicate with others effectively. The implementation presented in this paper proves a successful smart flexible search and professional networking tool that makes the working environment for students—including PWDs—more productive. The Smart Help feature facilitates the usage of a system for PWD that works on their voice commands by voice recognition system. Furthermore, an intervention event and behavior-based algorithm, as well as a direction movement algorithm have been devised, which help in identifying the behavior of PWD students and taking the right actions to inform caregivers about any issues so that they can be aware of PWD critical activities. This tracking of behavior assists PWDs by alerting them or their caregiver about future or missing events. Moreover, the global acceptance and the cultural aspects of this type of technology can be realized from the positive feedback received from the users. The feedback that has been received actually reflects the global acceptance of the proposed technologies.
The proposed universal interface based on USCAS solution to support multiple PWD users on the university campus proved to be successful in meeting their needs while moving indoors, sitting in the class rooms, and moving outdoors on the university campus. This is done through smart help and communication solutions as well as smart tools that are context-aware and help PWD users and students to do their task effectively. However, some of the issues related to conducting extensive user studies to evaluate the diverse environment and conducting relevant usability studies need to be done, so that a better solution can be achieved. Furthermore, addressing the privacy issues of PWDs needs to be improved in the future in order to accommodate all impairment conditions and cultural diversity. Future research should incorporate PWD students’ needs at the university level to address the curriculum for different departments, to integrate the student registration interface, and to allow PWD students to access their related course materials and interact with them online. Moreover, the AT smart interface should support student forums to address their needs so that feedback can be collected from these forums to improve the design interface continuously.
For future endeavors, the authors would evaluate the interconnectivity of the proposed system with other technologies, such as cars, bikes, etc. They will also consider implementing the solution to run on different mobile platforms, such as Android and Windows.

Supplementary Materials

The following are available online at www.mdpi.com/2073-431X/5/3/18/s1, Video S1: CommInter-BehAnala-ProfileSetup.mp4; Video S2: CommonUnified.mp4; Video S3: MobileAddDataToDB.mp4; Video S4: SmartHelpDesktop.mp4.

Acknowledgments

This Project was funded by The National Plan for Science Technology and Innovation (MARRIFAH), King Abdul-Aziz City for Science and Technology, Kingdom of Saudi Arabia, Award Number (12-ELE3220–02).

Author Contributions

The first author has conceived and designed the system; all authors analyzed users’ needs and determined the matching of AT solution. A development team consists of 4 engineers have developed the proposed solution. All authors contributed to writing the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
PWDPeople with disabilities
RFIDRadio-frequency identification
GISGeographic Information System
USCASUniversity-based Smart and Context Aware Solution

Appendix A1

The following questionnaire was developed to collect more users’ requirements. It was sent to a number of users from all four groups.
Computers 05 00018 i001

Appendix A2

The results are shown in the table below. The users were different from those who participated in the other elicitation methods. A total of 100 responses were collected and analyzed (25 for SI, 25 for HI, 25 for partial VI, and 25 for partial MI).
Strongly AgreeAgreeNeutralDisagreeStrongly Disagree
Using Assistive Technology at work is important97%3%0%0%0%
I prefer to use AT to achieving different tasks at work97%3%0%0%0%
The _______ is my best means to insert input to a computing device:
   Keyboard48%2%3%2%45%
   Gesture control30%23%19%5%23%
   Sign language2%7%2%40%49%
   Speech to text45%10%10%30%5%
   Mouse49%4%1%5%41%
   Braille keyboard0%3%1%26%70%
The _______ is my best means to receive output to a computing device:
   Vibration0%5%30%45%20%
   Screen70%8%0%12%10%
   Text to speech22%7%50%15%6%
   Speech and volume control21%8%50%12%9%
   Magnification80%15%5%0%0%
   Flashing15%13%40%27%5%
At work, Getting info from the system about other employees is helpful 80%15%5%0%0%
At work, Getting info from the system about the building is helpful 95%5%0%0%0%
At work, Getting info from the system about work related issues is helpful98%2%0%0%0%
At work, Getting help in editing a doc from the system is helpful96%4%0%0%0%
At work, a system that provides easier communications with others is helpful85%10%5%0%0%
At work, a system that provides magnification is useful30%20%45%5%0%
At work, a system that provides font and background color control is useful27%17%35%21%0%
At work, a system that provides health monitoring and intervention is useful99%1%0%0%0%
At work, a system that provides zooming is useful32%19%47%2%0%
At work, a system that provides movements monitoring and navigation is useful45%17%5%25%8%
At work, a system that provides behavior analysis is useful75%13%8%2%2%
At work, a system that provides behavior prediction is useful72%19%5%1%3%

Appendix A3

Table A1. Different impairment classifications.
Table A1. Different impairment classifications.
SerialSeePartially SeeHearingSpeakMove HandGroup
100000NA (at least partially see)
200001NA (at least partially see)
300010NA (at least partially see)
400011NA (at least partially see)
500100NA (at least partially see)
600101NA (at least partially see)
700110NA (at least partially see)
800111NA (at least partially see)
901000NA (Must hear if partially see)
1001001NA (Must hear if partially see)
1101010NA (Must hear if partially see)
1201011NA (Must hear if partially see)
1301100NA (at least speak or move hand)
14011018
15011107
1601111NA (overlap with group 7 and 8)
1710000NA (at least speak or move hand)
18100012
19100101
20100115
2110100NA (at least speak or move hand)
22101014
23101103
24101116
2511000NA (at least speak or move hand)
2611001NA(Same as 2)
2711010NA(Same as 1)
2811011NA(Same as 5)
2911100NA (at least speak or move hand)
3011101NA(Same as 4)
3111110NA(Same as 3)
3211111NA(Same as 6)

Appendix B

Table B1. Details of voice recognition software applications.
Table B1. Details of voice recognition software applications.
TTS—STT (C3.1)Web Based (C3.2)Downloads Number (Last Month) (C3.3)Cost (C4.3)Operating System(s) (C4.4)Available Language(s) (C4.5)Available Voices(s) (C4.6)Programming language (C4.7)
FreeTTS [78]TTSNo1965FreeCross-platformEnglishSeveralJava
Sphinx-4 [79]STTNo16,204FreeCross-platformEnglishSeveralJava
Voce [75]BothNo508FreeCross-platformEnglishSeveralJava/C++
J.A.R.V.I.S [80] Speech APIBothyes5792FreeCross-platformEnglishSeveralJava
iSpeech’s [81]Bothyes900CommercialWindows, AndroidEnglish20Java
Neospeech [82]TTSyes6991CommercialWindowsEnglish13C/C++/Java
Elan Speech Cube [83]TTSNoNAFreeCross-platformEnglishSeveralJava
AT&T Natural Voices [84]TTSNoNACommercialWindowsEnglish20C++
CereProc [85]TTSNo144CommercialWindows Android, iOSEnglish21C/Python
eSpeak [86]TTSNo9339FreeWindowsEnglishSeveralC++
Festival Speech Synthesis System [87]TTSNo10FreeWindowsEnglishSeveralC++
Table B2. Comparison of speech-recognition software applications.
Table B2. Comparison of speech-recognition software applications.
FreeTTSSphinx-4VoceJ.A.R.V.I.S. Speech APIiSpeech’sNeospeechElan Speech CubeAT&T Natural VoicesCereProceSpeakFestival Speech Synthesis System
C1.181010886668108
C1.255101010555555
C1.355510101055555
C1.4810510510555105
Tot_12630303833312121233023
C2.181010886668108
C2.281010886668108
C2.3101010105555585
C2.4202020201052051555
C2.555555555555
C2.655551085101088
C2.7101015101015105555
Tot_26670756556505742565144
Total9210010510389817863798167

References

  1. World Health Organization. World Report on Disability “The Way Forward: Recommendations, Chp 9”. 2011. Available online: http://www.who.int/disabilities/world_report/2011/chapter9.pdf (accessed on 26 July 2016).
  2. Antona, M. The Universal Access Handbook; CRC Press: London, UK; New York, NY, USA, 2009. [Google Scholar]
  3. Chen, C.-M. Intelligent location-based mobile news service system with automatic news summarization. Expert Syst. Appl. 2010, 37, 6651–6662. [Google Scholar] [CrossRef]
  4. Longley, P.A.; Goodchild, M.F.; Maguire, D.J.; Rhind, D.W. Geographic Information Systems and Science; Wiley: New York, NY, USA, 2001. [Google Scholar]
  5. Li, C. User preferences, information transactions and location-based services: A study of urban pedestrian wayfinding. Comput. Environ. Urban Syst. 2006, 30, 726–740. [Google Scholar] [CrossRef]
  6. Zipf, A.; Jöst, M. Implementing adaptive mobile GI services based on ontologies: Examples from pedestrian navigation support. Comput. Environ. Urban Syst. 2006, 30, 784–798. [Google Scholar] [CrossRef]
  7. Huang, Y.; Bian, L. A Bayesian network and analytic hierarchy process based personalized recommendations for tourist attractions over the internet. Expert Syst. Appl. 2009, 36, 933–943. [Google Scholar] [CrossRef]
  8. Gervais, E.; Liu, H.; Nussbaum, D.; Roh, Y.-S.; Sack, J.-R.; Yi, J. Intelligent map agents—An ubiquitous personalized GIS. ISPRS J. Photogramm. Remote Sens. 2007, 62, 347–365. [Google Scholar] [CrossRef]
  9. Aoidh, E.M.; McArdle, G.; Petitc, M.; Ray, C.; Bertolotto, M.; Claramuntc, C.; Wilson, D. Personalization in adaptive and interactive GIS. Ann. GIS 2009, 15, 23–33. [Google Scholar] [CrossRef]
  10. Brossard, A.; Abed, M.; Kolski, C. Taking context into account in conceptual models using a Model Driven Engineering approach. Inform. Softw. Technol. 2011, 53, 1349–1369. [Google Scholar] [CrossRef]
  11. Jung, J.J. Contextualized mobile recommendation service based on interactive social network discovered from mobile users. Expert Syst. Appl. 2009, 36, 11950–11956. [Google Scholar] [CrossRef]
  12. Chang, S.K.; Yung, D.; Colace, F.; Greco, L.; Lemma, S.; Lombardi, M. An Adaptive Contextual Recommender System: A Slow Intelligence Perspective. In Proceedings of the 27th International Conference on Software Engineering and Knowledge Engineering, SEKE, Pittsburgh, PA, USA, 6–8 July 2015; pp. 64–71.
  13. Rebollo-Monedero, D.; Forné, J.; Solanas, A.; Martínez-Ballesté, A. Private location-based information retrieval through user collaboration. Comput. Commun. 2010, 33, 762–774. [Google Scholar] [CrossRef] [Green Version]
  14. Kbar, G.; Aldosari, A. Interactive—Smart and Context Aware Search for KSU University Campus (I-SCASUC). In Proceedings of the 2nd International IT Conference, Tangier, Morocco, 10–12 May 2012.
  15. Schofield, G. Novel GPS tracking of sea turtles as a tool for conservation management. J. Exp. Mar. Biol. Ecol. 2007, 347, 58–68. [Google Scholar] [CrossRef]
  16. Gupta, R.; Reddy, B.V.R. GPS and GPRS Based Cost Effective Human Tracking System Using Mobile Phones. Viewpoint 2011, 2, 39–45. [Google Scholar]
  17. Chandra, A.; Jain, S.; Qadeer, M.A. GPS Locator: An Application for Location Tracking and Sharing Using GPS for Java Enabled Handhelds. In Proceedings of the International Conference on Computational Intelligence and Communication Networks (CICN), Gwalior, India, 7–9 October 2011; pp. 406–410.
  18. Ekahau. Available online: http://www.ekahau.com/real-time-location-system/technology/how-rtls-works (accessed on 28 February 2016).
  19. Sanitag. Available online: http://www.sanitag.com/ (accessed on 28 February 2016).
  20. Areoscout. Available online: http://www.aeroscout.com/ (accessed on 28 February 2016).
  21. Ning, H.; Fu, Y.; Hu, S.; Liu, H. Tree-Code modeling and addressing for non-ID physical objects in the Internet of Things. Telecommun. Syst. 2015, 58, 195–204. [Google Scholar] [CrossRef]
  22. Dao, T.K.; Nguyen, H.L.; Pham, T.T.; Castelli, E.; Nguyen, V.T.; Nguyen, D.V. User Localization in Complex Environments by Multimodal Combination of GPS, Wi-Fi, RFID, and Pedometer Technologies. Sci. World J. 2014, 2014. [Google Scholar] [CrossRef]
  23. Li, Z.; Yan, L.; Pan, W.; Luo, B.; Ye, J.; Jiang, H. Simultaneous Transmission of Multiple Wireless Services Over Fiber with Reduced Network Complexities. J. Opt. Commun. Netw. 2014, 6, 26–32. [Google Scholar] [CrossRef]
  24. Szewcyzk, S.; Dwan, K.; Minor, B.; Swedlove, B.; Cook, D. Annotating Smart Environment Sensor Data for Activity Learning. Technol. Health Care 2009, 17, 161–169. [Google Scholar] [PubMed]
  25. Reddy, G.Y.; Reddy, M.R. Design of Ultrasonic Spectacles, Flex Sensor Based Voice Generation and Speech to Text Conversion Techniques for Physically Disable People. Int. J. Res. Electron. Commun. Eng. 2013, 1, 12–22. [Google Scholar]
  26. Chan, M.; Campo, E.; Estève, D.; Fourniolsa, J.Y. Smart homes—Current features and future perspectives. Maturitas 2009, 64, 90–97. [Google Scholar] [CrossRef] [PubMed]
  27. Lê, Q.; Nguyen, H.B.; Barnett, T. Smart Homes for Older People: Positive Aging in a Digital World. Future Int. 2012, 4, 607–617. [Google Scholar] [CrossRef]
  28. Chen, J.; Zhi, L. Blind Path Identification System Design Based on RFID. In Proceedings of the International conference on Electrical and Control Engineering, Wuhan, China, 25–27 June 2010; pp. 548–551.
  29. Hesch, J.A.; Roumeliotis, S.I. Design and Analysis of a Portable Indoor Localization Aid for the Visually Impaired. Int. J. Robot. Res. 2010, 29, 1400–1415. [Google Scholar] [CrossRef]
  30. Kammoun, S.; Parseihian, G.; Gutierrez, O.; Brilhault, A.; Serpa, A.; Raynal, M.; Oriola, B.; Macé, M.J.-M.; Auvray, M.; Denis, M.; et al. Navigation and space perception assistance for the visually impaired: The NAVIG project. IRBM 2012, 33, 182–189. [Google Scholar] [CrossRef]
  31. Darrah, M. The Use of Touch Technology in Science to Provide Access for Students Who Are Visually Impaired. J. Technol. Integr. Classr. 2012, 4, 45–50. [Google Scholar]
  32. Bhatlawande, S.S.; Mukhopadhyay, J.; Mahadevappa, M. Ultrasonic spectacles and waist-belt for visually impaired and blind person. In Proceedings of the National Conference on Communication, Kharagpur, India, 3–5 February 2012; pp. 1–4.
  33. Gallaghera, T.; Wise, E.; Yam, H.C.; Li, B.; Ramsey-Stewart, E.; Dempster, A.G.; Rizos, C. Indoor navigation for people who are blind or vision impaired: Where are we and where are we going? J. Locat. Based Serv. 2014, 8, 54–73. [Google Scholar] [CrossRef]
  34. Willis, S.; Helal, S. RFID Information Grid for Blind Navigation and Wayfinding. In Proceedings of the 9th IEEE International Symposium on Wearable Computers, Osaka, Japan, 18–21 October 2005; pp. 34–37.
  35. Ghiani, G.; Leporini, B.; Paterno, F. Vibrotactile feedback as an orientation aid for blind users of mobile guides. In Proceedings of the 10th International Conference on Human Computer Interaction with Mobile Devices and Services, New York, NY, USA, 2–5 September 2008; pp. 431–434.
  36. Colace, F.; DeSanto, M.; Greco, L.; Lemma, S.; Lombardi, M.; Moscato, V.; Picariello, A. A Context-Aware Framework for Cultural Heritage Applications. In Proceedings of the Tenth International Conference on Signal-Image Technology & Internet-Based Systems, Naples, Italy, 28 November–1 December 2014; pp. 469–476.
  37. Murad, M.; Rehman, A.; Shah, A.A.; Ullah, S.; Fahad, M.; Yahya, K.M. RFAIDE—An RFID Based Navigation and Object Recognition Assistant for Visually Impaired People. In Proceedings of the International Conference on Emerging Technologies, Islamabad, Pakistan, 5–6 September 2011; pp. 1–4.
  38. El-Alamy, L.; Lhaddad, S.; Maalal, S.; Taybi, Y.; Salih-Alj, Y. Bus Identification System for Visually Impaired Person. In Proceedings of the Sixth International Conference on Next Generation Mobile Applications, Services and Technologies, Paris, France, 12–14 September 2012; pp. 13–17.
  39. Jain, D. Path-Guided Indoor Navigation for the Visually Impaired Using Minimal Building Retrofitting. In Proceedings of the 16th International ACM SIGACCESS Conference on Computers & Accessibility, Rochester, NY, USA, 20–22 October 2014; pp. 225–232.
  40. Helal, A.; Moore, S.; Ramachandran, B. Drishti: An Integrated Navigation System for Visually Impaired and Disabled. In Proceedings of the 5th International Symposium on Wearable Computer, Zurich, Switzerland, 8–9 October 2001.
  41. Gomez, J.D.; Bologna, G.; Pun, T. See ColOr: An extended sensory substitution device for the visually impaired. J. Assist. Technol. 2014, 8, 77–94. [Google Scholar] [CrossRef]
  42. Iannizzotto, G.; Costanzo, C.; Lanzafame, P.; LaRosa, F. Badge3D for Visually Impaired. In Proceedings IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; p. 29.
  43. Ando, B.; Graziani, S. Multisensor Strategies to Assist Blind People: A Clear-Path Indicator. IEEE Trans. Instrum. Measur. 2009, 58, 2488–2494. [Google Scholar] [CrossRef]
  44. Cooper, R.A.; Cooper, R. Quality-of-Life Technology for People with Spinal Cord Injuries Review Article. Phys. Med. Rehabilit. Clin. N. Am. 2010, 21, 1–13. [Google Scholar] [CrossRef] [PubMed]
  45. Domingo, M.C. An overview of the Internet of Things for people with disabilities. J. Netw. Comp. Appl. 2012, 35, 584–596. [Google Scholar] [CrossRef]
  46. Pascual, J.; Sanjua’n, J.M.; Cueva, O.; Pelayo, B.C.; A’lvarez, M.; Gonzalez, A. Modelling architecture for collaborative virtual objects based on services. J. Netw. Comp. Appl. 2011, 34, 1634–1647. [Google Scholar] [CrossRef]
  47. Giannetsos, T.; Dimitriou, T.; Prasad, N.R. People-centric sensing in assistive healthcare: Privacy challenges and directions. Secur. Commun. Netw. 2011, 4, 1295–1307. [Google Scholar] [CrossRef]
  48. Doukas, C.; Maglogiannis, I.; Koufi, V.; Malamateniou, F.; Vassilacopoulos, G. Enabling data protection through PKI encryption in IoT m-Health devices. In Proceedings of the IEEE 12th International Conference on Bioinformatics & Bioengineering (BIBE), Larnaca, Cyprus, 11–13 November 2012.
  49. Yau, S.S.; Yin, Y.; Ho, A.H. An Adaptive Approach to Optimizing Tradeoff between Service Performance and Security in Service-Based Systems. Int. J. Web Serv. Res. (IJWSR) 2011, 8, 18. [Google Scholar] [CrossRef]
  50. Templeton, S.J. Security aspects of cyber-physical device safety in assistive environments. In Proceedings of the 4th International Conference on Pervasive Technologies Related to Assistive Environments, Crete, Greece, 25–27 May 2011.
  51. Kaur, M.; Sandhu, M.; Mohan, N.; Sandhu, P.S. RFID Technology Principles, Advantages, Limitations & Its Applications. Int. J. Comp. Electr. Eng. 2011, 3, 1793–8163. [Google Scholar]
  52. Federici, S.; Scherer, M.J. The Assistive Technology Assessment Model and Basic Definitions. In Assistive Technology Assessment Handbook; Federici, S., Scherer, M.J., Eds.; Rehabilitation Science in Practice Series; CRC Press: Boca Raton, FL, USA, 2012. [Google Scholar]
  53. Borg, J.; Lantz, A.; Gulliksen, J. Accessibility to electronic communication for people with cognitive disabilities: A systematic search and review of empirical evidence. Univers. Access Inf. Soc. 2015, 14, 547–562. [Google Scholar] [CrossRef]
  54. Persson, H.; Åhman, H.; Yngling, A.; Gulliksen, J. Universal, inclusive, accessible, design for all; Different concepts—one goal? On the concept of accessibility—Historical, methodological and philosophical aspects. Univers. Access Inf. Soc. 2014. [Google Scholar] [CrossRef]
  55. Mourouzis, A.; Leonidis, A.; Foukarakis, M.; Antona, M.; Maglaveras, N. A Novel Design Approach for Multi-device Adaptable User Interfaces: Concepts, Methods and Examples. In Universal Access in HCI; Springer: Berlin, Germany, 2011; pp. 400–409. [Google Scholar]
  56. Karpov, A.; Ronzhin, A. A Universal Assistive Technology with Multimodal Input and Multimedia Output Interfaces. In Universal Access in Human-Computer Interaction. Design and Development Methods for Universal Access, Proceedings of the International Conference on Universal Access in Human-Computer Interaction, Heraklion, Greece, 22–27 June 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 369–378. [Google Scholar]
  57. Cook, A.M.; Hussey, S.M. Assistive Technology: Principles and Practice, 2nd ed.; Elsevier Health Sciences: Mosby, MO, USA, 2002. [Google Scholar]
  58. Green, M.G.; Rajan, P. Product usage context: Improving customer needs gathering and design target setting. In Proceedings of the ASME Design Engineering Technical Conferences, Salt Lake City, UT, USA, 28 September–2 October 2004.
  59. Department of Education and Early Childhood Development. Human Resources Report on Disability and Reasonable Adjustment in the Workplace. State Government: Victoria, Australia, 2013. Available online: http://www.education.vic.gov.au/hrweb/Documents/Disability-Reasonable-Adj-in-Workplace.pdf (accessed on 31 December 2015). [Google Scholar]
  60. Anttila, V.; Hyppönen, H.; Rathmayer, R.; Mankkinen, E. User needs for personal navigation services—group discussions and diaries. In Deliverable 3, NAVITarve Consortium; VTT Building and Transport: Espoo, Finland, 2001. [Google Scholar]
  61. Hannukainen, P.; Hölttä-Otto, K. Identifying Customer Needs—Disabled Persons as Lead Users. In Proceedings of the ASME International Design Engineering Technical Conferences & Computers and Information in Engineering Conference, Philadelphia, PA, USA, 10–13 September 2006.
  62. Butterfield, T.M.; Ramseur, J.H. Research and case study findings in the area of workplace accommodations including provisions for assistive technology: A literature review. Technol. Disabil. 2004, 16, 201–210. [Google Scholar]
  63. Yeager, P.; Kay, H.S.; Reed, M.; Doe, T.M. Assistive technology and employment: Experiences of Californians with disabilities. Work 2006, 27, 333–344. [Google Scholar] [PubMed]
  64. Galvin, J.C.; Scherer, M.J. Evaluating, Selecting, and Using Appropriate Assistive Technology; Aspen Publication: Gaithersburg, MD, USA, 1996; p. 394. [Google Scholar]
  65. Blind Navigator. Available online: https://play.google.com/store/apps/details?id=com.blindnavigator&hl=en (accessed on 30 May 2015).
  66. Ray Vision. Available online: http://www.project-ray.com/ (accessed on 30 May 2015).
  67. Lend An Eye. Available online: https://play.google.com/store/apps/details?id=air.lend.an.eye (accessed on 30 May 2015).
  68. Talking Camera. Available online: https://itunes.apple.com/us/app/talking-camera-pro-for-visually/id544295997?mt=8 (accessed on 30 May 2015).
  69. TapTapSee. Available online: http://www.taptapseeapp.com/ (accessed on 30 May 2015).
  70. The Deaf and Hearing Impaired. Available online: https://play.google.com/store/apps/details?id=kr.ac.kaist.isilab.doyeob.humanityProject2&hl=en (accessed on 30 May 2015).
  71. SpeechTrans Hearing Impaired. Available online: https://play.google.com/store/apps/details?id=com.nuance.nmdp.ultimate&hl=en (accessed on 30 May 2015).
  72. Speech Assistant AAC. Available online: https://play.google.com/store/apps/details?id=nl.asoft.speechassistant&hl=en (accessed on 26 July 2015).
  73. uSound. Available online: https://play.google.com/store/apps/details?id=com.newbrick.usound&hl=en (accessed on 30 May 2015).
  74. AAC Speech Communicator. Available online: http://freetts.sourceforge.net/docs/index.php (accessed on 20 May 2015).
  75. Tyler Streeter Voce: Open Source Speech Interaction. Available online: http://voce.sourceforge.net/ (accessed on 20 May 2015).
  76. World Health Organization (WHO). International Classification of Impairments, Disabilities and Handicaps; World Health Organization: Geneva, Switzerland, 1980. [Google Scholar]
  77. Barnes, C. Disabled People in Britain and Discrimination: A Case for Anti-Discrimination Legislation; Hurst & Co.: London, UK, 1994. [Google Scholar]
  78. Free TTS. FreeTTS 1.2—A Speech Synthesizer Written Entirely in the JavaTM Programming Language. 2015. Available online: http://freetts.sourceforge.net/docs/index.php (accessed on 26 July 2016).
  79. CMU Sphinx Letter to Phoneme Conversion in CMU Sphinx-4. Available online: http://cmusphinx.sourceforge.net/wiki/lettertophoneme (accessed on 20 May 2015).
  80. Java2s Download java speech api Free Java Code. Available online: http://www.java2s.com/Open-Source/Java_Free_Code/API_Google/Download_java_speech_api_Free_Java_Code.htm (accessed on 22 May 2015).
  81. iSpeech. Speech-Enable Your Java® Application. Available online: https://www.ispeech.org/developers/java (accessed on 22 May 2015).
  82. Neospeech VoiceText Text-To-Speech Engine SDK. Available online: http://www.neospeech.com/products (accessed on 23 May 2015).
  83. Acapela Elan Speech Cube. Available online: http://www.elantts.com/products/fp_speechcube.html (accessed on 26 May 2015).
  84. AT&T AT&T Natural Voices Crystal. Available online: http://at-t-natural-voices-crystal-v.software.informer.com/ (accessed on 24 May 2015).
  85. Cereproc CereVoice Engine Text-to-Speech SDK. Available online: https://www.cereproc.com/en/products/sdk (accessed on 25 May 2015).
  86. Sourceforge. eSpeak Text to Speech. Available online: http://espeak.sourceforge.net/ (accessed on 26 May 2015).
  87. Festvox University of Edinburgh Festival Speech Synthesis System. Available online: http://www.festvox.org/festival/ (accessed on 25 May 2015).
Figure 1. Combined design models of assistive technology assessment (ATA) and human activity assistive technology (HAAT).
Figure 1. Combined design models of assistive technology assessment (ATA) and human activity assistive technology (HAAT).
Computers 05 00018 g001
Figure 2. King Saud University (KSU) Smart and Context Aware University Campus Networking for PWD (USCAS-PWD).
Figure 2. King Saud University (KSU) Smart and Context Aware University Campus Networking for PWD (USCAS-PWD).
Computers 05 00018 g002
Figure 3. RFID–WiFi network inside RTV building.
Figure 3. RFID–WiFi network inside RTV building.
Computers 05 00018 g003
Figure 4. Usability scores of voice recognition software.
Figure 4. Usability scores of voice recognition software.
Computers 05 00018 g004
Figure 5. Sustainability and Maintainability scores of voice recognition software application.
Figure 5. Sustainability and Maintainability scores of voice recognition software application.
Computers 05 00018 g005
Figure 6. Login Page.
Figure 6. Login Page.
Computers 05 00018 g006
Figure 7. Customize Font Size First Page.
Figure 7. Customize Font Size First Page.
Computers 05 00018 g007
Figure 8. First Window application for USCAS-PWD.
Figure 8. First Window application for USCAS-PWD.
Computers 05 00018 g008
Figure 9. Features in Help option of the interface (Interface relevant to PWD).
Figure 9. Features in Help option of the interface (Interface relevant to PWD).
Computers 05 00018 g009
Figure 10. Main Search Example.
Figure 10. Main Search Example.
Computers 05 00018 g010
Figure 11. Communication and Event Setup.
Figure 11. Communication and Event Setup.
Computers 05 00018 g011
Figure 12. Sequence diagram of Auto Emergency Response and Help program.
Figure 12. Sequence diagram of Auto Emergency Response and Help program.
Computers 05 00018 g012
Figure 13. Light Smart Universal Interface, Smart Help, for iPhone.
Figure 13. Light Smart Universal Interface, Smart Help, for iPhone.
Computers 05 00018 g013
Figure 14. Light Smart Employee Interface, Search, for iPhone.
Figure 14. Light Smart Employee Interface, Search, for iPhone.
Computers 05 00018 g014
Figure 15. USCAS-PWD User Interface using keyword-search and Find Direction.
Figure 15. USCAS-PWD User Interface using keyword-search and Find Direction.
Computers 05 00018 g015
Figure 16. USCAS-PWD User Interface using Location-based search, event booking, and Find Direction on map.
Figure 16. USCAS-PWD User Interface using Location-based search, event booking, and Find Direction on map.
Computers 05 00018 g016
Figure 17. USCAS-PWD User Interface using User Profile-based search, using previous keyword search and sending email.
Figure 17. USCAS-PWD User Interface using User Profile-based search, using previous keyword search and sending email.
Computers 05 00018 g017
Figure 18. USCAS-PWD User Interface using User Profile-based search, using previous keyword (KW) specialist search and sending SMS.
Figure 18. USCAS-PWD User Interface using User Profile-based search, using previous keyword (KW) specialist search and sending SMS.
Computers 05 00018 g018
Figure 19. Personal Reminder—Event Algorithm to run on server.
Figure 19. Personal Reminder—Event Algorithm to run on server.
Computers 05 00018 g019
Figure 20. Personal Reminder—Noting Algorithm to run on server.
Figure 20. Personal Reminder—Noting Algorithm to run on server.
Computers 05 00018 g020
Table 1. Technologies essential in an Assistive System for different impairment conditions.
Table 1. Technologies essential in an Assistive System for different impairment conditions.
Type of InterfaceType of Assistive TechnologyHearing ImpairmentMotor ImpairmentSpeech ImpairmentVisual Impairment
Input to System1. Voice RecognitionXX X
2. Voice to TextXX
3. Keyboard and mouseX X
4. Gesture controlXXXX
Output from System5. Text To Voice XXX
6. Display ScreenXXX
7. Vibration or flashing LEDXXXX
Mobile Input and Output System8. Mobile platform such as laptop and SmartphoneXXXX
9. Tracking Location and Behavior using RFID sensors, and Wi-FiXXXX
10. Flexible adaptable interfaceXXXX
Table 2. Assessment criteria for commercially available applications for people with disabilities (PWDs).
Table 2. Assessment criteria for commercially available applications for people with disabilities (PWDs).
CriteriaDescription
ParentChild
1. Usability1.1 Understandability of how to use the programEasily understood application functionality and use
1.2 DocumentationComprehensive well-structured user document
1.3 InstallabilityStraightforward to install on a supported system
1.4 Comprehensiveness of the SolutionSupported features relevant to PWD
1.5 User RatingsUser rating about the software design and performance
2. Support and Maintenance2.1 Web basedInternet-based support for the user
2.2 Licensing costCost of the application or system
2.3 Quick to respond to the queryTime taken to respond to the user query
2.4 UpgradabilityHow often the system upgrades for better performance
2.5 Portability in terms of Operating SystemsHow compatible the system/app is with different operating systems
Table 3. Assessment results of 10 commercially available assistive applications. Note that scale is between 0 to 10, where 0 for no evidence and 10 for full evidence.
Table 3. Assessment results of 10 commercially available assistive applications. Note that scale is between 0 to 10, where 0 for no evidence and 10 for full evidence.
UsabilitySupport and Maintenance
Understandability for Using the ProgramDocumentationInstall-AbilityComprehensiveness of SolutionUser RatingsWeb BasedLicensing CostQuick to Respond to QueryUpgradabilityPortability Operating System
Blind Navigator60957010004
RAY Vision80939010704
Lend An Eye75978910974
Talking Camera7095265307
TapTapSee70838010087
Deaf and Hearing Impaired87959810974
SpeechTrans Hearing Impaired8595464567
uSound (Hearing Assistant)8595767884
Speech Assistant AAC87968910994
AAC Speech communicator98967810774
Table 4. Evaluation based on design consideration to support PWDs needs (Weightage: Yes, No, Partially Yes).
Table 4. Evaluation based on design consideration to support PWDs needs (Weightage: Yes, No, Partially Yes).
Input to SystemOutput of SystemMobile Input and Output System
Voice RecognitionVoice-to- TextKeyboard and MouseGesture ControlText-to- VoiceDisplay ScreenVibration or Flashing LEDMobile PlatformTracking Location and BehaviorFlexible Adaptable Interface
Blind Navigator [65]YesYesYesNoYesYesPartially YesYesPartially Yes (Only Tracking)No
RAY Vision [66]YesYesYesYesYesYesYesYesPartially Yes (Only Tracking)No
Lend An Eye [67]YesYesYesNoYesYesNoYesPartially Yes (Only Tracking)No
Talking Camera [68]NoNoYesNoYesYesNoYesNoNo
TapTapSee [69]NoNoYesNoYesYesNoYesNoNo
Deaf and Hearing Impaired [70]YesYesYesNoYesYesYesYesNoNo
SpeechTrans Hearing Impaired [71]YesYesYesNoYesYesNoYesNoNo
uSound (Hearing Assistant) [72]YesYesYesNoNoYesNoYesNoNo
Speech Assistant AAC [73]NoNoYesNoYesYesNoYesNoNo
AAC Speech communicator [74]NoNoYesNoYesYesNoYesNoNo
USCAS-PWDYes using VOCE [75]Yes using VOCE [75]YesNo (as this extra feature is optionalYes using VOCE [75]YesNo (as this extra feature is optional Yes (using Laptop and Smartphone)Yes (using RFID–WiFi Location tracking server, event and behavior algorithm, and movement direction algorithm)Yes (User can customize his/her interface in addition to selecting interface based on group profile)
Table 5. Different groups according to impairment conditions.
Table 5. Different groups according to impairment conditions.
RequirementCategoryG1G2G3G4G5G6G7G8G9G10G11
MustSeeXX
Must if cannot see fullyPartially seeXXXXXXXXX
Optional if can see and must if partially seeHearingXXXXXX
Must if cannot move hand and optional if can move handSpeakXXXX
Must if not speaking and optional if speakingMove handXXXX
Optional if can see and must if can partially seePartially hearXXXXXXXX
Table 6. Assessment criteria for voice recognition software.
Table 6. Assessment criteria for voice recognition software.
CriteriaDescriptionScore
ParentChildMinMax
3. Usability3.1 Ability to understand, learn, build, and install with right documentationEasily understood, straightforward to build and install, comprehensive user documentation (10 very flexible; 8 medium flexible; 5 less flexible; 0 non flexible)010
3.2 STT and TTSSpeech-to-text (STT), text-to-speech (TTS) (10 for both; 5 for only one)510
3.3 Web basedWeb based accessibility (5 for no, 10 for yes)510
3.4 DownloadsNumber of downloads in a specific period (month) 5 for (500 < downloads < 1000), and 8 (1000 ≤ downloads ≤ 2000), 10 for (downloads >2000)010
4. Sustainability and maintainability4.1 Identity, copyright and community, and accessibilityClear identity and ownership, community usage, and ability to download (10 Excellent; 8 very good; 6 good; 0–5 poor)010
4.2 Supportability, testability, and changeabilityDeveloper support, testability of source code, Aaility to modify and credit to developers (10 Excellent; 8 very good; 6 good; 0–5 poor)010
4.3 CostAdoption of appropriate license (10 free; 8 good price for all packages; 6 expensive price for all packages; 0–5 price per bandwidth)010
4.4 Portability operating system(s)Usable on multiple platforms (20 for many, 15 Windows, Android, and IOS, 10 for Windows and Android, and 5 for Windows only)020
4.5 Available language(s)English and Arabic support (5 English, 10 Arabic and English)010
4.6 Available voices(s)Number of voices available for speech-to-text and text-to-speech (5 for voices < 8; 8 for 8 < voices < 15; and 10 for voices ≥ 15)010
4.7 Programming language(s)Programming languages (10 Java , 5 C++, 15 for Java and C++)015
Total0125

Share and Cite

MDPI and ACS Style

Kbar, G.; Abidi, M.H.; Hammad Mian, S.; Al-Daraiseh, A.A.; Mansoor, W. A University-Based Smart and Context Aware Solution for People with Disabilities (USCAS-PWD). Computers 2016, 5, 18. https://doi.org/10.3390/computers5030018

AMA Style

Kbar G, Abidi MH, Hammad Mian S, Al-Daraiseh AA, Mansoor W. A University-Based Smart and Context Aware Solution for People with Disabilities (USCAS-PWD). Computers. 2016; 5(3):18. https://doi.org/10.3390/computers5030018

Chicago/Turabian Style

Kbar, Ghassan, Mustufa Haider Abidi, Syed Hammad Mian, Ahmad A. Al-Daraiseh, and Wathiq Mansoor. 2016. "A University-Based Smart and Context Aware Solution for People with Disabilities (USCAS-PWD)" Computers 5, no. 3: 18. https://doi.org/10.3390/computers5030018

APA Style

Kbar, G., Abidi, M. H., Hammad Mian, S., Al-Daraiseh, A. A., & Mansoor, W. (2016). A University-Based Smart and Context Aware Solution for People with Disabilities (USCAS-PWD). Computers, 5(3), 18. https://doi.org/10.3390/computers5030018

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop