Next Article in Journal
Vertical Balance of an Autonomous Two-Wheeled Single-Track Electric Vehicle
Previous Article in Journal
Effect of Oscillating Area on Generating Microbubbles from Hollow Ultrasonic Horn
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Cane for Assisting the Visually Impaired

by
Claudiu-Eugen Panazan
and
Eva-Henrietta Dulf
*
Department of Automation, Faculty of Automation and Computer Science, Technical University of Cluj-Napoca, Memorandumului Str. 28, 400014 Cluj-Napoca, Romania
*
Author to whom correspondence should be addressed.
Technologies 2024, 12(6), 75; https://doi.org/10.3390/technologies12060075
Submission received: 21 March 2024 / Revised: 21 May 2024 / Accepted: 23 May 2024 / Published: 27 May 2024
(This article belongs to the Section Assistive Technologies)

Abstract

:
Those with visual impairments, including complete blindness or partial sight loss, constitute a significant global population. According to estimates by the World Health Organization (WHO), there are at least 2.2 billion people worldwide who have near or distance vision disorders. Addressing their needs is crucial. Introducing a smart cane tailored for the blind can greatly improve their daily lives. This paper introduces a significant technical innovation, presenting a smart cane equipped with dual ultrasonic sensors for obstacle detection, catering to the visually impaired. The primary focus is on developing a versatile device capable of operating in diverse conditions, ensuring efficient obstacle alerts. The strategic placement of ultrasonic sensors facilitates the emission and measurement of high-frequency sound waves, calculating obstacle distances and assessing potential threats to the user. Addressing various obstacle types, two ultrasonic sensors handle overhead and ground-level barriers, ensuring precise warnings. With a detection range spanning 2 to 400 cm, the device provides timely information for user reaction. Dual alert methods, including vibrations and audio signals, offer flexibility to users, controlled through intuitive switches. Additionally, a Bluetooth-connected mobile app enhances functionality, activating audio alerts if the cane is misplaced or too distant. Cost-effective implementation enhances accessibility, supporting a broader user base. This innovative smart cane not only represents a technical achievement but also significantly improves the quality of life for visually impaired individuals, emphasizing the social impact of technology. The research underscores the importance of technological research in addressing societal challenges and highlights the need for solutions that positively impact vulnerable communities, shaping future directions in research and technological development.

1. Introduction

Aristotle underscored the vital role of vision as one of the five senses essential for comprehending the external world. Globally, over 2.2 billion individuals contend with vision disorders, a statistic highlighted by the World Health Organization [1]. Childhood visual impairments can exert lasting impacts on motor, verbal, and emotional development, often translating into academic hurdles. In adulthood, these impairments frequently correlate with decreased workforce engagement, diminished productivity, and an increased susceptibility to depression.
Visual impairment, characterized by limited visual fields or low acuity, presents challenges for daily activities. Traditional aids like canes, while economical, cannot detect obstacles. On the other hand, guide dogs, though effective, come with a considerable cost and a finite functional lifespan.
The smart cane is a contemporary, cost-effective solution endowed with obstacle-detection capabilities [1]. Guide dogs, while potent aids demand extensive training and financial considerations [1].
Visual acuity is often measured using the Snellen diagram, with blindness defined by a visual field of 20° or less horizontally with both eyes open [1]. This diagnostic tool gauges acuity based on the ability to identify letters from a specific distance.
The head of the World Health Organization (WHO), Margaret Chan, stated, “Almost everyone will experience a permanent or temporary disability at some point in their life” [2]. Due to this, We believe that humanity should be more understanding and accepting of people with disabilities. Each of us desires a life with fewer limitations and challenges, but for some individuals on Earth, life is not entirely satisfactory. In my opinion, everyone who thinks this way should begin to understand that, regardless of the difficulties we encounter, we should become more tolerant, especially towards those with different disabilities. According to the WHO, currently, over one billion people worldwide (1 in 7 or 15% of the population) experience some form of disability and 50% lack access to medical care.
Various types of disabilities individuals face include vision, hearing, mobility, memory, and communication, with depression being the most common (60%), followed by hearing and vision impairments. The WHO also provides data on individuals with visual impairments and those who are completely blind. Thus, global statistics indicate that there are currently 237 million people with moderate or severe visual impairments, with 55% being women.
Individuals less fortunate are those with congenital defects, as they lack any visual references and are completely blind from birth. This category also includes those who lose visual function in the first year of life. On the other hand, some people lose their vision later in life, retaining visual references of certain images memorized in the past. For this category of individuals, memory training and mental adaptation are relatively easier since they retain some residual vision [3].
As one can imagine, the lives of these individuals are challenging due to the difficulties posed by visual impairments. The visually impaired require more confidence and freedom than they currently have, as they are much more aware of their surroundings and the position of objects around them than one can imagine. For those who are blind or visually impaired, discovering things of interest and navigating complicated routes can be difficult undertakings, and there are currently insufficient infrastructures to make these activities simpler [4].
The primary objective of this research is to enhance the quality of life for individuals with visual impairments, addressing the societal marginalization they face and the associated psychological challenges. This research evaluates existing ideas to improve and integrate previous solutions into a smart cane equipped with ultrasonic sensors. The ultimate goal is to create a valuable tool for individuals with visual impairments while keeping production costs low. The research involves testing the assembled cane and defining optimal usage conditions for optimal results.
The secondary objectives complement the primary goal of creating an intelligent cane for individuals with visual disabilities. The journey starts with an in-depth exploration of specialized literature on visual impairments and the evolution of traditional and smart canes. Clear system requirements and an adaptable architecture are crucial, ensuring accessibility beyond the engineering realm.
A meticulous component analysis guides the acquisition of necessary elements, considering technical specifications for optimal selection. The subsequent phase involves designing and implementing an obstacle-detection solution using advanced sensor technology.
Testing and analysis follow, comparing results with existing solutions to identify strengths and weaknesses. This evaluation informs iterative improvements, fostering innovation. Key principles emphasize the clarity and adaptability of system elements, detailed technical analysis for component selection, and a dynamic design for future enhancements.
In essence, these objectives shape a user-centric solution at the intersection of technology and social impact, poised for adaptability and continual improvement.
This research seeks to enhance the quality of life for individuals with visual impairments, placing a strong emphasis on affordability to ensure widespread accessibility. To align with this overarching goal and the defined secondary objectives, the research specifications encompass various key aspects.
The cost is a crucial consideration, with the total not exceeding 60$, aimed at ensuring both affordability and competitiveness in the market. Additionally, an adjustable cane size is proposed, providing a predetermined adjustable size for user comfort.
The sensor range is specified to detect objects within a range of 5 cm to 3 m with high accuracy, contributing to effective obstacle detection. The choice of durable cane material, coupled with provisions for vibration motors in the handle, adds a tactile dimension to the user experience.
The system’s user-friendliness is prioritized, intending to create an intuitive system that allows users to seamlessly switch between indoor and outdoor modes. Finally, the incorporation of a rechargeable battery is proposed to power the components, aligning with modern energy-efficient practices. These specifications collectively aim to create an accessible, efficient, and user-centric solution for individuals with visual impairments.

2. Literature Review

The cane is the most common tool used by the visually impaired for mobility. Materials used include aluminum, graphite, fiberglass, or carbon fiber. Canes can be fixed, foldable, or adjustable in length. Studies suggest longer canes provide greater mobility and safety. Over time, various types of assisting canes have been invented. History attributes the invention of the long white cane to the English photographer James Biggs, who, after losing his sight in an accident, painted his walking cane white to be more visible to drivers around his home [5].
In the realm of canes designed to assist individuals with visual impairments, various types serve distinct purposes:
The Long Cane is instrumental in detecting possible objects at a distance, and its usage allows freedom of movement in different directions without strict rules.
The Identification Cane, also known as “ID cane” or “symbol cane”, serves the purpose of alerting others to the user’s visual impairment. Similar to the Long Cane, it provides flexibility in movement without stringent usage rules.
The Guiding Cane, slightly shorter than the Long Cane, is employed for detecting low-lying objects like curbs. It can also be used diagonally for added protection.
Designed to aid individuals with visual, balance, or stability issues, the Support Cane enhances physical stability, providing valuable assistance where needed.
The Green Cane, utilized in countries like Argentina, serves as an indicator of visual impairment. Notably, the white color of this cane denotes complete blindness. These diverse types of canes cater to specific needs, offering tailored support for individuals with visual impairments.
The primary cane used by the visually impaired is the long white cane, instrumental in enhancing mobility but with limitations in identifying obstacles, especially those at various heights or distances greater than the cane’s length, roughly one meter when factoring in arm extension. Technological advancements have led to the development of smart canes incorporating sensors, particularly ultrasonic ones, to overcome these limitations [6].
Ultrasonic sensors operate by emitting and receiving acoustic waves. They offer millimeter precision in detecting obstacles and are known for their reliability under various conditions, including dust, rain, or dirt. These sensors calculate distances by leveraging the time of wave propagation and reception, using the speed of sound (≈340 [m/s]).
Smart canes with ultrasonic sensors can detect obstacles well in advance, up to a distance of 4 [m] [7]. This represents a substantial improvement over traditional white canes, as they provide not only early obstacle detection but also additional features like acoustic alerts or vibrations. These features empower visually impaired individuals to react promptly and navigate their surroundings more effectively [8].
Arduino is commonly employed in smart cane projects due to its accessibility and cost-effectiveness. These projects, often initiated by technology enthusiasts, serve as valuable prototypes, lacking refined designs. The fundamental principle involves obstacle detection using an ultrasonic sensor, triggering alerts based on a predefined distance limit. Hardware components are simple, featuring an Arduino development board and an HC-SR04 ultrasonic sensor. The prototype’s functionality is straightforward: detecting obstacles within a pre-set distance prompts the microcontroller to activate a buzzer for user alert. More advanced smart cane models, such as “WeWALK” and “Phoenix”, exist in the field, offering intricate functionalities beyond basic prototypes.
While the advanced features offered by smart canes like WeWALK and Phoenix significantly enhance the mobility of visually impaired individuals, it’s essential to acknowledge certain disadvantages. One notable drawback is the relatively high cost associated with these sophisticated technologies. The WeWALK smart cane, for instance, is priced at around 500$, and the Phoenix system may range from 800$ to 1500$. The incorporation of cutting-edge functionalities, such as obstacle detection, GPS navigation, and AI assistance, contributes to a premium price point, potentially limiting accessibility for individuals with lower financial resources. Additionally, the complexity of these devices may pose a learning curve for some users. Balancing the advanced functionalities with affordability and user-friendly interfaces remains a challenge in the development and adoption of smart cane technologies.
Sensor technology, computer image processing technology, computer vision technology, and voice recognition technology are the three phases of development that positioning and navigation technology has gone through since it was first presented in the field of blind guides in the 1960s [9]. The least that any guiding device should be able to do is measure the separation between obstacles, potholes and stairs, and low-lying obstructions [10]. It should also be able to identify both static and moving objects, operate continuously day and night, and deliver clear and trustworthy information about the surroundings in a matter of seconds [9]. Distinct guidance systems installed on various guiding apparatuses are made up of various assistive guidance technologies. One innovative solution for the visually handicapped is the guide helmet [11], which uses cutting-edge sensor technology like LiDAR to identify obstacles and provide real-time spatial awareness. Users are provided with an intuitive grasp of their surroundings through the processing and transformation of the acquired data into haptic or audible warnings. Furthermore, certain smart helmets come equipped with GPS navigation systems, which provide easy movement and orientation. These helmets are a potential way to increase the freedom and quality of life for those who are visually impaired because they combine cutting-edge sensor technology with intuitive interfaces.
Dernayka et al. (2021) [12] suggested the Tom Pouce III smart blind cane. The cane included two separate sensors that operated separately: an infrared sensor and a laser detector. If you use the infrared sensor alone, the laser detector can identify objects up to 12 m away, but it can only detect objects up to 6 m away. Compared to the Teletact II laser blind cane, the Tom Pouce III is less vulnerable to interference from natural light and is still capable of detecting obstacles in bright sunshine. However, the visually handicapped must walk slowly with their blind cane in order to avoid obstacles that are 4–50 cm from the front at their typical walking pace. This is due to the blind cane’s detecting reaction rate.
Another smart and portable device, an environment-aware smart blind cane system called INSPEX was suggested by J. Foucault et al. (2019) [13]. It is made up of ultrasound, ultra-wideband radar, depth camera TOF sensor, and Lidar. The smart cane’s integrated Uwiband radar system and ultrasonic sensor can significantly enhance its capacity for environmental anti-interference and ultra-short-range detection. The system’s measurement range is up to 10 m, and it can identify obstructions surrounding the blind cane with effectiveness.
This research centers on creating an experimental smart cane designed to significantly improve the independence and mobility of individuals with visual impairments. The model, employing ultrasonic sensors, excels in addressing the challenges related to mobility and orientation. A standout feature is its impressive 98% accuracy in obstacle detection, setting it apart from other smart canes on the market. The system effectively alerts users in advance, minimizing the risk of collisions.
The experimental model is both accurate and cost-effective, rendering it an optimal choice for visually impaired individuals. The combination of affordability and exceptional accuracy positions this technology as a significant advancement in assistive technology.

3. Materials and Methods

3.1. Solution Analysis and Design

The environment is undergoing a rapid evolution in response to technological advances and the development of intelligent solutions, which are having a positive impact on people’s lives. In light of the desire for independence, particularly among those with visual impairments, the subject of the research was the development of an experimental smart cane with ultrasonic sensors. The objective of this tool is to assist visually impaired individuals in navigating safely towards their chosen destinations.
Establishing the foundational requirements for the system is crucial in addressing mobility challenges for individuals with visual impairments. The concise yet comprehensive system requirements aim to:
-
Prioritize user-friendly design and comfort.
-
Enable flexible deactivation of intelligent features and the entire system.
-
Ensure user comfort with vibrations and sounds.
-
Enable Bluetooth connectivity for app-controlled acoustic signals.
-
Allow seamless mode switching for indoor and outdoor use.
The research is organized into sequential sub-chapters for efficient development, as illustrated in Figure 1. Subsequent chapters will delve into the details of each stage.

3.2. Analysis of the Most Important Components Used

The goal is to develop a competitive experimental model offering advanced support for blind individuals. By integrating smart components and leveraging existing ideas, the model emphasizes enhanced obstacle detection, a feature absent in traditional white canes. The first step involves a detailed analysis of selected components to drive the development process.

3.2.1. Development Board

The crucial component in our experimental model is the ESP-WROOM-32 development board, a vital element for its successful implementation. Previous projects often relied on development boards like Arduino Uno or Arduino Nano, owing to their ease of programming and cost-effectiveness. However, the limitations posed by the size of the Arduino Uno made it unsuitable for our envisioned robust system designed to attach to a blind person’s cane. Considering an alternative, the smaller Arduino Nano was contemplated, but its lack of Wi-Fi and Bluetooth capabilities rendered it insufficient for our purposes.
To overcome these constraints and enhance the functionality of our experimental model, the ESP-WROOM-32 DevKit emerged as the preferred choice. One of the key advantages that tipped the scales in its favor is its incorporation of both Wi-Fi and Bluetooth technologies, addressing the connectivity needs essential for our smart cane research. Furthermore, the ESP-WROOM-32 offers a larger memory capacity and a greater number of programmable pins compared to both the Arduino Uno and Arduino Nano. This makes it a more versatile and powerful option, aligning better with the comprehensive requirements of our innovative smart cane. The programming language used by the development board can be extended through C++ libraries, like on the Arduino part. Arduino is open source, both in terms of hardware and software [14]. For a clearer visual representation of these development boards, refer to Figure 2 and Figure 3 showcasing the Arduino Uno and Arduino Nano, respectively.
The selected development board for this research is, in conclusion, the ESP-WROOM-32 DevKit. The primary advantage of this board is its incorporation of the aforementioned technologies, expanded memory, and a greater number of programmable pins, rendering it more suitable and powerful overall. A comparison of the characteristics of the chosen development board with those of the Arduino Uno and Arduino Nano is presented in Table 1.
The ESP-WROOM-32 operates at a supply voltage of 3.3 [V], and its operating temperature ranges from −40 [°C] to 85 [°C]. In addition to Bluetooth and Wi-Fi, it also incorporates Bluetooth Low Energy (BLE), enabling users to connect mobile phones or transmit BLE messages. The model of the board used for the research is presented in Figure 4, which is a powerful and reliable development board from all analytical perspectives of this experimental model. The software that was developed is subsequently implemented on this component, constituting the heart of this research.

3.2.2. Ultrasonic Sensors

The two HC-SR04 ultrasonic sensors used in implementing the solution are crucial components. These sensors are among the most widely used for measuring the distance to a target object, offering a favorable quality-to-price ratio. Since the aim for the experimental model was to be affordable for everyone interested, these ultrasonic sensors were used, which have an operating range starting from 2 cm and reaching up to 4 m. The sensor module used is depicted in Figure 5.
These sensors aim to detect potential obstacles in the traveler’s surroundings. Ultrasonic sensors have predefined threshold limits: they operate from 3 [cm] to 3.5 [m] in most situations. The used HC-SR04 ultrasonic sensor, presented in Figure 5, has four pins in its composition: the VCC power pin, GND ground pin, Trigger, and Echo. [16] The operating principle is as follows: the Trigger pin, an input pin, is held “high” for a period of 10 [μs] to initialize the measurement by sending ultrasonic waves, and the Echo pin, an output pin, is held “high” for a time equal to the round-trip time of the ultrasonic wave returning to the sensor [17].
Generally, ultrasonic sensors emit waves at a frequency higher than 20 [kHz], but the frequency of the sensors chosen for use in this research is 40 [kHz]. The distance to the reference object can be calculated, taking into account the speed of sound, which has an approximate value of 343 [m/s] or 0.0343 [cm/μs], and the time taken by the ultrasonic wave emitted by the sensor to travel through the air. Thus, using the general formula for speed, which is equal to the ratio of distance to time, the calculation formula can be expressed and used to determine the distance at which objects detected by the sensors are located. Therefore, this formula is as follows [17]:
Distance   [ cm ] = Sound speed [ cm / μ s ] · Time [ μ s ] 2
As mentioned earlier, the development board has a voltage drop on the output pins of only 3.3 [V], which is insufficient for the ultrasonic sensors used, requiring a supply voltage of 5 [V]. The main characteristics of these sensors, extracted from the datasheet, are presented in Table 2.

3.2.3. Other Components Used

The haptic feedback system incorporates compact button-type vibration motors, which, weighing in at around 0.9 [g] and measuring 10 [mm] × 3.2 [mm], operate at a voltage of 3 [V] to provide tactile alerts upon detecting obstacles. In the auditory domain, adjustable passive buzzers are employed for sound notifications, offering flexibility in frequency modifications. Powering the entire system is a rechargeable 9 [V] voltage accumulator, utilizing Li-ion technology with a capacity of 5400 [mWh]/600 [mAh], ensuring both user convenience and eco-friendly recharging through USB. User control is facilitated by vertical switches, allowing seamless toggling of audio and haptic functions. Complementing these features is a radio-frequency remote control, which, operating at 12 [V] and 315 [MHz], possesses a range of 50–80 [m], aiding in locating the cane through emitted sound signals upon button activation.

3.3. System Arhitecture

In the construction of a more intricate system, the system architecture assumes a pivotal role. It serves as a framework upon which the system is built, providing information about how its parts interact with one another. The system’s capacity to fulfil the imposed requirements while maintaining reliability, scalability and efficiency is contingent upon its architecture. For the purposes of this experimental model, the block diagram representing the system architecture is presented in Figure 6. A description of each component and its connectivity will be presented in the following sections.
The most complex component of this system is the ESP-WROOM-32 development board. For the development of this experimental model, the chosen kit was DevKit model with 30 pins and 2 dual-core processors, as presented in the previous subsection. Out of these, 25 pins are programmable, 18 are ADC (Analog-Digital Converter) pins, 10 are GPIO pins with capacitive sensitivity (aiding in the implementation of functions like touch buttons, and touchpads), 2 are DAC (Digital-Analog Converter) pins, 3 are UART (Universal Asynchronous Receiver-Transmitter) pins, 2 are I2C (Inter-Integrated Circuit) pins, 16 are connected to PWM (Pulse Width Modulation) channels, allowing control over light intensity, motor speed, or similar applications, 2 are I2S (Inter-IC Sound) pins, and 3 are SPI (Serial Peripheral Interface) pins. In addition to these pins, there are also those related to VIN and 3.3 [V] power supply, as well as 2 GND ground pins. All this information was extracted from the development board’s datasheet.
For this experimental model, 16 pins out of the total 30 were utilized, leaving the remaining 14 unused, available for future researchs or further development of the current one. In Figure 7 below, the used pins and their meanings are represented.
In Figure 8, the complete hardware diagram of the entire system, namely the smart cane for the visually impaired with ultrasonic sensors, is presented. This diagram was created using the EasyEDA (Standard) tool. It is a free and user-friendly program that can run both online and on a personal computer after being downloaded. The main purpose of this program is to create and simulate electrical schematics and printed circuit boards (PCBs).

3.4. Working Principle

Figure 9 illustrates the functioning of the entire system through a flowchart. The central element is the microcontroller, which has not been included in it. This element controls all the components in the system.
When the power switch is activated, the system initiates its operational functions, including the ultrasonic sensors and mobile application. Once the system has been initialized, the sensors commence scanning for obstacles. Upon the detection of an obstacle, the sensor data is transmitted to the microcontroller, where it is used to calculate the distance to the obstacle. The system mode (indoor or outdoor) is set by the system switch. If the obstacle is within a predefined distance threshold and both the haptic and sound alert switches are activated, an alert is triggered. In the event that only one alert switch is activated, the corresponding alert will be triggered.
On the mobile application branch, after its execution, the next step is to establish a Bluetooth connection between the application and the development board. Upon successful connection, the user can change the buzzer state using buttons in the application, enabling the start and stop of sound alerts through the buzzer to locate the cane in case of loss. This is possible even if the audio alert switch is in the OFF state.
If the power switch is in the OFF state, the system will not function, and the cane will act as a traditional support cane for visually impaired individuals.

3.5. Software Implementation

In the implementation of the final algorithm, the microcontroller code was developed using the Arduino IDE. Pin assignments for various components were defined using the “#define” directive in the C/C++ programming language. This approach allowed the use of symbolic names for constant values, improving code readability. Pins for ultrasonic sensors, vibration motors, audio components, and switches were identified using this method.
The next step involved setting threshold distances for the ultrasonic sensors based on the system’s operating mode (indoor or outdoor). Three threshold distances were defined for each mode. Global variables, such as sound speed and various switch states, were initialized in the code.
The core functionality of obstacle detection and alerting was implemented in the “loop()” function. The distances from ultrasonic sensors were continuously monitored, and alerts were triggered based on predefined threshold values. The alert system included both haptic (vibrations) and audio (buzzer and headphones) signals. The code allowed users to customize alert preferences through switches, providing flexibility in system behavior.
Additionally, a Bluetooth-enabled mobile application was developed using MIT App Inventor 2, an online platform for mobile app design [18]. Initially developed by Google and currently maintained by the Massachusetts Institute of Technology (MIT), MIT App Inventor 2 aims to assist users who may not be proficient in traditional programming by employing a block-based programming approach. The platform provides pre-built components like buttons and labels, making it efficient for building mobile applications. Utilizing a “drag and drop” method, the application interface, depicted in Figure 10, was easily constructed. The visual programming approach enhances user-friendliness and accessibility, aligning to facilitate mobile app development for individuals with limited programming expertise.

3.6. Experimental Setup

To integrate all the components into a functional and reliable system, their placement on the cane is based on their size and function. To ensure that users can feel the signals through vibrations, the position of the vibration motors was on the handle of the cane. Below the handle was created a space for the switches related to the cane’s functionalities.
Recognizing the need for users to easily activate or deactivate the entire system, the switch responsible for this is at the top of the support. Simultaneously, for easier memorization of switch functionalities, the system mode switch is on the left side of the support, with the switches for audio and haptic alerts on the right.
The positioning of the vibration motors was influenced by the hand’s placement on the handle. There were used three motors because a single one was insufficient to perceive vibrations effectively. After testing with one and two motors, the conclusion was that three motors were adequate for my experimental model, ensuring users could sense vibrations when encountering obstacles. Figure 11 illustrates these components, showing two motors on one side of the handle and the third on the opposite side.
To explain the placement of other essential components, Figure 12 illustrates the assembly of ultrasonic sensors, the buzzer, and the Jack plug on the cane. The two sensors were placed at the base and approximately in the middle of the cane, angled at about 45 degrees to detect obstacles perpendicular to the user’s path.
After assembling all the components on the cane and testing its functionalities, the entire circuit was covered with a plastic cap, later concealed with a black cover for a more aesthetically pleasing effect. The assembled circuit can be observed in its entirety in Figure 13.
The cane has a compact appearance, uncluttered by wires and various components, making it easy for me to handle. The cane is one meter long but its height can be adjusted as well as the position of the sensor on it. Now, the sensors are positioned 10 cm and 60 cm from the ground, as a stick has been created for medium height. The cane weighs approximately 600–650 g with all components attached, making it easy and simple to handle. In terms of the water resistance of the circuits, they are encased in a plastic case which increases their resistance to moisture and wind. The casing covering the basic circuit is located under the cane’s handle. As in any field, improvements can be made, and thus, this research represents only my experimental model of a cane with ultrasonic sensors designed for individuals with visual impairments.

4. Results and Discussions

The testing phase stands as a pivotal juncture in the iterative process of refining an experimental model, playing a critical role in elucidating the model’s practical performance in real-world scenarios. This phase is strategically designed to furnish an impartial and comprehensive evaluation of the system’s efficacy, providing an unbiased estimate of its predictive capabilities. By conducting these tests under genuine, non-laboratory conditions, the scope is to capture the model’s performance in an authentic environment, free from artificial enhancements that might influence measurement quality.
The primary objective of this rigorous testing is to ensure the authenticity of the model’s performance, steering clear of any distortions that could arise from manual adjustments made to validation data. This meticulous approach not only yields a genuine measure of the model’s real-world functionality but also contributes significantly to the holistic evaluation of its effectiveness and potential impact.
In the specific validation process of the experimental model illustrated in Figure 13, a series of comprehensive tests were conducted in a real-world environment. This environment intentionally lacked the controlled conditions of a laboratory setting and eschewed any features that could artificially amplify measurement quality. The testing protocol involved the collection of five measurements for six distinct distances from an obstacle, with the obstacle represented by a sufficiently large wall for identification purposes. The wealth of data gleaned from these tests, recorded with precision, is systematically documented in the detailed tabulation presented in Table 3.
As shown in the table, for each actual distance measured with the ruler, data was collected five times to calculate an average distance obtained from the sensors. This approach is taken because inexpensive sensors may introduce errors. These errors were calculated as an average of the errors obtained in each measurement for each distance interval. The average error was then expressed as a percentage using Equation (2), where “ D r ” is the actual distance measured with the ruler, and “ D m ” represents the distance obtained from the sensor.
ϵ = D r D m D r · 100 [ % ]
Additionally, a comparison between the actual and measured distances was conducted. As shown in Figure 14, the differences are not significant, indicating the reliability of the system.
The validation phase emerges as a pivotal juncture in the iterative refinement process of the experimental model, assuming a critical role as a benchmark for an exhaustive evaluation of its performance in comparison to an array of existing models. The relentless pursuit of precision and accuracy during this phase becomes indispensable, wielding a profound influence on the overarching dependability, reliability, and robustness of the system.
Upon delving into a comprehensive analysis of the outcomes stemming from the proposed experimental model, a resounding affirmation of its stature as a beacon of reliability within the expansive domain of assistive devices for individuals with visual impairments becomes evident. The system not only showcases an extraordinary average accuracy of 98%, a remarkable feat in itself but also distinguishes itself as a pertinent solution when subjected to an intricately objective comparative assessment against a diverse spectrum of models. This evaluative scrutiny unfolds meticulously in the illuminating tableau of Table 4, providing a nuanced and insightful perspective on the experimental model’s prowess and its distinctive standing among contemporary assistive technologies.
From Table 3, it can be observed that the maximum accuracy, 100%, was achieved when the obstacle was placed at a distance of 25 [cm], while the lowest accuracy, 96.8%, was recorded at a distance of 100 [cm]. In Figure 15, the evolution of the system’s accuracy is graphically depicted based on the real distance at which the object was positioned.
Concurrently, a perusal of the graphical representation delineated in Figure 16 unveils a succinct yet illuminating insight into the system’s error dynamics. This graphical elucidation harmonizes seamlessly with the accuracy metrics of the model, meticulously calculated through the nuanced lens of Equation (2). As the graph unfolds its narrative, a striking revelation comes to light—the system consistently maintains an average error rate below 2%. This empirical revelation serves as a testament to the system’s heightened relevance in the expansive landscape of existing solutions, solidifying its stature as a beacon of precision and efficacy.
Furthermore, in an endeavor to ascertain the system’s efficacy across diverse real-world scenarios, a comprehensive array of tests was conducted to evaluate its adeptness in detecting obstacles crafted from an assortment of materials. The sensors showcased commendable proficiency in discerning various surfaces and objects, including cabinets, stairs, and mirrors. Nevertheless, a nuanced challenge surfaced when the sensors grappled with detecting objects positioned on coarse surfaces, wherein the undulating waves tended to be absorbed by these particular materials. The encapsulated outcomes of these discerning tests are meticulously documented in Table 5, providing a nuanced exposition of the system’s material-specific detection prowess.
As shown in Table 5, the sensors detect different obstacles very well, except for those with a rough surface. This type of surface was detected by the sensors six times out of ten trials. This is because the transmitted waves can be distributed at different angles to the angle at which the sensor expects to receive a wave. This can also happen when the sensor sends waves from one angle and they are reflected by obstacles in other directions. This is why sensor calibration on the stick has played a very important role in obstacle detection. These types of measurements were made by visually impaired people to be most relevant to our type of system. Other experiments have been carried out by the blind, such as taking an obstacle course from one point to another, and the results have been good.
The final step in the validation process was to analyze the accuracy of the ultrasonic sensors and validate their use against other types of sensors. So, Table 6 shows the comparison between the characteristics of the sensor types that could have been used in the research.
Due to the research’s objective of achieving low implementation costs, the choice of sensors came down to infrared and ultrasonic options. After conducting numerous tests and analyzing the characteristics of these sensor types, ultrasonic sensors were selected. This decision was based on their larger distance range and broader bandwidth. Another significant advantage is that ultrasonic sensors are minimally affected by atmospheric conditions in the surrounding environment, as the speed of sound is proportional to temperature. This ensures more reliable measurements under various temperature conditions. Considering all these aspects, the validation of the proposed experimental model is conducted objectively, making the proposed smart cane for visually impaired individuals a reliable and robust solution.

5. Conclusions and Future Steps

Within the research, the goal was to create an experimental model of a smart cane to assist people with visual impairments. We chose this theme because the aim was to develop a system that enhances the independence and mobility of these individuals. The experimental model was created using ultrasonic sensors. As mentioned in the first chapter of this work, one of the most crucial sensory senses is vision, as it must replace smell and hearing when these are absent. Moreover, it constitutes the first sense that detects the path one takes from one place to another, an activity everyone performs daily. Therefore, the main challenge faced by visually impaired individuals is mobility and orientation. To improve these factors, the white cane for the blind was introduced in 1940, and it now serves as the primary assistance system for the blind.
Over time and with technological advancements, various types of canes have emerged, evolving in terms of length, material, and the issues they address. The longer the cane, the faster it detects obstacles, providing the user with more reaction time. Traditional canes can detect obstacles within a maximum distance of one meter, depending on the user’s height and the cane’s tilt angle. However, a major issue is the obstacle localization time, as users are only alerted after making contact, resulting in a longer detection time. These issues have been addressed in this research by implementing a smart cane with ultrasonic sensors, which are used to detect obstacles. These obstacles can be low or hanging, requiring the use of two sensors, one for each type of obstacle. The operating range of the sensors is between 2 [cm] and 400 [cm]. To create a reliable smart cane in various situations, obstacle warnings can be delivered through both haptic and audio alerts played through headphones and a buzzer. Users can start and stop these alerts using two vertical switches. Another switch allows users to choose the system mode, as it contains two modes of operation: indoors and outdoors. Another intelligent functionality implemented was a mobile application connected to the development board via Bluetooth, enabling the triggering of audio signals from a buzzer in case of cane loss. A backup solution was also implemented using a radiofrequency remote control with the same purpose. Following all the proposed steps in creating this system, an experimental model was developed to alert the user in advance if there is an obstacle in their walking path, preventing collisions. The closer the distance, the more frequent the signals to warn the user of approaching obstacles. In this way, visually impaired individuals can avoid collisions using a rechargeable and robust experimental model.
The biggest advantage of this system is the very low cost, which is one of the main objectives achieved. The total cost was about 60 $ which makes it a very strong competitor on the market compared to other systems such as: the WeWALK device which has a cost of 600 $ with some extra features or the Phoenix device with a price of about 128 $ , which is double the proposed system. The 60 $ price includes the time and work developed, which would be a big plus for it.
For testing and validating the model, real-world scenarios were created, and the obtained results were very good. The model has an accuracy of 98%, which, when compared to other models of assistance systems for visually impaired people, is exceptionally good. The system can recognize objects made of various materials, providing robustness. Another important aspect is the cost. This experimental model was created at a very low cost compared to existing models on the market, as studied earlier in the work. Therefore, the model manages to be a combination of important functionalities of the studied models but at a much lower cost.
Any field can undergo improvements with the development of technology and automation. The smart cane created in this research is just an experimental model and may contain improvements, either in terms of functionality or appearance. One potential improvement could involve introducing a GPS (Global Positioning System) module to track the cane’s location in case of loss or to provide assistance to the user. Another aspect could be the incorporation of a fall sensor to monitor the user’s condition during system usage and cover the possibility of the user hitting an obstacle and falling.
One idea that could improve this system and bring it to life would be to include deep learning or machine learning for the results obtained from the ultrasound sensors, but also in the event of obstacle detection with the camera. Some methods that could be relevant for the application of deep learning are also presented in [26].
Another idea for improving this experimental model could be replacing the wired headphones with a Jack socket for audio alerts in Bluetooth-connected headphones. Unlike the previous ideas that required additional components, this development idea does not need extra components as the development board incorporates the Bluetooth technology module. This aspect would make the cane more in line with current technology trends, where most users prefer Bluetooth-connected headphones over wired ones.
On the other hand, changing the appearance of the cane could be another enhancement to this research. Creating a cane with fewer but more powerful vibration motors would reduce the number of components used in the system and improve the cane’s appearance. Additionally, further development could involve integrating ultrasonic sensors with a camera for image processing and obstacle detection and filtering. If this feature is desired, cost considerations should be taken into account, as it was a deciding factor in the system created in this research.
In conclusion, after testing and validating the experimental model, it can be summarized that a reliable and robust smart cane for visually impaired individuals has been created, with high accuracy and low cost, thereby increasing the user’s reaction time. The primary goal of improving the quality of life for people with visual impairments has been achieved, along with the secondary objectives of the research.

Author Contributions

Conceptualization, C.-E.P.; methodology, C.-E.P.; software, C.-E.P.; validation, E.-H.D.; formal analysis, E.-H.D.; investigation, C.-E.P. and E.-H.D.; writing—original draft preparation, C.-E.P.; writing—review and editing, E.-H.D.; supervision, E.-H.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Available on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Farooq, M.S.; Shafi, I.; Khan, H.; Díez, I.D.L.T.; Breñosa, J.; Espinosa, J.C.M.; Ashraf, I. IoT Enabled Intelligent Stick for Visually Impaired People for Obstacle Recognition. Sensors 2022, 22, 8914. [Google Scholar] [CrossRef] [PubMed]
  2. Silberner, J. Nearly 1 in 7 People on Earth Is Disabled, Survey Finds. National Public Radio, 9 June 2011. Available online: https://www.npr.org/sections/health-shots/2011/06/09/137084239/nearly-1-in-7-people-on-earth-are-disabled-survey-finds(accessed on 22 May 2024).
  3. Lessard, N.; Paré, M.; Lepore, F.; Lassonde, M. Early-blind human subjects localize sound sources better than sighted subjects. Nature 1998, 395, 278–280. [Google Scholar] [CrossRef] [PubMed]
  4. Messaoudi, M.D.; Menelas, B.-A.J.; Mcheick, H. Autonomous Smart White Cane Navigation System for Indoor Usage. Technologies 2020, 8, 37. [Google Scholar] [CrossRef]
  5. Strong, P. The History of the White Cane. PathstoLiteracy. 2009. Available online: https://in-sight.org/wp-content/uploads/WC-History.pdf (accessed on 22 May 2024).
  6. Kim, S.Y.; Cho, K. Usability and design guidelines of smart canes for users with visual impairments. Int. J. Des. 2013, 7, 99–110. [Google Scholar]
  7. Morgan, E.J. HC-SR04 Ultrasonic Sensor. November 2014. Available online: https://docplayer.net/23967545-Hc-sr04-ultrasonic-sensor-elijah-j-morgan-nov.html (accessed on 22 May 2024).
  8. Zhmud, V.A.; Kondratiev, N.O.; Kuznetsov, K.A.; Trubin, V.G.; Dimitrov, L.V. Application of ultrasonic sensor for measuringdistances in robotics. J. Phys. Conf. Ser. 2018, 1015, 032189. [Google Scholar] [CrossRef]
  9. Elmannai, W.; Elleithy, K. Sensor-based assistive devices for visually-impaired people: Current status, challenges, and future directions. Sensors 2017, 17, 565. [Google Scholar] [CrossRef] [PubMed]
  10. Benjamin, J.J. A review of the veterans administration. Bull. Prosthet. Res. 1968, 9, 63. [Google Scholar]
  11. Kassem, K.; Caramazza, P.; Mitchell, K.J.; Miller, M.; Emadi, A.; Faccio, D. Real-time scene monitoring with haptic feedback. Sensors 2022, 22, 7136. [Google Scholar] [CrossRef] [PubMed]
  12. Dernayka, A.; Amorim, M.A.; Leroux, R.; Bogaert, L.; Farcy, R. Tom Pouce III, an electronic white cane for blind people: Ability to detect obstacles and mobility performances. Sensors 2021, 21, 6854. [Google Scholar] [CrossRef] [PubMed]
  13. Foucault, J.; Lesecq, S.; Dudnik, G.; Correvon, M.; O’Keeffffe, R.; Palma, V.D.; Passoni, M.; Quaglia, F.; Ouvry, L.; Buckley, S.; et al. INSPEX: Optimize range sensors for environment perception as a portable system. Sensors 2019, 19, 4350. [Google Scholar] [CrossRef] [PubMed]
  14. Lin, S.S.; Zhu, K.Y.; Zhang, X.H.; Liu, Y.C.; Wang, C.Y. Development of a Microservice-Based Storm Sewer Simulation System with IoT Devices for Early Warning in Urban Areas. Smart Cities 2023, 6, 3411–3426. [Google Scholar] [CrossRef]
  15. Kondaveeti, H.K.; Kumaravelu, N.K.; Vanambathina, S.D.; Mathe, S.E.; Vappangi, S. A systematic literature review on prototyping with Arduino: Applications, challenges, advantages, and limitations. Comput. Sci. Rev. 2021, 40, 100364. [Google Scholar] [CrossRef]
  16. Meliones, A.; Filios, C.; Llorente, J. Reliable Ultrasonic Obstacle Recognition for Outdoor Blind Navigation. Technologies 2022, 10, 54. [Google Scholar] [CrossRef]
  17. Rahman, M.M.; Islam, M.M.; Ahmmed, S.; Khan, S.A. Obstacle and fall detection to guide the visually impaired people with real time monitoring. SN Comput. Sci. 2020, 1, 219. [Google Scholar]
  18. Wolber, D. App inventor and real-world motivation. In Proceedings of the 42nd ACM Technical Symposium on Computer Science Education, Dallas, TX, USA, 9–12 March 2011; pp. 601–606. [Google Scholar]
  19. Sadi, M.S.; Mahmud, S.; Kamal, M.M.; Bayazid, A.I. Automated walk-in assistant for the blinds. In Proceedings of the 2014 International Conference on Electrical Engineering and Information & Communication Technology, Dhaka, Bangladesh, 10–12 April 2014; pp. 1–4. [Google Scholar]
  20. Zhou, D.; Yang, Y.; Yan, H. A Smart “Virtual Eye” Mobile System for the Visually Impaired. IEEE Potentials 2016, 35, 13–20. [Google Scholar] [CrossRef]
  21. Sharma, S.; Gupta, M.; Kumar, A.; Tripathi, M.; Gaur, M.S. Multiple distance sensors based smart stick for visually impaired people. In Proceedings of the 2017 IEEE 7th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 9–11 January 2017; pp. 1–5. [Google Scholar]
  22. Shaha, A.; Rewari, S.; Gunasekharan, S. SWSVIP-smart walking stick for the visually impaired people using low latency communication. In Proceedings of the 2018 International Conference on Smart City and Emerging Technology (ICSCET), Mumbai, India, 5 January 2018; pp. 1–5. [Google Scholar]
  23. Khanom, M.; Sadi, M.S.; Islam, M.M. A comparative study of walking assistance tools developed for the visually impaired people. In Proceedings of the 2019 1st International Conference on Advances in Science, Engineering and Robotics Technology (ICASERT), Dhaka, Bangladesh, 3–5 May 2019; pp. 1–5. [Google Scholar]
  24. Shahira, K.C.; Tripathy, S.; Lijiya, A. Obstacle detection, depth estimation and warning system for visually impaired people. In Proceedings of the TENCON 2019-2019 IEEE Region 10 Conference (TENCON), Kochi, India, 17–20 October 2019; pp. 863–868. [Google Scholar]
  25. Chang, W.J.; Chen, L.B.; Chen, M.C.; Su, J.P.; Sie, C.Y.; Yang, C.H. Design and implementation of an intelligent assistive system for visually impaired people for aerial obstacle avoidance and fall detection. IEEE Sens. J. 2020, 20, 10199–10210. [Google Scholar] [CrossRef]
  26. Cha, Y.J.; Ali, R.; Lewis, J.; Büyüköztürk, O. Deep learning-based structural health monitoring. Autom. Constr. 2024, 161, 105328. [Google Scholar] [CrossRef]
Figure 1. Stages of experimental model development.
Figure 1. Stages of experimental model development.
Technologies 12 00075 g001
Figure 2. Arduino Uno.
Figure 2. Arduino Uno.
Technologies 12 00075 g002
Figure 3. Arduino Nano.
Figure 3. Arduino Nano.
Technologies 12 00075 g003
Figure 4. ESP-WROOM-32 development board.
Figure 4. ESP-WROOM-32 development board.
Technologies 12 00075 g004
Figure 5. HC-SR04 Ultrasonic sensor.
Figure 5. HC-SR04 Ultrasonic sensor.
Technologies 12 00075 g005
Figure 6. Component block diagram.
Figure 6. Component block diagram.
Technologies 12 00075 g006
Figure 7. Used pins of ESP-WROOM-32 development doard.
Figure 7. Used pins of ESP-WROOM-32 development doard.
Technologies 12 00075 g007
Figure 8. Complete hardware diagram of the smart cane system.
Figure 8. Complete hardware diagram of the smart cane system.
Technologies 12 00075 g008
Figure 9. Smart cane system flowchart.
Figure 9. Smart cane system flowchart.
Technologies 12 00075 g009
Figure 10. Smart cane application interface.
Figure 10. Smart cane application interface.
Technologies 12 00075 g010
Figure 11. Positioning of components on the stick handle.
Figure 11. Positioning of components on the stick handle.
Technologies 12 00075 g011
Figure 12. Positioning of sensors, buzzer and jack plug.
Figure 12. Positioning of sensors, buzzer and jack plug.
Technologies 12 00075 g012
Figure 13. Experimental prototype of smart cane.
Figure 13. Experimental prototype of smart cane.
Technologies 12 00075 g013
Figure 14. Comparison between real distances and those measured by the ultrasonic sensor.
Figure 14. Comparison between real distances and those measured by the ultrasonic sensor.
Technologies 12 00075 g014
Figure 15. Accuracy rate reached by the proposed system.
Figure 15. Accuracy rate reached by the proposed system.
Technologies 12 00075 g015
Figure 16. Error rate reached by the proposed system.
Figure 16. Error rate reached by the proposed system.
Technologies 12 00075 g016
Table 1. Comparison between ESP-WROOM-32, Arduino Uno, and Arduino Nano [15].
Table 1. Comparison between ESP-WROOM-32, Arduino Uno, and Arduino Nano [15].
CharacteristicESP-WROOM-32Arduino UnoArduino Nano
ProcessorDual-core processor with a frequency of up to 240 [MHz]Single-core processor with a frequency of up to 16 [MHz]Single-core processor with a frequency of up to 16 [MHz]
ConnectivityBuilt-in Wi-Fi and Bluetooth connectivityNo built-in connectivity, requires additional hardwareNo built-in connectivity, requires additional hardware
Number of pins30 pins20 digital input/output pins and 6 analog input pins22 digital input/output pins and 8 analog input pins
Dimensions [mm]18 × 25.554 × 6918 × 45
PriceHigher priceLower priceLower price
Table 2. Main characteristics of HC-SR04 ultrasonic sensors.
Table 2. Main characteristics of HC-SR04 ultrasonic sensors.
CharacteristicValue
Supply Voltage5 [V]
Operating Current<15 [mA]
Measurement Precision±3 [mm]
Coverage Angle15°
Dimensions [mm]45 × 20 × 15
Table 3. Data collected from ultrasonic sensors.
Table 3. Data collected from ultrasonic sensors.
Real Distance ( D r ) [cm]Measured Distance ( D m ) [cm]Mean Distance [cm]Mean Error [cm]Accuracy [%]Error ( ϵ ) [%]
12345
2525252525252501000
504950494849491982
100979697979796.83.296.83.2
150146146147146146146.23.897.52.5
200196195196195196195.64.497.82.2
300295294293294295294.25.898.11.9
Table 4. Comparison of the proposed system with other related systems based on accuracy and error.
Table 4. Comparison of the proposed system with other related systems based on accuracy and error.
Developed SystemYearAccuracy [%]Error [cm]
Sadi et al. [19]201496.016
Zhou et al. [20]201685.458.4
Sharma et al. [21]201787.749.2
Shaha et al. [22]201890.139.6
Khanom et al. [23]201995.219.2
Shahira et al. [24]201997.012
Chang et al. [25]202094.522
Rahman et al. [17]202096.314.8
Proposed system202398.08
Table 5. Ability of the system to detect different obstacles.
Table 5. Ability of the system to detect different obstacles.
Obstacle TypeResult [%]
Wall100
Refrigerator100
Closet100
Stair100
Rough surface60
Table100
Mirror90
Table 6. Comparison of the characteristics of different types of sensors.
Table 6. Comparison of the characteristics of different types of sensors.
UltrasonicInfraredLaserRadar
Principle of operationTransmission and reception of acoustic wavesTransmission and reception of IR light pulsesTransmission and reception of light wavesTransmission and reception of microwaves
Range range2–400 [cm]20–150 [cm]Short range: 15–120 [cm] Long range: 10–50 [m]Short range: 0–200 [m] Long range: >3000 [m]
BandwidthWideQuite narrowNarrowDepends on antenna size
Weather conditionsVery little affectedAffectedAffectedAffected
CostLowLowVery highHigh
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Panazan, C.-E.; Dulf, E.-H. Intelligent Cane for Assisting the Visually Impaired. Technologies 2024, 12, 75. https://doi.org/10.3390/technologies12060075

AMA Style

Panazan C-E, Dulf E-H. Intelligent Cane for Assisting the Visually Impaired. Technologies. 2024; 12(6):75. https://doi.org/10.3390/technologies12060075

Chicago/Turabian Style

Panazan, Claudiu-Eugen, and Eva-Henrietta Dulf. 2024. "Intelligent Cane for Assisting the Visually Impaired" Technologies 12, no. 6: 75. https://doi.org/10.3390/technologies12060075

APA Style

Panazan, C. -E., & Dulf, E. -H. (2024). Intelligent Cane for Assisting the Visually Impaired. Technologies, 12(6), 75. https://doi.org/10.3390/technologies12060075

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop