Most navigation aids for visually impaired individuals require users to pay close attention and actively understand the instructions or feedback of guidance, which impose considerable cognitive loads in long-term usage. The “Aerial Guide Dog”, a helium balloon aerostat drone designed for indoor guidance, which leverages gentle tugs in real time for directional guidance, ensuring a seamless and intuitive guiding experience. The introduced Aerial Guide Dog has been evaluated in terms of directional guidance and path following in the pilot study, focusing on assessing its accuracy in orientation and the overall performance in navigation.
1. Introduction
According to August 2023 statistics from the World Health Organization (WHO), at least 1 billion individuals suffer from vision impairments or blindness. Vision disorders or blindness affect people of all ages, potentially limiting their educational development, reducing labor participation rates, social interaction, and independence, and often lead to a high prevalence of depression. These issues significantly impact the quality of life for individuals with visual impairments
[1][2][3][4][5].
Vision is one of humans’ most important senses, essential for normal living and normal moving around. When travelling in unknown environments, it aids individuals in recognizing environmental features to find the correct path and avoid potential hazards on the way
[6][7]. However, for individuals with visual impairments, navigating unfamiliar environments and complex buildings is particularly challenging. They often cannot identify key features to guide movements for negotiating stairs/steps/doors or avoid obstacles, such as walls/people/furniture, etc., to reach their desired destination, resulting in feelings of insecurity and anxiety. In fact, up to 70% of individuals with visual impairments tend to avoid moving around independently in indoor spaces, perceiving shopping malls as one of the most challenging environments. When it is essential for them to go shopping, they must rely on getting help from sighted people, which not only undermines their confidence and independence, but also significantly affects their ability to gain more experience in carrying out indoor everyday activities
[7][8][9][10][11][12].
Due to the difficulty visual impaired individuals face in recognizing their surroundings, white canes and guide dogs have become preferred solutions because of their simplicity and intuitive nature. However, they have limitations in that they primarily help in identifying objects near the user and are thus mainly suitable for individuals with reasonable confidence and the ability to move around effectively
[5][7][13]. The cane relies on tactile feedback transmitted mainly from its tip when swung at ground level, making it difficult to detect obstacles higher up above the swinging range, placing the users in potential hazardous situations
[6][10][13]. While guide dogs offer an intuitive and solution, they are limited by an insufficient supply of trained dogs whose lifespan is relatively short (about 6 to 8 years) and high training costs (≈USD 42,000)
[5].
Systems designed to enhance the walking autonomy of blind individuals through various technological solutions are generally referred to as Electronic Travel Aids (ETAs). The design of ETAs is particularly filled with challenges
[10], because very demanding requirements need to be met, such as real-time guidance, portability, power limitations, suitable interfaces, continuous availability, independence from infrastructure, low-cost solutions, and minimal training. Simultaneously, the system should be easy to use, clear, and user-friendly
[5][6][12][13][14]. However, various studies on assistive technology for the blind have primarily focused on object recognition, navigation, and mobility
[7], exploring the diverse needs of visually impaired individuals in different activity scenarios. These studies aim to solve context-specific challenges by developing various technological solutions. Currently, no single assistive device has been developed that can be used as extensively and long term in the lives of visually impaired individuals as traditional white canes and guide dogs. Therefore, the focus should be on developing cost-effective, user-friendly long-term solutions able to be used in real-world situations, rather than solely advancing technology
[11].
In the past decade of research on indoor ETAs, substituting visual perception via alternative methods has been a mainstream approach
[15] stemming from the theory of sensory substitution neuroplasticity. This refers to the capability of the brain to assimilate specific sensory information in alternative ways
[12] and requires individuals to consciously integrate their sensory disability with their other functioning senses; e.g., visual impairment can be replaced with auditory and/or tactile senses
[6][11].
Based on such sensory substitution approaches, indoor ETAs have been designed relying on methods to generate stimuli that substitute for vision, and users need to learn how to understand the (auditory/tactile) signals to successfully complete activities such as travelling tasks in complex environments. Complexity in understanding the environmental information has been recognized as placing a large cognitive load on the user
[15]. Cognitive Load Theory suggests that our working memory is only able to hold a small amount of information at any one time and that instructional methods should avoid overloading it in order to maximize learning
[16], and researchers like Giudice et al. have suggested that developers should focus on assisting users in performing specific and necessary tasks, while minimizing the amount of information passed to the user. Indoor ETAs involve utilizing “perceptual and cognitive factors related to processing non-visual information”; however, the bandwidth of non-visual senses such as auditory, tactile, and olfactory is much smaller than that of vision
[17]. Hence, balancing the relationship between the minimum and the necessary information becomes crucial
[12][18].
The main sensory approaches for replacing visual information are via auditory and tactile methods. Audio methods can be divided into audio description and spatial audio; audio description can provide general guidance but often lacks the detail needed for precise movements
[19][20][21][22][23], whereas spatial audio, which links sound source locations to intended directions, is more intuitive for the user. However, spatial audio can interfere with environmental sounds, which can cause hazardous situations to arise
[24][25][26][27][28][29][30]. Tactile methods involve vibrotactile and kinesthetic approaches. Vibrotactile feedback uses vibrations to convey environmental information and can be felt on different body parts, but its effectiveness varies due to factors like body part sensitivity and clothing thickness
[30][31][32][33][34][35][36][37][38]. Kinesthetic devices use traction force for providing directional cues. For instance, Antolini et al. proposed a method of providing kinesthetic stimulation to users by tilting a flywheel inside the device, allowing users to determine left or right directions based on the sensation of motion simulated by the flywheel, thereby guiding user navigation
[39]. Another method includes devices that change shape to provide directional clues
[38]. For example, Spiers et al. proposed a cube-like device with a top section that rotates and extends, providing tactile feedback on various finger areas to indicate direction
[38][40][41].
2. A Low-Cognitive-Load Indoor Electronic Travel Aid for Visually Impaired Individuals
The preferences, suggestions, and actual needs of visually impaired individuals regarding ETAs in both indoor and outdoor environments are crucial references for researchers developing suitable commercial ETA solutions. In this regard, Plikynas et al.
[6] conducted comprehensive interviews with 25 blind experts, revealing that 16 of them avoided using any ETAs for indoor navigation due to the absence of suitable and convenient commercial solutions. Therefore, compared to existing outdoor solutions, the market still demands further development and enhancement of suitable tactile and auditory devices for indoor orientation and navigation
[8].
The pros and cons of navigation system feedback methods are a qualitative assessment, varying according to the specific needs and capabilities of users in different environments. Plikynas et al. indicate that, taking voice commands as an example, visually impaired individuals tend to prefer this type of audio feedback for outdoor navigation as compared to indoor environments
[6]. Therefore, it is crucial to provide users with appropriate feedback methods tailored to specific situations and needs. Although tactile feedback may encounter limitations in comprehending all transmitted information in areas of perception after prolonged use, visually impaired individuals still show a preference for receiving commands or information through this feedback method in indoor environments
[6].Hence, considering tactile feedback as a more accepted method for visually impaired individuals in indoor environments, it should be prioritized as a vital sensory alternative in the development of indoor ETAs.
Enhancing vibration-sensed signals through advanced signal processing algorithms to convey directional information to visually impaired users is a common tactile and effective feedback method. Robert et al. proposed a method called ALVU (Array of Lidars and Vibrotactile Units), which includes a sensor belt worn around the waist and a separate tactile belt worn around the upper abdomen
[42]. The sensor belt operates by emitting infrared light pulses to measure the distance between the person and nearby obstacles, effectively detecting obstacles around the individual. In contrast, the tactile belt utilizes vibrating motors to provide feedback. These motors adjust their vibration frequency and intensity based on the distance to detected obstacles, as measured by the sensor belt, thus conveying the distance information of these obstacles to the user
[39]. This system has been identified as an effective method of feedback. Khusro et al. developed a real-time feedback system for indoor navigation that utilizes the vibration motors within smartphones to deliver rich tactile information based on vibration characteristics such as frequency, rhythm, and duration. By systematically arranging different lengths of patterns in the manner of Morse code, this system mimics natural tones familiar to users, such as ‘heartbeat’ and ‘knocking’, thereby greatly improving the learnability and understandability of the information received by users
[43]. See et al. utilized a robotic operating system to integrate depth camera sensors and obstacle localization algorithms, employing tactile feedback to detect obstacles surrounding the user. This wearable device, equipped with vibration motors in various areas on the user’s body, conveys the location of obstacles by activating the corresponding directional motor and indicates the distance to these obstacles through the intensity of the vibrations. Users can stop and make necessary adjustments based on the specific vibration cues to navigate around all types of obstacles
[44].
In most approaches, the tactile signals of the assistive devices developed rely on coding and requiring users to learn to understand the “coded information” corresponding to different vibration signals which can demand significant effort to learn and memorize. Additionally, while the information provided via vibration-based mechanisms is generally effective, prolonged use can lead to fatigue and numbness, resulting in individuals being unable to comprehend all the information for effective use
[6]. Another approach using force feedback for guidance has been found to be more intuitive and less demanding cognitively. Federica et al.
[10] proposed an ETA system where users receive directional haptic feedback through forces provided by motors worn around in an armband. The device works through the motors spinning in opposite directions to tighten or loosen the armband, advising the user to walk or stop, and the motors spinning in the same direction, causing the armband to slide up or down the arm advising the user to turn left or right. This simple method has been evaluated to convey clear directional information through pressure and skin stretching on specific body parts, akin to a volunteer holding a visually impaired person’s arm for guidance. Navigating by replicating such familiar approaches from the experiences of visual impaired persons is clearly a valid method to adopt in realizing effective user-centered designs that can work well in real-world situations. However, an issue that needs to be addressed is that the thickness of clothing needs to be taken into account as it can affect the user’s perception of the signals. Therefore, compared to reproducing the method of volunteers guiding individuals with visual impairments, the Aerial Guide Dog chooses the more sensitive finger pulp area for tactile feedback. By emulating the working method of guide dogs to lead the visually impaired, it can enhance the effectiveness of perception and reduce the impact of other external factors.
Avila et al.
[24] demonstrated that an assistive navigation system with a drone as the guidance module is an efficient and accurate method of guiding, as it provides continuous directional feedback
[45]. Notably, the drones developed utilize a soft rope to relay the forces to the user to enhance the independent navigation abilities of visually impaired users. However, due to the use of a soft rope connection, users must maintain a strict relative spatial position with the drone to fully perceive the traction force, as any change in relative position renders it ineffective. When users follow the drone for navigation, changing their walking speed can cause a mismatch between the expected and actual traction forces provided by the drone, leading to ambiguous directional guidance
[46]. Compared to the guidance systems of commercial drones, the Aerial Guide Dog utilizes a quieter helium balloon aerostat drone and uses a flexible carbon rod for the traction rope, ensuring that users can clearly perceive directional signals by merely holding the handle. This makes the new approach presented here more in line with the visually impaired users’ requirements as well as being cost-effective
[7][11].
Compared to the traditional robotic guide dog method developed by Hwang et al., the advantage of the Aerial Guide Dog lies in its flying guidance approach
[47], which reduces ground interaction challenges with the environment encountered
[7]. Being above the ground, it also has a wider field of view, thereby improving its range to provide more complete environmental information to the user. Furthermore, this aerial approach reduces the wear and tear often seen in ground-based systems due to continuous contact with irregularities of the ground surface.
The introduction of the Aerial Guide Dog as an indoor ETA is felt to represent a significant advancement in technology to help individuals with visual impairments move around effectively. Furthermore, it underscores the importance and practical applicability of the Aerial Guide Dog’s tactile sensory substitution approach, which needs to be investigated in future research on indoor ETAs.
This entry is adapted from the peer-reviewed paper 10.3390/s24010297