Innovative technology for the hearing impaired makes use of visual cues

Ophthalmology Times EuropeOphthalmology Times Europe October 2022
Volume 18
Issue 08

New visually-guided hearing aid uses eye gaze to focus amplification of sound signals from the direction of the user’s attention.

Reviewed by Dr Gerald Kidd.

Innovative technology for the hearing impaired makes use of visual cues

The concept of a visually-guided hearing aid (VGHA) is intriguing and encompasses a number of different disciplines that are not for the faint of heart, considering the complexity of the technologies being brought together.

Although not a new idea, the VGHA, which evolved out of work on spatial hearing in the Psychoacoustics Laboratory at Boston University, Massachusetts, United States, is a prototype that is currently used for research that incorporates acoustic beamforming.

Beamforming is a way of harnessing audio-frequency signals to focus amplification of a signal from a specific direction. In the VGHA, the formed beam is manipulated by eye gaze to improve the ability of users to better focus on one sound despite the presence of nearby competing sounds.

Dr Gerald Kidd Jr, a professor in the department of Speech, Language and Hearing Sciences and director of the Psychoacoustics Laboratory at Boston University, explained that one of the early concepts motivating this work was the finding that individuals can listen very selectively in space. “Along the dimension left to right—i.e., azimuth—they can focus their attention on a particular point of interest and attenuate sources of sound that are off the axis of the focus of attention,” he said.

The evidence for this spatial tuning effect with natural hearing was first described in 2000.1 Investigators then incorporated the effect into studies on hearing loss and improving hearing aids.

“In considering how we could improve individuals’ hearing in situations in which there are multiple spatially distributed sources of sound, which has been a perennial problem for people who wear hearing aids, we worked on devising an algorithm to replicate normal spatial hearing for those who cannot hear well; that is, by separating the sources in azimuth,” Dr Kidd said.

Options for steering the beam

Dr Kidd and his colleagues worked in collaboration with investigators from Sensimetrics of Malden, Massachusetts, US. They designed a way to work with beamforming as a method to improve hearing aids by enhancing the signal-to-noise ratio for sounds that are immediately in front of the beamformer. This research extended earlier work at the Research Laboratory of Electronics at the Massachusetts Institute of Technology (MIT) in Cambridge, Massachusetts, US.2

One option was to mount microphones on a spectacle frame. Working with Dr Joseph Desloge, professor, Department of speech, Language and Hearing Sciences and director of the Psychocaoustics Laboratory, Boston University, the two groups identified a problem with a beamformer mounted on a spectacle frame—namely, the inability to move it from one place to another.

For example, it could not easily be used to follow conversations in a group. The listener’s head would have to move each time a different individual spoke, to point the beamformer in the direction of the speaker.

Another method of steering the beamformer was also considered; this involved using a hand dial or phone controls to move the beam into a desired direction. However, the most natural way of steering the beam was in plain sight—moving the beam with eye gaze. The linking of the auditory and visual attention was born.

The eyes and the beam move in tandem from left to right with the change in the source of the sound. This was considered to be a powerful way to improve amplification for individuals with hearing loss, Dr Kidd explained.

VGHA design and operation

The technology will be most beneficial for nearby sound sources

Dr Kidd and Dr Desloge settled on a system that used the signals from a commercially available eye tracker combined with a custom-made microphone array comprised of four rows of four microphones, flush-mounted on a flexible band that can be positioned across the top of the user’s head.3

Dr Kidd explained that the outputs from the microphones (the acoustic component) are combined using an algorithm that was applied to audio beamforming devised by the MIT group, among others. “This configuration optimises the response of the microphone array to the direction chosen by the user,” he said.

As Dr Kidd went on to explain, the signals from the microphone array are processed in such a way as to be maximally responsive to an azimuth that is determined by gaze as detected by the eye tracker. A requirement of the eye tracker is that it should have both a world-view camera that points outwards and a camera that points inwards to track pupillary location. The two used together calibrate where the eyes are positioned relative to where the external camera thinks the eyes are positioned.

Associated software also contains a previously measured set of head-related impulse responses. “These responses provide the values across frequency for the algorithm that determines the optimal phase response, or time delay, and amplitude used to weight the response of each microphone to optimise the responsiveness to a particular azimuth,” Dr Kidd stated.

The system was fitted on a mannequin and a recorded set of impulse responses obtained from a number of different locations in the front hemifield. The resolution obtained was good.

When placed on a user, the eye tracker senses the angle at which the eyes are positioned, selects the head-related transfer function that corresponds to that azimuth and convolves that with the approaching stimulus. “This provides a very highly directional response … as if the sound were coming from that location,” Dr Kidd explained.

Sound filtering and quality

The filtering function of the VGHA is sharpest at the higher frequencies (short wavelengths) and broadest at the lower frequencies (long wavelengths), resulting in the beamformer being most sharply tuned at higher frequencies.

Dr Kidd reported that sounds falling outside the focus of the beam are sharply attenuated at high frequencies and less attenuated at low frequencies, depending on the distance from the focus. “This typically alters the quality of the sounds that are off axis a bit,” he said. “However, the beamformer can provide a great improvement in the signal-to-noise ratio for nearby sound sources and in rooms with good acoustics.”

This technology seems to be most effective when used in a small conference room with individuals sitting around a table, or in a “cocktail party” scenario in which selective hearing can be impaired when multiple individuals are speaking simultaneously. In larger settings, such as at a concert or a play, the technology would not be beneficial for selecting one speaker from among others at a greater distance. “The basic physics of this technology is that it will be more beneficial for nearby sound sources,” Dr Kidd stated.

A step forward

A recent beneficial advancement in Dr Kidd’s research has been the development of a triple beamformer that should provide enhanced hearing for individuals with cochlear implants.4 By way of comparison, Dr Kidd noted that the original beamformer had a single-channel output that comprised one spatial filter used to provide sound to one or both ears, but without any difference in the sound that went to the two ears.

“With single channel output, [although] hearing is enhanced, our natural binaural hearing that provides a big advantage in locating sounds is lost,” he said. The investigators are working on multiple beams to focus on the primary sound source of interest. A second beam that is pointed to the right funnels sound only to the right ear, and a third beam pointed to the left funnels sound only to the left ear.

“This approach restores some of the normal binaural hearing in addition to the benefit of the original beam former,” Dr Kidd explained. This restoration, the major feature of the triple beamformer, works by improving the signal-to-noise ratio of the single-channel beam and improves the spatial hearing, by improving the ability to locate sound sources outside the beam.5

In patients with bilateral cochlear implants, the devices work mostly independently of each other, which results in loss of a great deal of the normal binaural spatial hearing. “The triple beamformer enhances the intra-aural differences that lead to spatial location and source segregation,” Dr Kidd stated. He believes that this research approach is promising for this patient population.

The VGHA research is currently confined to the laboratory, and computers perform the signal processing. In addition, for the technology to become commercially available, the portable system would have to be miniaturised and cosmetically acceptable in order to be worn on the head.

Dr Kidd said he anticipates that the technology may be first used in the cochlear implant community. “I hope that this research will lead to building a better hearing aid for solving situations such as the problem at the classical cocktail party,” he concluded. “In addition, the technology is not intended only for people with hearing loss. We believe that even people with normal hearing might benefit from this technology and it could have wide application.”

Gerald Kidd, PhD
Dr Kidd has no financial interest in this technology.
1. Arbogast TL, Kidd G Jr. Evidence for spatial tuning in informational masking using the probe-signal method. J Acoust Soc Am. 2000;108:1803-1810.
2. Desloge JG, Rabinowitz WM, Zurek PM. Microphone-array hearing aids with binaural output. I. Fixed-processing systems. IEEE Trans Audio Speech Lang Process.1997;5:529-542.
3. Kidd G Jr, Favrot S, Desloge J, et al. Design and preliminary testing of a visually-guided hearing aid. J Acoust Soc Am. 2013;133:EL202-EL207.
4. Kidd G Jr, Jennings TR, Byrne AJ. Enhancing the perceptual segregation and localization of sound sources with a triple beamformer. J Acoust Soc Am. 2020;148:3598.
5. Yun D, Jennings TR, Kidd G Jr, et al. Benefits of triple acoustic beamforming during speech-on-speech masking and sound localization for bilateral cochlear-implant users. J Acoust Soc Am. 2021;149:3052-3072.
Related Videos
ARVO 2024: Andrew D. Pucker, OD, PhD on measuring meibomian gland morphology with increased accuracy
 Allen Ho, MD, presented a paper on the 12 month results of a mutation agnostic optogenetic programme for patients with severe vision loss from retinitis pigmentosa
Noel Brennan, MScOptom, PhD, a clinical research fellow at Johnson and Johnson
ARVO 2024: President-elect SriniVas Sadda, MD, speaks with David Hutton of Ophthalmology Times
Elias Kahan, MD, a clinical research fellow and incoming PGY1 resident at NYU
Neda Gioia, OD, sat down to discuss a poster from this year's ARVO meeting held in Seattle, Washington
Eric Donnenfeld, MD, a corneal, cataract and refractive surgeon at Ophthalmic Consultants of Connecticut, discusses his ARVO presentation with Ophthalmology Times
John D Sheppard, MD, MSc, FACs, speaks with David Hutton of Ophthalmology Times
Paul Kayne, PhD, on assessing melanocortin receptors in the ocular space
Osamah Saeedi, MD, MS, at ARVO 2024
© 2024 MJH Life Sciences

All rights reserved.