Ahsan Adeel
Towards Next-Generation Lip-Reading Driven Hearing-Aids: A preliminary Prototype Demo
Adeel, Ahsan; Gogate, Mandar; Hussain, Amir
Authors
Dr. Mandar Gogate M.Gogate@napier.ac.uk
Principal Research Fellow
Prof Amir Hussain A.Hussain@napier.ac.uk
Professor
Abstract
Speech enhancement aims to enhance the perceived speech quality and intelligibility in the presence of noise. Classical speech enhancement methods are mainly based on audio only processing which often perform poorly in adverse conditions, where overwhelming noise is present. This paper presents an interactive prototype demo, as part of a disruptive cognitivelyinspired multimodal hearing-aid being researched and developed at Stirling, as part of an EPSRC funded project (COGAVHEAR). The proposed technology contextually utilizes and integrates multimodal cues such as lip-reading, facial expressions, gestures, and noisy audio, to further enhance the quality and intelligibility of the noise-filtered speech signal. However, the preliminary work presented in this paper has used only lip-reading and noisy audio. Lip-reading driven deep learning algorithms are exploited to learn noisy audio-visual to clean audio mappings, leading to enhanced Weiner filtering for more effective noise cancellation. The term context-aware signifies the device’s learning and adaptable capabilities, which could be exploited in a wide-range of real-world applications, ranging from hearing-aids, listening devices, cochlear implants and telecommunications, to need for ear defenders in extreme noisy environments. Hearing-impaired users could experience more intelligible speech by contextually learning and switching between audio and visual cues. The preliminary interactive Demo employs randomly selected, real noisy speech videos from YouTube to qualitatively benchmark the performance of the proposed contextual audio-visual approach against a stateof-the-art deep learning based audio-only speech enhancement method.
Citation
Adeel, A., Gogate, M., & Hussain, A. (2017, August). Towards Next-Generation Lip-Reading Driven Hearing-Aids: A preliminary Prototype Demo. Presented at 1st International Workshop on Challenges in Hearing Assistive Technology (CHAT 2017), Stockholm, Sweden
Presentation Conference Type | Conference Paper (published) |
---|---|
Conference Name | 1st International Workshop on Challenges in Hearing Assistive Technology (CHAT 2017) |
Start Date | Aug 19, 2017 |
Publication Date | 2017 |
Deposit Date | May 28, 2024 |
Peer Reviewed | Peer Reviewed |
Pages | 61-64 |
Book Title | Proceedings of the 1st International Workshop on Challenges in Hearing Assistive Technology (CHAT 2017) |
You might also like
Statistical Downscaling Modeling for Temperature Prediction
(2024)
Book Chapter
Federated Learning for Market Surveillance
(2024)
Book Chapter
Robust Real-time Audio-Visual Speech Enhancement based on DNN and GAN
(2024)
Journal Article
Downloadable Citations
About Edinburgh Napier Research Repository
Administrator e-mail: repository@napier.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search