Skip to main content

Research Repository

Advanced Search

All Outputs (3)

ASPIRE - Real noisy audio-visual speech enhancement corpus (2020)
Data
Gogate, M., Dashtipour, K., Adeel, A., & Hussain, A. (2020). ASPIRE - Real noisy audio-visual speech enhancement corpus. [Data]. https://doi.org/10.5281/zenodo.4585619

ASPIRE is a a first of its kind, audiovisual speech corpus recorded in real noisy environment (such as cafe, restaurants) which can be used to support reliable evaluation of multi-modal Speech Filtering technologies. This dataset follows the same sen... Read More about ASPIRE - Real noisy audio-visual speech enhancement corpus.

Robust Visual Saliency Optimization Based on Bidirectional Markov Chains (2020)
Journal Article
Jiang, F., Kong, B., Li, J., Dashtipour, K., & Gogate, M. (2021). Robust Visual Saliency Optimization Based on Bidirectional Markov Chains. Cognitive Computation, 13, 69–80. https://doi.org/10.1007/s12559-020-09724-6

Saliency detection aims to automatically highlight the most important area in an image. Traditional saliency detection methods based on absorbing Markov chain only take into account boundary nodes and often lead to incorrect saliency detection when t... Read More about Robust Visual Saliency Optimization Based on Bidirectional Markov Chains.

CochleaNet: A robust language-independent audio-visual model for real-time speech enhancement (2020)
Journal Article
Gogate, M., Dashtipour, K., Adeel, A., & Hussain, A. (2020). CochleaNet: A robust language-independent audio-visual model for real-time speech enhancement. Information Fusion, 63, 273-285. https://doi.org/10.1016/j.inffus.2020.04.001

Noisy situations cause huge problems for the hearing-impaired, as hearing aids often make speech more audible but do not always restore intelligibility. In noisy settings, humans routinely exploit the audio-visual (AV) nature of speech to selectively... Read More about CochleaNet: A robust language-independent audio-visual model for real-time speech enhancement.