Leandro A. Passos
Multimodal audio-visual information fusion using canonical-correlated Graph Neural Network for energy-efficient speech enhancement
Passos, Leandro A.; Papa, Jo�o Paulo; Del Ser, Javier; Hussain, Amir; Adeel, Ahsan
Authors
Abstract
This paper proposes a novel multimodal self-supervised architecture for energy-efficient audio-visual (AV) speech enhancement that integrates Graph Neural Networks with canonical correlation analysis (CCA-GNN). The proposed approach lays its foundations on a state-of-the-art CCA-GNN that learns representative embeddings by maximizing the correlation between pairs of augmented views of the same input while decorrelating disconnected features. The key idea of the conventional CCA-GNN involves discarding augmentation-variant information and preserving augmentation-invariant information while preventing capturing of redundant information. Our proposed AV CCA-GNN model deals with multimodal representation learning context. Specifically, our model improves contextual AV speech processing by maximizing canonical correlation from augmented views of the same channel and canonical correlation from audio and visual embeddings. In addition, it proposes a positional node encoding that considers a prior-frame sequence distance instead of a feature-space representation when computing the node’s nearest neighbors, introducing temporal information in the embeddings through the neighborhood’s connectivity. Experiments conducted on the benchmark ChiME3 dataset show that our proposed prior frame-based AV CCA-GNN ensures a better feature learning in the temporal context, leading to more energy-efficient speech reconstruction than state-of-the-art CCA-GNN and multilayer perceptron.
Citation
Passos, L. A., Papa, J. P., Del Ser, J., Hussain, A., & Adeel, A. (2023). Multimodal audio-visual information fusion using canonical-correlated Graph Neural Network for energy-efficient speech enhancement. Information Fusion, 90, 1-11. https://doi.org/10.1016/j.inffus.2022.09.006
Journal Article Type | Article |
---|---|
Acceptance Date | Sep 6, 2022 |
Online Publication Date | Sep 11, 2022 |
Publication Date | 2023-02 |
Deposit Date | Sep 28, 2022 |
Publicly Available Date | Mar 12, 2024 |
Journal | Information Fusion |
Print ISSN | 1566-2535 |
Publisher | Elsevier |
Peer Reviewed | Peer Reviewed |
Volume | 90 |
Pages | 1-11 |
DOI | https://doi.org/10.1016/j.inffus.2022.09.006 |
Keywords | Canonical correlation analysis, Graph Neural Networks, Multimodal learning, Positional encoding, Prior frames neighborhood |
Public URL | http://researchrepository.napier.ac.uk/Output/2917348 |
Files
Multimodal audio-visual information fusion using canonical-correlated Graph Neural Network for energy-efficient speech enhancement (accepted version)
(1.1 Mb)
PDF
You might also like
Transition-aware human activity recognition using an ensemble deep learning framework
(2024)
Journal Article
Downloadable Citations
About Edinburgh Napier Research Repository
Administrator e-mail: repository@napier.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search