Soujanya Poria
Fusing audio, visual and textual clues for sentiment analysis from multimodal content
Poria, Soujanya; Cambria, Erik; Howard, Newton; Huang, Guang-Bin; Hussain, Amir
Authors
Abstract
A huge number of videos are posted every day on social media platforms such as Facebook and YouTube. This makes the Internet an unlimited source of information. In the coming decades, coping with such information and mining useful knowledge from it will be an increasingly difficult task. In this paper, we propose a novel methodology for multimodal sentiment analysis, which consists in harvesting sentiments from Web videos by demonstrating a model that uses audio, visual and textual modalities as sources of information. We used both feature- and decision-level fusion methods to merge affective information extracted from multiple modalities. A thorough comparison with existing works in this area is carried out throughout the paper, which demonstrates the novelty of our approach. Preliminary comparative experiments with the YouTube dataset show that the proposed multimodal system achieves an accuracy of nearly 80%, outperforming all state-of-the-art systems by more than 20%.
Citation
Poria, S., Cambria, E., Howard, N., Huang, G.-B., & Hussain, A. (2016). Fusing audio, visual and textual clues for sentiment analysis from multimodal content. Neurocomputing, 174(Part A), 50-59. https://doi.org/10.1016/j.neucom.2015.01.095
Journal Article Type | Article |
---|---|
Acceptance Date | Jan 2, 2015 |
Online Publication Date | Aug 17, 2015 |
Publication Date | Jan 22, 2016 |
Deposit Date | Oct 4, 2019 |
Journal | Neurocomputing |
Print ISSN | 0925-2312 |
Publisher | Elsevier |
Peer Reviewed | Peer Reviewed |
Volume | 174 |
Issue | Part A |
Pages | 50-59 |
DOI | https://doi.org/10.1016/j.neucom.2015.01.095 |
Keywords | Multimodal fusion; Big social data analysis; Opinion mining; Multimodal sentiment analysis; Sentic computing |
Public URL | http://researchrepository.napier.ac.uk/Output/1792721 |
You might also like
MTFDN: An image copy‐move forgery detection method based on multi‐task learning
(2024)
Journal Article
Transition-aware human activity recognition using an ensemble deep learning framework
(2024)
Journal Article
Downloadable Citations
About Edinburgh Napier Research Repository
Administrator e-mail: repository@napier.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search