Research Output
Benchmarking multimodal sentiment analysis
  We propose a deep-learning-based framework for multimodal sentiment analysis and emotion recognition. In particular, we leverage on the power of convolutional neural networks to obtain a performance improvement of 10% over the state of the art by combining visual, text and audio features. We also discuss some major issues frequently ignored in multimodal sentiment analysis research, e.g., role of speaker-independent models, importance of different modalities, and generalizability. The framework illustrates the different facets of analysis to be considered while performing multimodal sentiment analysis and, hence, serves as a new benchmark for future research in this emerging field.

Citation

Cambria, E., Hazarika, D., Poria, S., Hussain, A., & Subramanyam, R. (2018). Benchmarking multimodal sentiment analysis. In Computational Linguistics and Intelligent Text Processing. , (166-179). https://doi.org/10.1007/978-3-319-77116-8_13

Authors

Keywords

Multimodal sentiment analysis, Emotion detection, Deep learning, Convolutional neural networks

Monthly Views:

Available Documents