Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Target-oriented Sentiment Classification with Sequential Cross-modal Semantic Graph (2208.09417v2)

Published 19 Aug 2022 in cs.CV and cs.AI

Abstract: Multi-modal aspect-based sentiment classification (MABSC) is task of classifying the sentiment of a target entity mentioned in a sentence and an image. However, previous methods failed to account for the fine-grained semantic association between the image and the text, which resulted in limited identification of fine-grained image aspects and opinions. To address these limitations, in this paper we propose a new approach called SeqCSG, which enhances the encoder-decoder sentiment classification framework using sequential cross-modal semantic graphs. SeqCSG utilizes image captions and scene graphs to extract both global and local fine-grained image information and considers them as elements of the cross-modal semantic graph along with tokens from tweets. The sequential cross-modal semantic graph is represented as a sequence with a multi-modal adjacency matrix indicating relationships between elements. Experimental results show that the approach outperforms existing methods and achieves state-of-the-art performance on two standard datasets. Further analysis has demonstrated that the model can implicitly learn the correlation between fine-grained information of the image and the text with the given target. Our code is available at https://github.com/zjukg/SeqCSG.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yufeng Huang (14 papers)
  2. Zhuo Chen (319 papers)
  3. Jiaoyan Chen (85 papers)
  4. Jeff Z. Pan (78 papers)
  5. Zhen Yao (18 papers)
  6. Wen Zhang (170 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.