Interpretable Multimodal Misinformation Detection with Logic Reasoning

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

4 Scopus Citations
View graph of relations

Author(s)

Related Research Unit(s)

Detail(s)

Original languageEnglish
Title of host publicationFindings of the Association for Computational Linguistics
Subtitle of host publicationACL 2023
Place of PublicationStroudsburg, PA
PublisherAssociation for Computational Linguistics (ACL)
Pages9781-9796
ISBN (print)978-1-959429-62-3
Publication statusPublished - Jul 2023

Publication series

NameProceedings of the Annual Meeting of the Association for Computational Linguistics
ISSN (Print)0736-587X

Conference

Title61st Annual Meeting of the Association for Computational Linguistics (ACL 2023)
LocationWestin Harbour Castle
PlaceCanada
CityToronto
Period9 - 14 July 2023

Abstract

Multimodal misinformation on online social platforms is becoming a critical concern due to increasing credibility and easier dissemination brought by multimedia content, compared to traditional text-only information. While existing multimodal detection approaches have achieved high performance, the lack of interpretability hinders these systems' reliability and practical deployment. Inspired by Neural-Symbolic AI which combines the learning ability of neural networks with the explainability of symbolic learning, we propose a novel logic-based neural model for multimodal misinformation detection which integrates interpretable logic clauses to express the reasoning process of the target task. To make learning effective, we parameterize symbolic logical elements using neural representations, which facilitate the automatic generation and evaluation of meaningful logic clauses. Additionally, to make our framework generalizable across diverse misinformation sources, we introduce five meta-predicates that can be instantiated with different correlations. Results on three public datasets (Twitter, Weibo, and Sarcasm) demonstrate the feasibility and versatility of our model. The implementation of our work can be found in this link. © 2023 Association for Computational Linguistics.

Citation Format(s)

Interpretable Multimodal Misinformation Detection with Logic Reasoning. / Liu, Hui; Wang, Wenya; Li, Haoliang.
Findings of the Association for Computational Linguistics: ACL 2023. Stroudsburg, PA: Association for Computational Linguistics (ACL), 2023. p. 9781-9796 (Proceedings of the Annual Meeting of the Association for Computational Linguistics).

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review