A (More) Realistic Evaluation Setup for Generalisation of Community Models on Malicious Content Detection
arXiv | ACL Anthology | Github
Abstract
Community models for malicious content detection, which take into account the context from a social graph alongside the content itself, have shown remarkable performance on benchmark datasets. Yet, misinformation and hate speech continue to propagate on social media networks. This mismatch can be partially attributed to the limitations of current evaluation setups that neglect the rapid evolution of online content and the underlying social graph. In this paper, we propose a novel evaluation setup for model generalisation based on our few-shot subgraph sampling approach. This setup tests for generalisation through few labelled examples in local explorations of a larger graph, emulating more realistic application settings. We show this to be a challenging inductive setup, wherein strong performance on the training graph is not indicative of performance on unseen tasks, domains, or graph structures. Lastly, we show that graph meta-learners trained with our proposed few-shot subgraph sampling outperform standard community models in the inductive setup. We make our code publicly available.
Citation
1@inproceedings{verhoeven-etal-2024-realistic,
2 title = "A (More) Realistic Evaluation Setup for Generalisation of Community Models on Malicious Content Detection",
3 author = "Verhoeven, Ivo and
4 Mishra, Pushkar and
5 Beloch, Rahel and
6 Yannakoudakis, Helen and
7 Shutova, Ekaterina",
8 editor = "Duh, Kevin and
9 Gomez, Helena and
10 Bethard, Steven",
11 booktitle = "Findings of the Association for Computational Linguistics: NAACL 2024",
12 month = jun,
13 year = "2024",
14 address = "Mexico City, Mexico",
15 publisher = "Association for Computational Linguistics",
16 url = "https://aclanthology.org/2024.findings-naacl.30",
17 doi = "10.18653/v1/2024.findings-naacl.30",
18 pages = "437--463",
19}
#misinformation #meta-learning #graph-neural-networks #generalisation #phd