Transforming Causal LLM into MLM Encoder for Detecting Social Media Manipulation in Telegram

Anton Bazdyrev, Ivan Bashtovyi, Ivan Havlytskyi, Oleksandr Kharytonov, Artur Khodakovskyi


Abstract
We participated in the Fourth UNLP shared task on detecting social media manipulation in Ukrainian Telegram posts, addressing both multilabel technique classification and token-level span identification. We propose two complementary solutions: for classification, we fine-tune the decoder-only model with class-balanced grid-search thresholding and ensembling. For span detection, we convert causal LLM into a bidirectional encoder via masked language modeling pretraining on large Ukrainian and Russian news corpora before fine-tuning. Our solutions achieve SOTA metric results on both shared task track. Our work demonstrates the efficacy of bidirectional pretraining for decoder-only LLMs and robust threshold optimization, contributing new methods for disinformation detection in low-resource languages.
Anthology ID:
2025.unlp-1.13
Volume:
Proceedings of the Fourth Ukrainian Natural Language Processing Workshop (UNLP 2025)
Month:
July
Year:
2025
Address:
Vienna, Austria (online)
Editor:
Mariana Romanyshyn
Venues:
UNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
112–119
Language:
URL:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.unlp-1.13/
DOI:
Bibkey:
Cite (ACL):
Anton Bazdyrev, Ivan Bashtovyi, Ivan Havlytskyi, Oleksandr Kharytonov, and Artur Khodakovskyi. 2025. Transforming Causal LLM into MLM Encoder for Detecting Social Media Manipulation in Telegram. In Proceedings of the Fourth Ukrainian Natural Language Processing Workshop (UNLP 2025), pages 112–119, Vienna, Austria (online). Association for Computational Linguistics.
Cite (Informal):
Transforming Causal LLM into MLM Encoder for Detecting Social Media Manipulation in Telegram (Bazdyrev et al., UNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.unlp-1.13.pdf