Can information theory unravel the subtext in a Chekhovian short story?

J. Nathanael Philipp, Olav Mueller-Reichau, Matthias Irmer, Michael Richter, Max Kölbl


Abstract
In this study, we investigate whether information-theoretic measures such as surprisal can quantify the elusive notion of subtext in a Chekhovian short story. Specifically, we conduct a series of experiments for which we enrich the original text once with (different types of) meaningful glosses and once with fake glosses. For the different texts thus created, we calculate the surprisal values using two methods: using either a bag-of-words model or a large language model. We observe enrichment effects depending on the method, but no interpretable subtext effect.
Anthology ID:
2025.bsnlp-1.10
Volume:
Proceedings of the 10th Workshop on Slavic Natural Language Processing (Slavic NLP 2025)
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Jakub Piskorski, Pavel Přibáň, Preslav Nakov, Roman Yangarber, Michal Marcinczuk
Venues:
BSNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
84–90
Language:
URL:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.bsnlp-1.10/
DOI:
Bibkey:
Cite (ACL):
J. Nathanael Philipp, Olav Mueller-Reichau, Matthias Irmer, Michael Richter, and Max Kölbl. 2025. Can information theory unravel the subtext in a Chekhovian short story?. In Proceedings of the 10th Workshop on Slavic Natural Language Processing (Slavic NLP 2025), pages 84–90, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Can information theory unravel the subtext in a Chekhovian short story? (Philipp et al., BSNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.bsnlp-1.10.pdf