Tanmoy Mukherjee
2025
Connecting Concept Layers and Rationales to Enhance Language Model Interpretability
Thomas Bailleux
|
Tanmoy Mukherjee
|
Pierre Marquis
|
Zied Bouraoui
Proceedings of the 14th Joint Conference on Lexical and Computational Semantics (*SEM 2025)
With the introduction of large language models, NLP has undergone a paradigm shift where these models now serve as the backbone of most developed systems. However, while highly effective, they remain opaque and difficult to interpret, which limits their adoption in critical applications that require transparency and trust. Two major approaches aim to address this: rationale extraction, which highlights input spans that justify predictions, and concept bottleneck models, which make decisions through human-interpretable concepts. Yet each has limitations. Crucially, current models lack a unified framework that connects where a model looks (rationales) with why it makes a decision (concepts). We introduce CLARITY, a model that first selects key input spans, maps them to interpretable concepts, and then predicts using only those concepts. This design supports faithful, multi-level explanations and allows users to intervene at both the rationale and concept levels. CLARITY, achieves competitive accuracy while offering improved transparency and controllability.
2018
Learning Unsupervised Word Translations Without Adversaries
Tanmoy Mukherjee
|
Makoto Yamada
|
Timothy Hospedales
Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing
Word translation, or bilingual dictionary induction, is an important capability that impacts many multilingual language processing tasks. Recent research has shown that word translation can be achieved in an unsupervised manner, without parallel seed dictionaries or aligned corpora. However, state of the art methods unsupervised bilingual dictionary induction are based on generative adversarial models, and as such suffer from their well known problems of instability and hyper-parameter sensitivity. We present a statistical dependency-based approach to bilingual dictionary induction that is unsupervised – no seed dictionary or parallel corpora required; and introduces no adversary – therefore being much easier to train. Our method performs comparably to adversarial alternatives and outperforms prior non-adversarial methods.
2016
Gaussian Visual-Linguistic Embedding for Zero-Shot Recognition
Tanmoy Mukherjee
|
Timothy Hospedales
Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing