@inproceedings{wei-zhu-2025-protolens,
    title = "{P}roto{L}ens: Advancing Prototype Learning for Fine-Grained Interpretability in Text Classification",
    author = "Wei, Bowen  and
      Zhu, Ziwei",
    editor = "Che, Wanxiang  and
      Nabende, Joyce  and
      Shutova, Ekaterina  and
      Pilehvar, Mohammad Taher",
    booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
    month = jul,
    year = "2025",
    address = "Vienna, Austria",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2025.acl-long.226/",
    doi = "10.18653/v1/2025.acl-long.226",
    pages = "4503--4523",
    ISBN = "979-8-89176-251-0",
    abstract = "In this work, we propose ProtoLens, a novel prototype-based model that provides fine-grained, sub-sentence level interpretability for text classification. ProtoLens uses a Prototype-aware Span Extraction module to identify relevant text spans associated with learned prototypes and a Prototype Alignment mechanism to ensure prototypes are semantically meaningful throughout training. By aligning the prototype embeddings with human-understandable examples, ProtoLens provides interpretable predictions while maintaining competitive accuracy. Extensive experiments demonstrate that ProtoLens outperforms both prototype-based and non-interpretable baselines on multiple text classification benchmarks. Code and data are available at \url{https://github.com/weibowen555/ProtoLens}."
}Markdown (Informal)
[ProtoLens: Advancing Prototype Learning for Fine-Grained Interpretability in Text Classification](https://preview.aclanthology.org/ingest-emnlp/2025.acl-long.226/) (Wei & Zhu, ACL 2025)
ACL