HYDRA: A Multi-Head Encoder-only Architecture for Hierarchical Text Classification

Fabian Karl, Ansgar Scherp


Abstract
We introduce HYDRA, a simple yet effective multi-head encoder-only architecture for hierarchical text classification that treats each level in the hierarchy as a separate classification task with its own label space. State-of-the-art approaches rely on complex components like graph encoders, label semantics, and autoregressive decoders. We demonstrate that such complexity is often unnecessary. Through parameter sharing and level-specific parameterization, HYDRA enables flat models to incorporate hierarchical awareness without architectural complexity. Experiments on four benchmarks (NYT, RCV1-V2, BGC, and WOS) demonstrate that HYDRA always increases the performance over flat models and matches or exceeds the performance of complex state-of-the-art methods.
Anthology ID:
2025.emnlp-main.472
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
9303–9314
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.472/
DOI:
Bibkey:
Cite (ACL):
Fabian Karl and Ansgar Scherp. 2025. HYDRA: A Multi-Head Encoder-only Architecture for Hierarchical Text Classification. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 9303–9314, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
HYDRA: A Multi-Head Encoder-only Architecture for Hierarchical Text Classification (Karl & Scherp, EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.472.pdf
Checklist:
 2025.emnlp-main.472.checklist.pdf