Atsuhiro Takasu
2025
Table-Text Alignment: Explaining Claim Verification Against Tables in Scientific Papers
Xanh Ho
|
Sunisth Kumar
|
Yun-Ang Wu
|
Florian Boudin
|
Atsuhiro Takasu
|
Akiko Aizawa
Findings of the Association for Computational Linguistics: EMNLP 2025
Scientific claim verification against tables typically requires predicting whether a claim is supported or refuted given a table. However, we argue that predicting the final label alone is insufficient: it reveals little about the model’s reasoning and offers limited interpretability. To address this, we reframe table–text alignment as an explanation task, requiring models to identify the table cells essential for claim verification. We build a new dataset by extending the SciTab benchmark with human-annotated cell-level rationales. Annotators verify the claim label and highlight the minimal set of cells needed to support their decision. After the annotation process, we utilize the collected information and propose a taxonomy for handling ambiguous cases. Our experiments show that (i) incorporating table alignment information improves claim verification performance, and (ii) most LLMs, while often predicting correct labels, fail to recover human-aligned rationales, suggesting that their predictions do not stem from faithful reasoning.
2024
Syllable-level lyrics generation from melody exploiting character-level language model
Zhe Zhang
|
Karol Lasocki
|
Yi Yu
|
Atsuhiro Takasu
Findings of the Association for Computational Linguistics: EACL 2024
The generation of lyrics tightly connected to accompanying melodies involves establishing a mapping between musical notes and syllables of lyrics. This process requires a deep understanding of music constraints and semantic patterns at syllable-level, word-level, and sentence-level semantic meanings. However, pre-trained language models specifically designed at the syllable level are publicly unavailable. To solve these challenging issues, we propose to exploit fine-tuning character-level language models for syllable-level lyrics generation from symbolic melody. In particular, our method aims to fine-tune a character-level pre-trained language model, allowing to incorporation of linguistic knowledge of the language model into the beam search process of a syllable-level Transformer generator network. Besides, by exploring ChatGPT-based evaluation of generated lyrics in addition to human subjective evaluation, we prove that our approach improves the coherence and correctness of generated lyrics, without the need to train expensive new language models.
Search
Fix author
Co-authors
- Akiko Aizawa 1
- Florian Boudin 1
- Xanh Ho 1
- Sunisth Kumar 1
- Karol Lasocki 1
- show all...