Difficulty Estimation in Natural Language Tasks with Action Scores

Aleksandar Angelov, Tsegaye Misikir Tashu, Matias Valdenegro-Toro


Abstract
This study investigates the effectiveness of the action score, a metric originally developed for computer vision tasks, in estimating sample difficulty across various natural language processing (NLP) tasks. Using transformer-based models, the action score is applied to sentiment analysis, natural language inference, and abstractive text summarization. The results demonstrate that the action score can effectively identify challenging samples in sentiment analysis and natural language inference, often capturing difficult instances that are missed by more established metrics like entropy. However, the effectiveness of the action score appears to be task-dependent, as evidenced by its performance in the abstractive text summarization task, where it exhibits a nearly linear relationship with entropy. The findings suggest that the action score can provide valuable insights into the characteristics of challenging samples in NLP tasks, particularly in classification settings. However, its application should be carefully considered in the context of each specific task and in light of emerging research on the potential value of hard samples in machine learning.
Anthology ID:
2025.trustnlp-main.24
Volume:
Proceedings of the 5th Workshop on Trustworthy NLP (TrustNLP 2025)
Month:
May
Year:
2025
Address:
Albuquerque, New Mexico
Editors:
Trista Cao, Anubrata Das, Tharindu Kumarage, Yixin Wan, Satyapriya Krishna, Ninareh Mehrabi, Jwala Dhamala, Anil Ramakrishna, Aram Galystan, Anoop Kumar, Rahul Gupta, Kai-Wei Chang
Venues:
TrustNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
351–364
Language:
URL:
https://preview.aclanthology.org/fix-sig-urls/2025.trustnlp-main.24/
DOI:
Bibkey:
Cite (ACL):
Aleksandar Angelov, Tsegaye Misikir Tashu, and Matias Valdenegro-Toro. 2025. Difficulty Estimation in Natural Language Tasks with Action Scores. In Proceedings of the 5th Workshop on Trustworthy NLP (TrustNLP 2025), pages 351–364, Albuquerque, New Mexico. Association for Computational Linguistics.
Cite (Informal):
Difficulty Estimation in Natural Language Tasks with Action Scores (Angelov et al., TrustNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/fix-sig-urls/2025.trustnlp-main.24.pdf