Toyin Aguda
2025
Advanced Messaging Platform (AMP): Pipeline for Automated Enterprise Email Processing
Simerjot Kaur
|
Charese Smiley
|
Keshav Ramani
|
Elena Kochkina
|
Mathieu Sibue
|
Samuel Mensah
|
Pietro Totis
|
Cecilia Tilli
|
Toyin Aguda
|
Daniel Borrajo
|
Manuela Veloso
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 6: Industry Track)
Understanding and effectively responding to email communication remains a critical yet complex challenge for current AI techniques, especially in corporate environments. These tasks are further complicated by the need for domain-specific knowledge, accurate entity recognition, and high precision to prevent costly errors. While recent advances in AI, specifically Large Language Models (LLMs), have made strides in natural language understanding, they often lack business-specific expertise required in such settings. In this work, we present Advanced Messaging Platform (AMP), a production-grade AI pipeline that automates email response generation at scale in real-world enterprise settings. AMP has been in production for more than a year, processing thousands of emails daily while maintaining high accuracy and adaptability to evolving business needs.
Conservative Bias in Large Language Models: Measuring Relation Predictions
Toyin Aguda
|
Erik Wilson
|
Allan Anzagira
|
Simerjot Kaur
|
Charese Smiley
Findings of the Association for Computational Linguistics: ACL 2025
Large language models (LLMs) exhibit pronounced conservative bias in relation extraction tasks, frequently defaulting to no_relation label when an appropriate option is unavailable. While this behavior helps prevent incorrect relation assignments, our analysis reveals that it also leads to significant information loss when reasoning is not explicitly included in the output. We systematically evaluate this trade-off across multiple prompts, datasets, and relation types, introducing the concept of Hobson’s choice to capture scenarios where models opt for safe but uninformative labels over hallucinated ones. Our findings suggest that conservative bias occurs twice as often as hallucination. To quantify this effect, we use SBERT and LLM prompts to capture the semantic similarity between conservative bias behaviors in constrained prompts and labels generated from semi-constrained and open-ended prompts.
Search
Fix author
Co-authors
- Simerjot Kaur 2
- Charese Smiley 2
- Allan Anzagira 1
- Daniel Borrajo 1
- Elena Kochkina 1
- show all...