Brady Bhalla
2025
TRACE: Real-Time Multimodal Common Ground Tracking in Situated Collaborative Dialogues
Hannah VanderHoeven
|
Brady Bhalla
|
Ibrahim Khebour
|
Austin C. Youngren
|
Videep Venkatesha
|
Mariah Bradford
|
Jack Fitzgerald
|
Carlos Mabrey
|
Jingxuan Tu
|
Yifan Zhu
|
Kenneth Lai
|
Changsoo Jung
|
James Pustejovsky
|
Nikhil Krishnaswamy
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (System Demonstrations)
We present TRACE, a novel system for live *common ground* tracking in situated collaborative tasks. With a focus on fast, real-time performance, TRACE tracks the speech, actions, gestures, and visual attention of participants, uses these multimodal inputs to determine the set of task-relevant propositions that have been raised as the dialogue progresses, and tracks the group’s epistemic position and beliefs toward them as the task unfolds. Amid increased interest in AI systems that can mediate collaborations, TRACE represents an important step forward for agents that can engage with multiparty, multimodal discourse.