@inproceedings{zhang-etal-2024-omagent,
title = "{O}m{A}gent: A Multi-modal Agent Framework for Complex Video Understanding with Task Divide-and-Conquer",
author = "Zhang, Lu and
Zhao, Tiancheng and
Ying, Heting and
Ma, Yibo and
Lee, Kyusong",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2024.emnlp-main.559/",
doi = "10.18653/v1/2024.emnlp-main.559",
pages = "10031--10045",
abstract = "Recent advancements in Large Language Models (LLMs) have expanded their capabilities to multimodal contexts, including comprehensive video understanding. However, processing extensive videos such as 24-hour CCTV footage or full-length films presents significant challenges due to the vast data and processing demands. Traditional methods, like extracting key frames or converting frames to text, often result in substantial information loss. To address these shortcomings, we develop OmAgent, efficiently stores and retrieves relevant video frames for specific queries, preserving the detailed content of videos. Additionally, it features an Divide-and-Conquer Loop capable of autonomous reasoning, dynamically invoking APIs and tools to enhance query processing and accuracy. This approach ensures robust video understanding, significantly reducing information loss. Experimental results affirm OmAgent{'}s efficacy in handling various types of videos and complex tasks. Moreover, we have endowed it with greater autonomy and a robust tool-calling system, enabling it to accomplish even more intricate tasks."
}
Markdown (Informal)
[OmAgent: A Multi-modal Agent Framework for Complex Video Understanding with Task Divide-and-Conquer](https://preview.aclanthology.org/fix-sig-urls/2024.emnlp-main.559/) (Zhang et al., EMNLP 2024)
ACL