Evelyn Duesterwald
2025
FLOW-BENCH: Towards Conversational Generation of Enterprise Workflows
Evelyn Duesterwald
|
Siyu Huo
|
Vatche Isahagian
|
K. R. Jayaram
|
Ritesh Kumar
|
Vinod Muthusamy
|
Punleuk Oum
|
Debashish Saha
|
Gegi Thomas
|
Praveen Venkateswaran
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
Large Language Models (LLMs) can be used to convert natural language (NL) instructions into structured business process automation (BPA) process artifacts.This paper contributes (i) FLOW-BENCH, a high quality dataset of paired NL instructions and business process definitions toevaluate NL-based BPA tools, and support research in this area, and (ii) FLOW-GEN,our approach to utilize LLMs to translate NL into an intermediate Python representation that facilitates final conversion into widely adopted business process definition languages, such as BPMN and DMN. We bootstrap FLOW-BENCH by demonstrating how it can be used to evaluate the components of FLOW-GEN across eight LLMs. We hope that FLOW-GEN and FLOW-BENCHcatalyze further research in BPA.
2023
DiSTRICT: Dialogue State Tracking with Retriever Driven In-Context Tuning
Praveen Venkateswaran
|
Evelyn Duesterwald
|
Vatche Isahagian
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing
Dialogue State Tracking (DST), a key component of task-oriented conversation systems, represents user intentions by determining the values of pre-defined slots in an ongoing dialogue. Existing approaches use hand-crafted templates and additional slot information to fine-tune and prompt large pre-trained language models and elicit slot values from the dialogue context. Significant manual effort and domain knowledge is required to design effective prompts, limiting the generalizability of these approaches to new domains and tasks. In this work, we propose DiSTRICT, a generalizable in-context tuning approach for DST that retrieves highly relevant training examples for a given dialogue to fine-tune the model without any hand-crafted templates. Experiments with the MultiWOZ benchmark datasets show that DiSTRICT outperforms existing approaches in various zero-shot and few-shot settings using a much smaller model, thereby providing an important advantage for real-world deployments that often have limited resource availability.
Search
Fix author
Co-authors
- Vatche Isahagian 2
- Praveen Venkateswaran 2
- Siyu Huo 1
- K. R. Jayaram 1
- Ritesh Kumar 1
- show all...