Thomas Gschwind
2025
Classifier-Augmented Generation for Structured Workflow Prediction
Thomas Gschwind
|
Shramona Chakraborty
|
Nitin Gupta
|
Sameep Mehta
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
ETL (Extract, Transform, Load) tools such as IBM DataStage allow users to visually assemble complex data workflows, but configuring stages and their properties remains time consuming and requires deep tool knowledge. We propose a system that translates natural language descriptions into executable workflows, automatically predicting both the structure and detailed configuration of the flow. At its core lies a Classifier-Augmented Generation (CAG) approach that combines utterance decomposition with a classifier and stage-specific few-shot prompting to produce accurate stage predictions. These stages are then connected into non-linear workflows using edge prediction, and stage properties are inferred from sub-utterance context. We compare CAG against strong single-prompt and agentic baselines, showing improved accuracy and efficiency, while substantially reducing token usage. Our architecture is modular, interpretable, and capable of end-to-end workflow generation, including robust validation steps. To our knowledge, this is the first system with a detailed evaluation across stage prediction, edge layout, and property generation for natural-language-driven ETL authoring.
2024
Adapting LLMs for Structured Natural Language API Integration
Robin Chan
|
Katsiaryna Mirylenka
|
Thomas Gschwind
|
Christoph Miksovic
|
Paolo Scotton
|
Enrico Toniato
|
Abdel Labbi
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: Industry Track
API integration is crucial for enterprise systems, as it enables seamless interaction between applications within workflows. However, the diversity and complexity of the API landscape present significant challenges in combining API calls based on user intent.Existing methods rely on named entity recognition (NER) and knowledge graphs, but struggle to generate more complex control flow structures, such as conditionals and loops.We propose a novel framework that leverages the success of large language models (LLMs) in code generation to integrate APIs based on natural language input. Our approach involves fine-tuning an LLM using automatically generated API flows derived from OpenAPI specifications.We further evaluate the effectiveness of enforcing the syntax and schema adherence through constrained decoding.To enable systematic comparison, we introduce targeted test suites to assess the generalization capabilities of these approaches and their ability to retain structured knowledge.Our findings show that LLMs fine-tuned on OpenAPI specifications can (a) learn structural API constraints implicitly during training, and (b) achieve significant improvements in both in-distribution and out-of-distribution performance over NER and retrieval-augmented generation (RAG)-based approaches.
Search
Fix author
Co-authors
- Shramona Chakraborty 1
- Robin Chan 1
- Nitin Gupta 1
- Abdel Labbi 1
- Sameep Mehta 1
- show all...