Alex Dimakis
2025
Large Language Models as Realistic Microservice Trace Generators
Donghyun Kim
|
Sriram Ravula
|
Taemin Ha
|
Alex Dimakis
|
Daehyeok Kim
|
Aditya Akella
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Workload traces are essential to understand complex computer systems’ behavior and manage processing and memory resources. Since real-world traces are hard to obtain, synthetic trace generation is a promising alternative. This paper proposes a first-of-a-kind approach that relies on training a large language model (LLM) to generate synthetic workload traces, specifically microservice call graphs. To capture complex and arbitrary hierarchical structures and implicit constraints in such traces, we propose to train LLMs to generate recursively, making call graph generation a sequence of more manageable steps. To further enforce learning constraints on the traces and generate uncommon situations, we apply additional instruction tuning steps to align our model with the desired trace features. With this method, we train TraceLLM, an LLM for microservice trace generation, and demonstrate that it produces diverse, realistic traces under varied conditions, outperforming existing approaches in both accuracy and validity. The synthetically generated traces can effectively replace real data to optimize important microservice management tasks. Additionally, TraceLLM adapts to downstream trace-related tasks, such as predicting key trace features and infilling missing data.
2024
Which questions should I answer? Salience Prediction of Inquisitive Questions
Yating Wu
|
Ritika Rajesh Mangla
|
Alex Dimakis
|
Greg Durrett
|
Junyi Jessy Li
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Inquisitive questions — open-ended, curiosity-driven questions people ask as they read — are an integral part of discourse processing and comprehension. Recent work in NLP has taken advantage of question generation capabilities of LLMs to enhance a wide range of applications. But the space of inquisitive questions is vast: many questions can be evoked from a given context. So which of those should be prioritized to find answers? Linguistic theories, unfortunately, have not yet provided an answer to this question. This paper presents QSalience, a salience predictor of inquisitive questions. QSalience is instruction-tuned over our dataset of linguist-annotated salience scores of 1,766 (context, question) pairs. A question scores high on salience if answering it would greatly enhance the understanding of the text. We show that highly salient questions are empirically more likely to be answered in the same article, bridging potential questions with Questions Under Discussion. We further validate our findings by showing that answering salient questions is an indicator of summarization quality in news.
Search
Fix author
Co-authors
- Aditya Akella 1
- Greg Durrett 1
- Taemin Ha 1
- Donghyun Kim 1
- Daehyeok Kim 1
- show all...