Tyler Derr
2026
A Survey on LLM-based Conversational User Simulation
Bo Ni | Yu Wang | Leyao Wang | Branislav Kveton | Franck Dernoncourt | Yu Xia | Hongjie Chen | Reuben Luera | Samyadeep Basu | Subhojyoti Mukherjee | Puneet Mathur | Nesreen K. Ahmed | Junda Wu | Li Li | Huixin Zhang | Ruiyi Zhang | Tong Yu | Sungchul Kim | Jiuxiang Gu | Zhengzhong Tu | Alexa Siu | Zichao Wang | Seunghyun Yoon | Nedim Lipka | Namyong Park | Zihao Lin | Trung Bui | Yue Zhao | Tyler Derr | Ryan A. Rossi
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Bo Ni | Yu Wang | Leyao Wang | Branislav Kveton | Franck Dernoncourt | Yu Xia | Hongjie Chen | Reuben Luera | Samyadeep Basu | Subhojyoti Mukherjee | Puneet Mathur | Nesreen K. Ahmed | Junda Wu | Li Li | Huixin Zhang | Ruiyi Zhang | Tong Yu | Sungchul Kim | Jiuxiang Gu | Zhengzhong Tu | Alexa Siu | Zichao Wang | Seunghyun Yoon | Nedim Lipka | Namyong Park | Zihao Lin | Trung Bui | Yue Zhao | Tyler Derr | Ryan A. Rossi
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
User simulation has long played a vital role in computer science due to its potential to support a wide range of applications. Language, as the primary medium of human communication, forms the foundation of social interaction and behavior. Consequently, simulating conversational behavior has become a key area of study. Recent advancements in large language models (LLMs) have significantly catalyzed progress in this domain by enabling high-fidelity generation of synthetic user conversation. In this paper, we survey recent advancements in LLM-based conversational user simulation. We introduce a novel taxonomy covering user granularity and simulation objectives. Additionally, we systematically analyze core techniques and evaluation methodologies. We aim to keep the research community informed of the latest advancements in conversational user simulation and to further facilitate future research by identifying open challenges and organizing existing work under a unified framework.
Ensemble Privacy Defense for Knowledge-Intensive LLMs against Membership Inference Attacks
Haowei Fu | Bo Ni | Han Xu | Kunpeng Liu | Dan Lin | Tyler Derr
Findings of the Association for Computational Linguistics: EACL 2026
Haowei Fu | Bo Ni | Han Xu | Kunpeng Liu | Dan Lin | Tyler Derr
Findings of the Association for Computational Linguistics: EACL 2026
Retrieval-Augmented Generation (RAG) and Supervised Finetuning (SFT) have become the predominant paradigms for equipping Large Language Models (LLMs) with external knowledge for diverse, knowledge-intensive tasks. However, while such knowledge injection improves performance, it also exposes new attack surfaces. Membership Inference Attacks (MIAs), which aim to determine whether a given data sample was included in a model’s training set, pose serious threats to privacy and trust in sensitive domains. To this end, we first systematically evaluate the vulnerability of RAG- and SFT-based LLMs to various MIAs. Then, to address the privacy risk, we further introduce a novel, model-agnostic defense framework, Ensemble Privacy Defense (EPD), which aggregates and evaluates the outputs of a knowledge-injected LLM, a base LLM, and a dedicated judge model to enhance resistance against MIAs. Comprehensive experiments show that, on average, EPD reduces MIA success by up to 27.8% for SFT and 526.3% for RAG compared to inference-time baseline, while maintaining answer quality.
2025
Demystifying the Power of Large Language Models in Graph Generation
Yu Wang | Ryan A. Rossi | Namyong Park | Nesreen K. Ahmed | Danai Koutra | Franck Dernoncourt | Tyler Derr
Findings of the Association for Computational Linguistics: NAACL 2025
Yu Wang | Ryan A. Rossi | Namyong Park | Nesreen K. Ahmed | Danai Koutra | Franck Dernoncourt | Tyler Derr
Findings of the Association for Computational Linguistics: NAACL 2025
Despite the unprecedented success of applying Large Language Models (LLMs) to graph discriminative tasks such as node classification and link prediction, its potential for graph structure generation remains largely unexplored. To fill this crucial gap, this paper presents a systematic investigation into the capability of LLMs for graph structure generation. Specifically, we design prompts triggering LLMs to generate codes that optimize network properties by injecting domain expertise from network science. Since graphs in different domains exhibit unique structural properties captured by various metrics (e.g., clustering coefficient capturing triangles in social networks while squares reflecting road segments in transportation networks), we first evaluate the capability of LLMs to generate graphs satisfying each structural property in different domains. After that, we select the optimal property configurations and benchmark the graph structure generation performance of LLMs against established graph generative models across multiple domains. Our findings shed light on generating graph structures from an LLM perspective. Our code is publically available https://github.com/yuwvandy/LLM-GraphGen.
Search
Fix author
Co-authors
- Nesreen K. Ahmed 2
- Franck Dernoncourt 2
- Bo Ni 2
- Namyong Park 2
- Ryan A. Rossi 2
- Yu Wang 2
- Samyadeep Basu 1
- Trung Bui 1
- Hongjie Chen 1
- Haowei Fu 1
- Jiuxiang Gu 1
- Sungchul Kim 1
- Danai Koutra 1
- Branislav Kveton 1
- Li Li 1
- Zihao Lin 1
- Dan Lin 1
- Nedim Lipka 1
- Kunpeng Liu 1
- Reuben Luera 1
- Puneet Mathur 1
- Subhojyoti Mukherjee 1
- Alexa Siu 1
- Zhengzhong Tu 1
- Leyao Wang 1
- Zichao Wang 1
- Junda Wu 1
- Yu Xia 1
- Han Xu 1
- Seunghyun Yoon 1
- Tong Yu 1
- Huixin Zhang 1
- Ruiyi Zhang 1
- Yue Zhao 1