Sangmook Lee
2025
Conditional [MASK] Discrete Diffusion Language Model
Hyukhun Koh
|
Minha Jhang
|
Dohyung Kim
|
Sangmook Lee
|
Kyomin Jung
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Although auto-regressive models excel in natural language processing, they often struggle to generate diverse text and provide limited controllability. Non-auto-regressive methods could be an alternative but often produce degenerate outputs and exhibit shortcomings in conditional generation. To address these challenges, we propose Diffusion-EAGS, a novel framework that integrates conditional masked language models into diffusion language models through the theoretical lens of a conditional Markov Random Field. In doing so, we propose entropy-adaptive Gibbs sampling and entropy-based noise scheduling to counterbalance each model’s shortcomings. Experimental results show that Diffusion-EAGS outperforms baselines and achieves the best quality-diversity tradeoff, demonstrating its effectiveness in non-autoregressive text generation.
ReflAct: World-Grounded Decision Making in LLM Agents via Goal-State Reflection
Jeonghye Kim
|
Sojeong Rhee
|
Minbeom Kim
|
Dohyung Kim
|
Sangmook Lee
|
Youngchul Sung
|
Kyomin Jung
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Recent advances in LLM agents have largely built on reasoning backbones like ReAct, which interleave thought and action in complex environments. However, ReAct often produces ungrounded or incoherent reasoning steps, leading to misalignment between the agent’s actual state and goals. Our analysis finds that this stems from ReAct’s inability to maintain consistent internal beliefs and goal alignment, causing compounding errors and hallucinations. To address this, we introduce ReflAct, a novel backbone that shifts reasoning from merely planning next actions to continuously reflecting on the agent’s state relative to its goal. By explicitly grounding decisions in states and enforcing ongoing goal alignment, ReflAct dramatically improves strategic reliability. This design delivers substantial empirical gains: ReflAct surpasses ReAct by 27.7% on average, achieving a 93.3% success rate in ALFWorld. Notably, ReflAct even outperforms ReAct with added enhancement modules (e.g., Reflexion, WKM), showing that strengthening the core reasoning backbone is key to reliable agent performance.
Search
Fix author
Co-authors
- Kyomin Jung 2
- Dohyung Kim 2
- Minha Jhang 1
- Jeonghye Kim 1
- Minbeom Kim 1
- show all...