Exploring Chain-of-Thought Reasoning for Steerable Pluralistic Alignment

Yunfan Zhang, Kathleen McKeown, Smaranda Muresan


Abstract
Large Language Models (LLMs) are typically trained to reflect a relatively uniform set of values, which limits their applicability to tasks that require understanding of nuanced human perspectives. Recent research has underscored the importance of enabling LLMs to support steerable pluralism — the capacity to adopt a specific perspective and align generated outputs with it. In this work, we investigate whether Chain-of-Thought (CoT) reasoning techniques can be applied to building steerable pluralistic models. We explore several methods, including CoT prompting, fine-tuning on human-authored CoT, fine-tuning on synthetic explanations, and Reinforcement Learning with Verifiable Rewards (RLVR). We evaluate these approaches using the Value Kaleidoscope and OpinionQA datasets. Among the methods studied, RLVR consistently outperforms others and demonstrates strong training sample efficiency. We further analyze the generated CoT traces with respect to faithfulness and safety.
Anthology ID:
2025.emnlp-main.1301
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
25647–25660
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1301/
DOI:
Bibkey:
Cite (ACL):
Yunfan Zhang, Kathleen McKeown, and Smaranda Muresan. 2025. Exploring Chain-of-Thought Reasoning for Steerable Pluralistic Alignment. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 25647–25660, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Exploring Chain-of-Thought Reasoning for Steerable Pluralistic Alignment (Zhang et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1301.pdf
Checklist:
 2025.emnlp-main.1301.checklist.pdf