@inproceedings{aytes-etal-2025-sketch,
    title = "Sketch-of-Thought: Efficient {LLM} Reasoning with Adaptive Cognitive-Inspired Sketching",
    author = "Aytes, Simon A.  and
      Baek, Jinheon  and
      Hwang, Sung Ju",
    editor = "Christodoulopoulos, Christos  and
      Chakraborty, Tanmoy  and
      Rose, Carolyn  and
      Peng, Violet",
    booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
    month = nov,
    year = "2025",
    address = "Suzhou, China",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1236/",
    pages = "24307--24331",
    ISBN = "979-8-89176-332-6",
    abstract = "Recent advances in large language models (LLMs) have enabled strong reasoning capabilities through Chain-of-Thought (CoT) prompting, which elicits step-by-step problem solving, but often at the cost of excessive verbosity in intermediate outputs, leading to increased computational overhead. We propose Sketch-of-Thought (SoT), a prompting framework that integrates cognitively inspired reasoning paradigms with linguistic constraints to reduce token usage while preserving reasoning accuracy. SoT is designed as a flexible, modular approach and is instantiated with three paradigms{---}Conceptual Chaining, Chunked Symbolism, and Expert Lexicons{---}each tailored to distinct reasoning tasks and selected dynamically at test-time by a lightweight routing model. Across 18 reasoning datasets spanning multiple domains, languages, and modalities, SoT achieves token reductions of up to 84{\%} with minimal accuracy loss. In tasks such as mathematical and multi-hop reasoning, it even improves accuracy while shortening outputs."
}Markdown (Informal)
[Sketch-of-Thought: Efficient LLM Reasoning with Adaptive Cognitive-Inspired Sketching](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1236/) (Aytes et al., EMNLP 2025)
ACL