@inproceedings{mi-etal-2025-language,
    title = "Language-to-Space Programming for Training-Free 3{D} Visual Grounding",
    author = "Mi, Boyu  and
      Wang, Hanqing  and
      Wang, Tai  and
      Chen, Yilun  and
      Pang, Jiangmiao",
    editor = "Christodoulopoulos, Christos  and
      Chakraborty, Tanmoy  and
      Rose, Carolyn  and
      Peng, Violet",
    booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
    month = nov,
    year = "2025",
    address = "Suzhou, China",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.191/",
    pages = "3844--3864",
    ISBN = "979-8-89176-332-6",
    abstract = "3D visual grounding (3DVG) is challenging due to the need to understand 3D spatial relations. While supervised approaches have achieved superior performance, they are constrained by the scarcity and high annotation costs of 3D vision-language datasets. Training-free approaches based on LLMs/VLMs eliminate the need for large-scale training data, but they either incur prohibitive grounding time and token costs or have unsatisfactory accuracy. To address the challenges, we introduce a novel method for training-free 3D visual grounding, namely **La**nguage-to-**S**pace **P**rogramming (LaSP). LaSP introduces LLM-generated codes to analyze 3D spatial relations among objects, along with a pipeline that evaluates and optimizes the codes automatically. Experimental results demonstrate that LaSP achieves 52.9{\%} accuracy on the Nr3D benchmark, ranking among the best training-free methods. Moreover, it substantially reduces the grounding time and token costs, offering a balanced trade-off between performance and efficiency."
}Markdown (Informal)
[Language-to-Space Programming for Training-Free 3D Visual Grounding](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.191/) (Mi et al., EMNLP 2025)
ACL