@inproceedings{chuang-etal-2025-judging,
title = "Judging It, Washing It: Scoring and Greenwashing Corporate Climate Disclosures using Large Language Models",
author = "Chuang, Marianne and
Chuang, Gabriel and
Chuang, Cheryl and
Chuang, John",
editor = "Dutia, Kalyan and
Henderson, Peter and
Leippold, Markus and
Manning, Christoper and
Morio, Gaku and
Muccione, Veruska and
Ni, Jingwei and
Schimanski, Tobias and
Stammbach, Dominik and
Singh, Alok and
Su, Alba (Ruiran) and
A. Vaghefi, Saeid",
booktitle = "Proceedings of the 2nd Workshop on Natural Language Processing Meets Climate Change (ClimateNLP 2025)",
month = jul,
year = "2025",
address = "Bangkok, Thailand",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/mtsummit-25-ingestion/2025.climatenlp-1.2/",
doi = "10.18653/v1/2025.climatenlp-1.2",
pages = "17--31",
ISBN = "979-8-89176-259-6",
abstract = "We study the use of large language models (LLMs) to both evaluate and greenwash corporate climate disclosures. First, we investigate the use of the LLM-as-a-Judge (LLMJ) methodology for scoring company-submitted reports on emissions reduction targets and progress. Second, we probe the behavior of an LLM when it is prompted to greenwash a response subject to accuracy and length constraints. Finally, we test the robustness of the LLMJ methodology against responses that may be greenwashed using an LLM. We find that two LLMJ scoring systems, numerical rating and pairwise comparison, are effective in distinguishing high-performing companies from others, with the pairwise comparison system showing greater robustness against LLM-greenwashed responses."
}
Markdown (Informal)
[Judging It, Washing It: Scoring and Greenwashing Corporate Climate Disclosures using Large Language Models](https://preview.aclanthology.org/mtsummit-25-ingestion/2025.climatenlp-1.2/) (Chuang et al., ClimateNLP 2025)
ACL