@inproceedings{chuang-etal-2025-judging,
    title = "Judging It, Washing It: Scoring and Greenwashing Corporate Climate Disclosures using Large Language Models",
    author = "Chuang, Marianne  and
      Chuang, Gabriel  and
      Chuang, Cheryl  and
      Chuang, John",
    editor = "Dutia, Kalyan  and
      Henderson, Peter  and
      Leippold, Markus  and
      Manning, Christoper  and
      Morio, Gaku  and
      Muccione, Veruska  and
      Ni, Jingwei  and
      Schimanski, Tobias  and
      Stammbach, Dominik  and
      Singh, Alok  and
      Su, Alba (Ruiran)  and
      A. Vaghefi, Saeid",
    booktitle = "Proceedings of the 2nd Workshop on Natural Language Processing Meets Climate Change (ClimateNLP 2025)",
    month = jul,
    year = "2025",
    address = "Vienna, Austria",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2025.climatenlp-1.2/",
    doi = "10.18653/v1/2025.climatenlp-1.2",
    pages = "17--31",
    ISBN = "979-8-89176-259-6",
    abstract = "We study the use of large language models (LLMs) to both evaluate and greenwash corporate climate disclosures. First, we investigate the use of the LLM-as-a-Judge (LLMJ) methodology for scoring company-submitted reports on emissions reduction targets and progress. Second, we probe the behavior of an LLM when it is prompted to greenwash a response subject to accuracy and length constraints. Finally, we test the robustness of the LLMJ methodology against responses that may be greenwashed using an LLM. We find that two LLMJ scoring systems, numerical rating and pairwise comparison, are effective in distinguishing high-performing companies from others, with the pairwise comparison system showing greater robustness against LLM-greenwashed responses."
}Markdown (Informal)
[Judging It, Washing It: Scoring and Greenwashing Corporate Climate Disclosures using Large Language Models](https://preview.aclanthology.org/ingest-emnlp/2025.climatenlp-1.2/) (Chuang et al., ClimateNLP 2025)
ACL