Houjun Liu
2025
Drop Dropout on Single Epoch Language Model Pretraining
Houjun Liu
|
John Bauer
|
Christopher D Manning
Findings of the Association for Computational Linguistics: ACL 2025
Originally, dropout was seen as a breakthrough regularization technique that reduced overfitting and improved performance in almost all applications of deep learning by reducing overfitting. Yet, single-epoch pretraining tasks common to modern LLMs yield minimal overfitting, leading to dropout not being used for large LLMs. Nevertheless, no thorough empirical investigation has been done on the role of dropout in LM pretraining. Through experiments in single-epoch pretraining of both masked (BERT) and autoregressive (Pythia 160M and 1.4B) LMs with varying levels of dropout, we find that downstream performance in language modeling, morpho-syntax (BLiMP), question answering (SQuAD), and natural-language inference (MNLI) improves when dropout is not applied during pretraining. We additionally find that the recently-introduced “early dropout” also degrades performance over applying no dropout at all. We further investigate the models’ editability, and find that models trained without dropout are more successful in gradient-based model editing (MEND) and equivalent in representation-based model editing (ReFT). Therefore, we advocate to **drop dropout** during single-epoch pretraining.
ASTPrompter: Preference-Aligned Automated Language Model Red-Teaming to Generate Low-Perplexity Unsafe Prompts
Amelia Hardy
|
Houjun Liu
|
Allie Griffith
|
Bernard Lange
|
Duncan Eddy
|
Mykel Kochenderfer
Findings of the Association for Computational Linguistics: EMNLP 2025
Existing LLM red-teaming approaches prioritize high attack success rate, often resulting in high-perplexity prompts. This focus overlooks low-perplexity attacks that are more difficult to filter, more likely to arise during benign usage, and more impactful as negative downstream training examples. In response, we introduce ASTPrompter, a single-step optimization method that uses contrastive preference learning to train an attacker to maintain low perplexity while achieving a high attack success rate (ASR). ASTPrompter achieves an attack success rate 5.1 times higher on Llama-8.1B while using inputs that are 2.1 times more likely to occur according to the frozen LLM. Furthermore, our attack transfers to Mistral-7B, Qwen-7B, and TinyLlama in both black- and white-box settings. Lastly, by tuning a single hyperparameter in our method, we discover successful attack prefixes along an efficient frontier between ASR and perplexity, highlighting perplexity as a previously under-considered factor in red-teaming.
2024
MSCAW-coref: Multilingual, Singleton and Conjunction-Aware Word-Level Coreference Resolution
Houjun Liu
|
John Bauer
|
Karel D’Oosterlinck
|
Christopher Potts
|
Christopher D. Manning
Proceedings of the Seventh Workshop on Computational Models of Reference, Anaphora and Coreference
Search
Fix author
Co-authors
- John Bauer 2
- Christopher D. Manning 2
- Karel D’Oosterlinck 1
- Duncan Eddy 1
- Allie Griffith 1
- show all...