SubmissionNumber#=%=#244 FinalPaperTitle#=%=#Weighted Layer Averaging RoBERTa for Black-Box Machine-Generated Text Detection ShortPaperTitle#=%=# NumberOfPages#=%=#4 CopyrightSigned#=%=#Ayan Datta JobTitle#==# Organization#==# Abstract#==#We propose a novel approach for machine-generated text detection using a RoBERTa model with weighted layer averaging and AdaLoRA for parameter-efficient fine-tuning. Our method incorporates information from all model layers, capturing diverse linguistic cues beyond those accessible from the final layer alone. To mitigate potential overfitting and improve generalizability, we leverage AdaLoRA, which injects trainable low-rank matrices into each Transformer layer, significantly reducing the number of trainable parameters. Furthermore, we employ data mixing to ensure our model encounters text from various domains and generators during training, enhancing its ability to generalize to unseen data. This work highlights the potential of combining layer-wise information with parameter-efficient fine-tuning and data mixing for effective machine-generated text detection. Author{1}{Firstname}#=%=#Ayan Author{1}{Lastname}#=%=#Datta Author{1}{Username}#=%=#advin4603 Author{1}{Email}#=%=#ayan.datta@research.iiit.ac.in Author{1}{Affiliation}#=%=#IIIT Hyderabad Author{2}{Firstname}#=%=#Aryan Ashok Author{2}{Lastname}#=%=#Chandramania Author{2}{Username}#=%=#aryanchandramania Author{2}{Email}#=%=#aryan.chandramania@research.iiit.ac.in Author{2}{Affiliation}#=%=#International Institute of Information Technology, Hyderabad Author{3}{Firstname}#=%=#Radhika Author{3}{Lastname}#=%=#Mamidi Author{3}{Username}#=%=#radhika Author{3}{Email}#=%=#radhika.mamidi@iiit.ac.in Author{3}{Affiliation}#=%=#Language Technologies Research Centre, IIIT Hyderabad ========== èéáğö