@inproceedings{mayfield-black-2020-fine,
title = "Should You Fine-Tune {BERT} for Automated Essay Scoring?",
author = "Mayfield, Elijah and
Black, Alan W",
booktitle = "Proceedings of the Fifteenth Workshop on Innovative Use of NLP for Building Educational Applications",
month = jul,
year = "2020",
address = "Seattle, WA, USA → Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2020.bea-1.15",
doi = "10.18653/v1/2020.bea-1.15",
pages = "151--162",
abstract = "Most natural language processing research now recommends large Transformer-based models with fine-tuning for supervised classification tasks; older strategies like bag-of-words features and linear models have fallen out of favor. Here we investigate whether, in automated essay scoring (AES) research, deep neural models are an appropriate technological choice. We find that fine-tuning BERT produces similar performance to classical models at significant additional cost. We argue that while state-of-the-art strategies do match existing best results, they come with opportunity costs in computational resources. We conclude with a review of promising areas for research on student essays where the unique characteristics of Transformers may provide benefits over classical methods to justify the costs.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="mayfield-black-2020-fine">
<titleInfo>
<title>Should You Fine-Tune BERT for Automated Essay Scoring?</title>
</titleInfo>
<name type="personal">
<namePart type="given">Elijah</namePart>
<namePart type="family">Mayfield</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Alan</namePart>
<namePart type="given">W</namePart>
<namePart type="family">Black</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2020-jul</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Fifteenth Workshop on Innovative Use of NLP for Building Educational Applications</title>
</titleInfo>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Seattle, WA, USA → Online</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Most natural language processing research now recommends large Transformer-based models with fine-tuning for supervised classification tasks; older strategies like bag-of-words features and linear models have fallen out of favor. Here we investigate whether, in automated essay scoring (AES) research, deep neural models are an appropriate technological choice. We find that fine-tuning BERT produces similar performance to classical models at significant additional cost. We argue that while state-of-the-art strategies do match existing best results, they come with opportunity costs in computational resources. We conclude with a review of promising areas for research on student essays where the unique characteristics of Transformers may provide benefits over classical methods to justify the costs.</abstract>
<identifier type="citekey">mayfield-black-2020-fine</identifier>
<identifier type="doi">10.18653/v1/2020.bea-1.15</identifier>
<location>
<url>https://aclanthology.org/2020.bea-1.15</url>
</location>
<part>
<date>2020-jul</date>
<extent unit="page">
<start>151</start>
<end>162</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Should You Fine-Tune BERT for Automated Essay Scoring?
%A Mayfield, Elijah
%A Black, Alan W.
%S Proceedings of the Fifteenth Workshop on Innovative Use of NLP for Building Educational Applications
%D 2020
%8 jul
%I Association for Computational Linguistics
%C Seattle, WA, USA → Online
%F mayfield-black-2020-fine
%X Most natural language processing research now recommends large Transformer-based models with fine-tuning for supervised classification tasks; older strategies like bag-of-words features and linear models have fallen out of favor. Here we investigate whether, in automated essay scoring (AES) research, deep neural models are an appropriate technological choice. We find that fine-tuning BERT produces similar performance to classical models at significant additional cost. We argue that while state-of-the-art strategies do match existing best results, they come with opportunity costs in computational resources. We conclude with a review of promising areas for research on student essays where the unique characteristics of Transformers may provide benefits over classical methods to justify the costs.
%R 10.18653/v1/2020.bea-1.15
%U https://aclanthology.org/2020.bea-1.15
%U https://doi.org/10.18653/v1/2020.bea-1.15
%P 151-162
Markdown (Informal)
[Should You Fine-Tune BERT for Automated Essay Scoring?](https://aclanthology.org/2020.bea-1.15) (Mayfield & Black, BEA 2020)
ACL
- Elijah Mayfield and Alan W Black. 2020. Should You Fine-Tune BERT for Automated Essay Scoring?. In Proceedings of the Fifteenth Workshop on Innovative Use of NLP for Building Educational Applications, pages 151–162, Seattle, WA, USA → Online. Association for Computational Linguistics.