@inproceedings{carreto-fidalgo-etal-2021-system,
title = "System description for {P}rof{NER} - {SMMH}: Optimized finetuning of a pretrained transformer and word vectors",
author = "Carreto Fidalgo, David and
Vila-Suero, Daniel and
Aranda Montes, Francisco and
Talavera Cepeda, Ignacio",
booktitle = "Proceedings of the Sixth Social Media Mining for Health ({\#}SMM4H) Workshop and Shared Task",
month = jun,
year = "2021",
address = "Mexico City, Mexico",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.smm4h-1.11",
doi = "10.18653/v1/2021.smm4h-1.11",
pages = "69--73",
abstract = "This shared task system description depicts two neural network architectures submitted to the ProfNER track, among them the winning system that scored highest in the two sub-tasks 7a and 7b. We present in detail the approach, preprocessing steps and the architectures used to achieve the submitted results, and also provide a GitHub repository to reproduce the scores. The winning system is based on a transformer-based pretrained language model and solves the two sub-tasks simultaneously.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="carreto-fidalgo-etal-2021-system">
<titleInfo>
<title>System description for ProfNER - SMMH: Optimized finetuning of a pretrained transformer and word vectors</title>
</titleInfo>
<name type="personal">
<namePart type="given">David</namePart>
<namePart type="family">Carreto Fidalgo</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Daniel</namePart>
<namePart type="family">Vila-Suero</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Francisco</namePart>
<namePart type="family">Aranda Montes</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Ignacio</namePart>
<namePart type="family">Talavera Cepeda</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2021-jun</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Sixth Social Media Mining for Health (#SMM4H) Workshop and Shared Task</title>
</titleInfo>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Mexico City, Mexico</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>This shared task system description depicts two neural network architectures submitted to the ProfNER track, among them the winning system that scored highest in the two sub-tasks 7a and 7b. We present in detail the approach, preprocessing steps and the architectures used to achieve the submitted results, and also provide a GitHub repository to reproduce the scores. The winning system is based on a transformer-based pretrained language model and solves the two sub-tasks simultaneously.</abstract>
<identifier type="citekey">carreto-fidalgo-etal-2021-system</identifier>
<identifier type="doi">10.18653/v1/2021.smm4h-1.11</identifier>
<location>
<url>https://aclanthology.org/2021.smm4h-1.11</url>
</location>
<part>
<date>2021-jun</date>
<extent unit="page">
<start>69</start>
<end>73</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T System description for ProfNER - SMMH: Optimized finetuning of a pretrained transformer and word vectors
%A Carreto Fidalgo, David
%A Vila-Suero, Daniel
%A Aranda Montes, Francisco
%A Talavera Cepeda, Ignacio
%S Proceedings of the Sixth Social Media Mining for Health (#SMM4H) Workshop and Shared Task
%D 2021
%8 jun
%I Association for Computational Linguistics
%C Mexico City, Mexico
%F carreto-fidalgo-etal-2021-system
%X This shared task system description depicts two neural network architectures submitted to the ProfNER track, among them the winning system that scored highest in the two sub-tasks 7a and 7b. We present in detail the approach, preprocessing steps and the architectures used to achieve the submitted results, and also provide a GitHub repository to reproduce the scores. The winning system is based on a transformer-based pretrained language model and solves the two sub-tasks simultaneously.
%R 10.18653/v1/2021.smm4h-1.11
%U https://aclanthology.org/2021.smm4h-1.11
%U https://doi.org/10.18653/v1/2021.smm4h-1.11
%P 69-73
Markdown (Informal)
[System description for ProfNER - SMMH: Optimized finetuning of a pretrained transformer and word vectors](https://aclanthology.org/2021.smm4h-1.11) (Carreto Fidalgo et al., SMM4H 2021)
ACL