2022
pdf
abs
Training Text-to-Text Transformers with Privacy Guarantees
Natalia Ponomareva
|
Jasmijn Bastings
|
Sergei Vassilvitskii
Findings of the Association for Computational Linguistics: ACL 2022
Recent advances in NLP often stem from large transformer-based pre-trained models, which rapidly grow in size and use more and more training data. Such models are often released to the public so that end users can fine-tune them on a task dataset. While it is common to treat pre-training data as public, it may still contain personally identifiable information (PII), such as names, phone numbers, and copyrighted material. Recent findings show that the capacity of these models allows them to memorize parts of the training data, and suggest differentially private (DP) training as a potential mitigation. While there is recent work on DP fine-tuning of NLP models, the effects of DP pre-training are less well understood: it is not clear how downstream performance is affected by DP pre-training, and whether DP pre-training mitigates some of the memorization concerns. We focus on T5 and show that by using recent advances in JAX and XLA we can train models with DP that do not suffer a large drop in pre-training utility, nor in training speed, and can still be fine-tuned to high accuracies on downstream tasks (e.g. GLUE). Moreover, we show that T5’s span corruption is a good defense against data memorization.
pdf
abs
Training Text-to-Text Transformers with Privacy Guarantees
Natalia Ponomareva
|
Jasmijn Bastings
|
Sergei Vassilvitskii
Proceedings of the Fourth Workshop on Privacy in Natural Language Processing
Recent advances in NLP often stem from large transformer-based pre-trained models, which rapidly grow in size and use more and more training data. Such models are often released to the public so that end users can fine-tune them on a task dataset. While it is common to treat pre-training data as public, it may still contain personally identifiable information (PII), such as names, phone numbers, and copyrighted material. Recent findings show that the capacity of these models allows them to memorize parts of the training data, and suggest differentially private (DP) training as a potential mitigation. While there is recent work on DP fine-tuning of NLP models, the effects of DP pre-training are less well understood it is not clear how downstream performance is affected by DP pre-training, and whether DP pre-training mitigates some of the memorization concerns. We focus on T5 and show that by using recent advances in JAX and XLA we can train models with DP that do not suffer a large drop in pre-training utility, nor in training speed, and can still be fine-tuned to high accuracy on downstream tasks (e.g. GLUE). Moreover, we show that T5s span corruption is a good defense against data memorization.
2019
pdf
abs
A Survey of the Perceived Text Adaptation Needs of Adults with Autism
Victoria Yaneva
|
Constantin Orasan
|
Le An Ha
|
Natalia Ponomareva
Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP 2019)
NLP approaches to automatic text adaptation often rely on user-need guidelines which are generic and do not account for the differences between various types of target groups. One such group are adults with high-functioning autism, who are usually able to read long sentences and comprehend difficult words but whose comprehension may be impeded by other linguistic constructions. This is especially challenging for real-world user-generated texts such as product reviews, which cannot be controlled editorially and are thus a particularly good applcation for automatic text adaptation systems. In this paper we present a mixed-methods survey conducted with 24 adult web-users diagnosed with autism and an age-matched control group of 33 neurotypical participants. The aim of the survey was to identify whether the group with autism experienced any barriers when reading online reviews, what these potential barriers were, and what NLP methods would be best suited to improve the accessibility of online reviews for people with autism. The group with autism consistently reported significantly greater difficulties with understanding online product reviews compared to the control group and identified issues related to text length, poor topic organisation, and the use of irony and sarcasm.
2013
pdf
Semi-supervised vs. Cross-domain Graphs for Sentiment Analysis
Natalia Ponomareva
|
Mike Thelwall
Proceedings of the International Conference Recent Advances in Natural Language Processing RANLP 2013
2012
pdf
Do Neighbours Help? An Exploration of Graph-based Algorithms for Cross-domain Sentiment Classification
Natalia Ponomareva
|
Mike Thelwall
Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning
2009
pdf
QALL-ME needs AIR: a portability study
Constantin Orăsan
|
Iustin Dornescu
|
Natalia Ponomareva
Proceedings of the Workshop on Adaptation of Language Resources and Technology to New Domains