Lavanya Shankar
2025
JHU’s Submission to the AmericasNLP 2025 Shared Task on the Creation of Educational Materials for Indigenous Languages
Tom Lupicki
|
Lavanya Shankar
|
Kaavya Chaparala
|
David Yarowsky
Proceedings of the Fifth Workshop on NLP for Indigenous Languages of the Americas (AmericasNLP)
This paper presents JHU’s submission to the AmericasNLP shared task on the creation of educational materials for Indigenous languages. The task involves transforming a base sentence given one or more tags that correspond to grammatical features, such as negation or tense. The task also spans four languages: Bribri, Maya, Guaraní, and Nahuatl. We experiment with augmenting prompts to large language models with different information, chain of thought prompting, ensembling large language models by majority voting, and training a pointer-generator network. Our System 1, an ensemble of large language models, achieves the best performance on Maya and Guaraní, building upon the previous successes in leveraging large language models for this task and highlighting the effectiveness of ensembling large language models.
JHU IWSLT 2025 Low-resource System Description
Nathaniel Romney Robinson
|
Niyati Bafna
|
Xiluo He
|
Tom Lupicki
|
Lavanya Shankar
|
Cihan Xiao
|
Qi Sun
|
Kenton Murray
|
David Yarowsky
Proceedings of the 22nd International Conference on Spoken Language Translation (IWSLT 2025)
We present the Johns Hopkins University’s submission to the 2025 IWSLT Low-Resource Task. We competed on all 10 language pairs. Our approach centers around ensembling methods – specifically Minimum Bayes Risk Decoding. We find that such ensembling often improves performance only slightly over the best performing stand-alone model, and that in some cases it can even hurt performance slightly.
Search
Fix author
Co-authors
- Tom Lupicki 2
- David Yarowsky 2
- Niyati Bafna 1
- Kaavya Chaparala 1
- Xiluo He 1
- show all...