Yanting Li
2025
Creolization versus code-switching: An agent-based cognitive model for bilingual strategies in language contact
Charles John Torres | Weijie Xu | Yanting Li | Richard Futrell
Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics
Charles John Torres | Weijie Xu | Yanting Li | Richard Futrell
Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics
Creolization and code-switching are closely related contact-induced linguistic phenomena, yet little attention has been paid to the connection between them. In this paper, we propose an agent-based cognitive model which provides a linkage between these two phenomena focusing on the statistical regularization of language use. That is, we identify that creolization as a conventionalization process and code-switching as flexible language choice can emerge from the same cognitive model in different social environments. Our model postulates a social structure of bilingual and monolingual populations, in which a set of agents seek for optimal communicative strategy shaped by multiple cognitive constraints. The simulation results show that our model successfully captures both phenomena as two ends of a continuum, characterized by varying degrees of regularization in the use of linguistic constructions from multiple source languages. The model also reveals a subtle dynamic between social structure and individual-level cognitive constraints.
LM2Protein: A Structure-to-Token Protein Large Language Model
Chang Zhou | Yuheng Shan | Pengan Chen | Xiangyu Shi | Zikang Wang | Yanting Li | Jiyue Jiang
Findings of the Association for Computational Linguistics: EMNLP 2025
Chang Zhou | Yuheng Shan | Pengan Chen | Xiangyu Shi | Zikang Wang | Yanting Li | Jiyue Jiang
Findings of the Association for Computational Linguistics: EMNLP 2025
Proteins are critical for various molecular functions, relying on their precise tertiary structures. This structure-sequence relationship is complex and degenerate, meaning multiple sequences can fold into a similar structure. The challenges in protein prediction, design, and modification increase with sequence complexity, while research on RNA-protein interactions, especially RNA-binding proteins (RBPs), is gaining importance. Large-scale pre-trained language models (LLMs) have shown promising results in handling biological sequences by treating them as natural language, though integrating spatial structures remains complex due to the need for specialized visual and 3D modeling approaches. We introduce a method to integrate protein 3D structural data within a sequence processing framework, converting 3D coordinates into discrete structure tokens using a VQ-VAE-like network. This simplifies the handling of 3D data, avoiding complex pipelines and facilitating a unified sequence-to-sequence processing model. Our approach demonstrates strong performance across a range of tasks, achieving high sequence recovery in inverse folding and protein-conditioned RNA design. These outstanding results demonstrate significant potential for application in complex biological systems research.
2024
On the communicative utility of code-switching
Yanting Li | Gregory Scontras | Richard Rutrell
Proceedings of the Society for Computation in Linguistics 2024
Yanting Li | Gregory Scontras | Richard Rutrell
Proceedings of the Society for Computation in Linguistics 2024
2021
Investigating Transfer Learning in Multilingual Pre-trained Language Models through Chinese Natural Language Inference
Hai Hu | He Zhou | Zuoyu Tian | Yiwen Zhang | Yina Patterson | Yanting Li | Yixin Nie | Kyle Richardson
Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021
Hai Hu | He Zhou | Zuoyu Tian | Yiwen Zhang | Yina Patterson | Yanting Li | Yixin Nie | Kyle Richardson
Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021
2020
CLUE: A Chinese Language Understanding Evaluation Benchmark
Liang Xu | Hai Hu | Xuanwei Zhang | Lu Li | Chenjie Cao | Yudong Li | Yechen Xu | Kai Sun | Dian Yu | Cong Yu | Yin Tian | Qianqian Dong | Weitang Liu | Bo Shi | Yiming Cui | Junyi Li | Jun Zeng | Rongzhao Wang | Weijian Xie | Yanting Li | Yina Patterson | Zuoyu Tian | Yiwen Zhang | He Zhou | Shaoweihua Liu | Zhe Zhao | Qipeng Zhao | Cong Yue | Xinrui Zhang | Zhengliang Yang | Kyle Richardson | Zhenzhong Lan
Proceedings of the 28th International Conference on Computational Linguistics
Liang Xu | Hai Hu | Xuanwei Zhang | Lu Li | Chenjie Cao | Yudong Li | Yechen Xu | Kai Sun | Dian Yu | Cong Yu | Yin Tian | Qianqian Dong | Weitang Liu | Bo Shi | Yiming Cui | Junyi Li | Jun Zeng | Rongzhao Wang | Weijian Xie | Yanting Li | Yina Patterson | Zuoyu Tian | Yiwen Zhang | He Zhou | Shaoweihua Liu | Zhe Zhao | Qipeng Zhao | Cong Yue | Xinrui Zhang | Zhengliang Yang | Kyle Richardson | Zhenzhong Lan
Proceedings of the 28th International Conference on Computational Linguistics
The advent of natural language understanding (NLU) benchmarks for English, such as GLUE and SuperGLUE allows new NLU models to be evaluated across a diverse set of tasks. These comprehensive benchmarks have facilitated a broad range of research and applications in natural language processing (NLP). The problem, however, is that most such benchmarks are limited to English, which has made it difficult to replicate many of the successes in English NLU for other languages. To help remedy this issue, we introduce the first large-scale Chinese Language Understanding Evaluation (CLUE) benchmark. CLUE is an open-ended, community-driven project that brings together 9 tasks spanning several well-established single-sentence/sentence-pair classification tasks, as well as machine reading comprehension, all on original Chinese text. To establish results on these tasks, we report scores using an exhaustive set of current state-of-the-art pre-trained Chinese models (9 in total). We also introduce a number of supplementary datasets and additional tools to help facilitate further progress on Chinese NLU. Our benchmark is released at https://www.cluebenchmarks.com
Search
Fix author
Co-authors
- Hai Hu 3
- Yina Patterson 3
- Zuoyu Tian 3
- Yiwen Zhang 3
- He Zhou 3
- Kyle Richardson 2
- Chenjie Cao 1
- Pengan Chen 1
- Yiming Cui 1
- Qianqian Dong 1
- Richard Futrell 1
- Jiyue Jiang 1
- Sandra Kübler 1
- Zhenzhong Lan 1
- Lu Li 1
- Yudong Li 1
- Junyi Li 1
- Chien-Jer Charles Lin 1
- Weitang Liu 1
- Shaoweihua Liu 1
- Yixin Nie 1
- Richard Rutrell 1
- Gregory Scontras 1
- Yuheng Shan 1
- Bo Shi 1
- Xiangyu Shi 1
- Kai Sun 1
- Yin Tian 1
- Charles John Torres 1
- Rongzhao Wang 1
- Zikang Wang 1
- Weijian Xie 1
- Liang Xu 1
- Yechen Xu 1
- Weijie Xu 1
- Zhengliang Yang 1
- Dian Yu 1
- Cong Yu 1
- Cong Yue 1
- Jun Zeng 1
- Xuanwei Zhang 1
- Xinrui Zhang 1
- Zhe Zhao 1
- Qipeng Zhao 1
- Chang Zhou 1