@inproceedings{chi-etal-2024-attention, title = "Attention Alignment and Flexible Positional Embeddings Improve Transformer Length Extrapolation", author = "Chi, Ta-Chung and Fan, Ting-Han and Rudnicky, Alexander", editor = "Duh, Kevin and Gomez, Helena and Bethard, Steven", booktitle = "Findings of the Association for Computational Linguistics: NAACL 2024", month = jun, year = "2024", address = "Mexico City, Mexico", publisher = "Association for Computational Linguistics", url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-naacl.10/", doi = "10.18653/v1/2024.findings-naacl.10", pages = "132--148" }