Xiaofeng Yang


2025

pdf bib
HELENE: Hessian Layer-wise Clipping and Gradient Annealing for Accelerating Fine-tuning LLM with Zeroth-order Optimization
Huaqin Zhao | Jiaxi Li | Yi Pan | Shizhe Liang | Xiaofeng Yang | Fei Dou | Tianming Liu | Jin Lu
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing

Fine-tuning large language models (LLMs) faces significant memory challenges due to the high cost of back-propagation. MeZO addresses this using zeroth-order (ZO) optimization, matching memory usage to inference but suffering from slow convergence due to varying curvatures across model parameters. To overcome this limitation, We propose HELENE, a scalable and memory-efficient optimizer that integrates annealed A-GNB gradients with diagonal Hessian estimation and layer-wise clipping as a second-order pre-conditioner. HELENE provably accelerates and stabilizes convergence by reducing dependence on total parameter space and scaling with the largest layer dimension. Experiments on RoBERTa-large and OPT-1.3B show up to a 20× speedup over MeZO with an average accuracy improvement of 1.5%. HELENE supports full and parameter-efficient fine-tuning, outperforming several state-of-the-art optimizers.

2008

pdf bib
Other-Anaphora Resolution in Biomedical Texts with Automatically Mined Patterns
Bin Chen | Xiaofeng Yang | Jian Su | Chew Lim Tan
Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008)

pdf bib
Coreference Systems Based on Kernels Methods
Yannick Versley | Alessandro Moschitti | Massimo Poesio | Xiaofeng Yang
Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008)

pdf bib
A Twin-Candidate Model for Learning-Based Anaphora Resolution
Xiaofeng Yang | Jian Su | Chew Lim Tan
Computational Linguistics, Volume 34, Number 3, September 2008

pdf bib
BART: A modular toolkit for coreference resolution
Yannick Versley | Simone Ponzetto | Massimo Poesio | Vladimir Eidelman | Alan Jern | Jason Smith | Xiaofeng Yang | Alessandro Moschitti
Proceedings of the Sixth International Conference on Language Resources and Evaluation (LREC'08)

Developing a full coreference system able to run all the way from raw text to semantic interpretation is a considerable engineering effort. Accordingly, there is very limited availability of off-the shelf tools for researchers whose interests are not primarily in coreference or others who want to concentrate on a specific aspect of the problem. We present BART, a highly modular toolkit for developing coreference applications. In the Johns Hopkins workshop on using lexical and encyclopedic knowledge for entity disambiguation, the toolkit was used to extend a reimplementation of Soon et al.’s proposal with a variety of additional syntactic and knowledge-based features, and experiment with alternative resolution processes, preprocessing tools, and classifiers. BART has been released as open source software and is available from http://www.sfs.uni-tuebingen.de/~versley/BART

pdf bib
An Entity-Mention Model for Coreference Resolution with Inductive Logic Programming
Xiaofeng Yang | Jian Su | Jun Lang | Chew Lim Tan | Ting Liu | Sheng Li
Proceedings of ACL-08: HLT

pdf bib
BART: A Modular Toolkit for Coreference Resolution
Yannick Versley | Simone Paolo Ponzetto | Massimo Poesio | Vladimir Eidelman | Alan Jern | Jason Smith | Xiaofeng Yang | Alessandro Moschitti
Proceedings of the ACL-08: HLT Demo Session

2007

pdf bib
Coreference Resolution Using Semantic Relatedness Information from Automatically Discovered Patterns
Xiaofeng Yang | Jian Su
Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics

2006

pdf bib
Kernel-Based Pronoun Resolution with Structured Syntactic Knowledge
Xiaofeng Yang | Jian Su | Chew Lim Tan
Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics

2005

pdf bib
A Twin-Candidate Model of Coreference Resolution with Non-Anaphor Identification Capability
Xiaofeng Yang | Jian Su | Chew Lim Tan
Second International Joint Conference on Natural Language Processing: Full Papers

pdf bib
Improving Pronoun Resolution Using Statistics-Based Semantic Compatibility Information
Xiaofeng Yang | Jian Su | Chew Lim Tan
Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics (ACL’05)

2004

pdf bib
An NP-Cluster Based Approach to Coreference Resolution
Xiaofeng Yang | Jian Su | GuoDong Zhou | Chew Lim Tan
COLING 2004: Proceedings of the 20th International Conference on Computational Linguistics

pdf bib
Improving Pronoun Resolution by Incorporating Coreferential Information of Candidates
Xiaofeng Yang | Jian Su | Guodong Zhou | Chew-Lim Tan
Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics (ACL-04)

2003

pdf bib
Coreference Resolution Using Competition Learning Approach
Xiaofeng Yang | Guodong Zhou | Jian Su | Chew Lim Tan
Proceedings of the 41st Annual Meeting of the Association for Computational Linguistics

2002

pdf bib
一種基於知網的語義排歧模型研究 (A Study of Semantic Disambiguation Based on HowNet) [In Chinese]
Xiaofeng Yang | Tangqiu Li
International Journal of Computational Linguistics & Chinese Language Processing, Volume 7, Number 1, February 2002: Special Issue on HowNet and Its Applications