Robert Gmyr
2024
i-Code V2: An Autoregressive Generation Framework over Vision, Language, and Speech Data
Ziyi Yang
|
Mahmoud Khademi
|
Yichong Xu
|
Reid Pryzant
|
Yuwei Fang
|
Chenguang Zhu
|
Dongdong Chen
|
Yao Qian
|
Xuemei Gao
|
Yi-Ling Chen
|
Robert Gmyr
|
Naoyuki Kanda
|
Noel Codella
|
Bin Xiao
|
Yu Shi
|
Lu Yuan
|
Takuya Yoshioka
|
Michael Zeng
|
Xuedong Huang
Findings of the Association for Computational Linguistics: NAACL 2024
The convergence of text, visual, and audio data is crucial towards human-like artificial intelligence, however the current Vision-Language-Speech landscape is dominated by encoder-only models that lack generative abilities. We propose closing this gap with i-Code V2, one of the first models capable of generating natural language from any combination of Vision, Language, and Speech data. i-Code V2 leverages state-of-the-art single-modality encoders, combining their outputs with a new modality-fusing encoder to project combinations of modalities into a shared representational space. Language tokens are generated from these representations via an autoregressive decoder. i-Code V2 is pretrained end-to-end on a large collection of dual- and single-modality datasets with a novel text completion objective that can be generalized across arbitrary combinations of modalities. i-Code V2 matches or outperforms state-of-the-art single- and dual-modality baselines on 7 multimodal tasks, demonstrating the power of generative multimodal pretraining across a diversity of tasks and signals.
2020
TED: A Pretrained Unsupervised Summarization Model with Theme Modeling and Denoising
Ziyi Yang
|
Chenguang Zhu
|
Robert Gmyr
|
Michael Zeng
|
Xuedong Huang
|
Eric Darve
Findings of the Association for Computational Linguistics: EMNLP 2020
Text summarization aims to extract essential information from a piece of text and transform the text into a concise version. Existing unsupervised abstractive summarization models leverage recurrent neural networks framework while the recently proposed transformer exhibits much more capability. Moreover, most of previous summarization models ignore abundant unlabeled corpora resources available for pretraining. In order to address these issues, we propose TED, a transformer-based unsupervised abstractive summarization system with pretraining on large-scale data. We first leverage the lead bias in news articles to pretrain the model on millions of unlabeled corpora. Next, we finetune TED on target domains through theme modeling and a denoising autoencoder to enhance the quality of generated summaries. Notably, TED outperforms all unsupervised abstractive baselines on NYT, CNN/DM and English Gigaword datasets with various document styles. Further analysis shows that the summaries generated by TED are highly abstractive, and each component in the objective function of TED is highly effective.
Search
Co-authors
- Ziyi Yang 2
- Chenguang Zhu 2
- Michael Zeng 2
- Xuedong Huang 2
- Eric Darve 1
- show all...