|Roberta: A robustly optimized bert pretraining approach|
Y Liu, M Ott, N Goyal, J Du, M Joshi, D Chen, O Levy, M Lewis, ...
arXiv preprint arXiv:1907.11692, 2019
|Spanbert: Improving pre-training by representing and predicting spans|
M Joshi, D Chen, Y Liu, DS Weld, L Zettlemoyer, O Levy
Transactions of the Association for Computational Linguistics 8, 64-77, 2020
|Cloze-driven pretraining of self-attention networks|
A Baevski, S Edunov, Y Liu, L Zettlemoyer, M Auli
arXiv preprint arXiv:1903.07785, 2019
|Mask-predict: Parallel decoding of conditional masked language models|
M Ghazvininejad, O Levy, Y Liu, L Zettlemoyer
Proceedings of the 2019 Conference on Empirical Methods in Natural Language …, 2019
|Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension|
M Lewis, Y Liu, N Goyal, M Ghazvininejad, A Mohamed, O Levy, ...
arXiv preprint arXiv:1910.13461, 2019
|Chromatographic peak alignment using derivative dynamic time warping|
C Bork, K Ng, Y Liu, A Yee, M Pohlscheidt
Biotechnology progress 29 (2), 394-402, 2013
|Multilingual Denoising Pre-training for Neural Machine Translation|
Y Liu, J Gu, N Goyal, X Li, S Edunov, M Ghazvininejad, M Lewis, ...
arXiv preprint arXiv:2001.08210, 2020