ACL2020论文阅读笔记:BART

背景 题目: BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension 机构:Facebook AI 作者:Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Ab
相关文章
相关标签/搜索