This is a brief summary of paper for me to study and organize it, BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension (Lewis et al., ACL 2020) that I read and studied.
Tho following is the material of my paper seminar on BART which is composed by me.
It consists of two types of presentation, 1) detailed version presentation and 2) short version presentation.
The below has video of the author presentation.
I hope someone who want to understand what is the BART and pre-training in natural language processing field
For detailed experiment analysis, you can found in BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension (Lewis et al., ACL 2020)
Reference
- Paper
- arXiv Version: BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension (Lewis et al., arXiv 2020)
- ACL Version: BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension (Lewis et al., ACL 2020)
- How to use html for alert
- For your information
- BART Presentation Video in ACL 2020
- Understand how the XLNet outperforms BERT in Language Modelling
- The Illustrated Transformer in Jay Alammar blog
- Layer Normalization in paperswithcode
- What is XLNet and why it outperforms BERT in towards data science by LIANG
- What is Two-Stream Self Attention in XLNet in towards data science by LIANG