This website requires JavaScript.
Explore
Help
Register
Sign In
e2hang
/
Literature
Watch
1
Star
0
Fork
0
You've already forked Literature
Code
Issues
Pull Requests
Actions
Packages
Projects
Releases
Wiki
Activity
Files
b9b18061cf5f24feabdfba4367f97319c5bca9a8
Literature
/
Attension-Transformer-预训练
History
e2hang
7e829ac119
New File
2025-10-16 19:18:12 +08:00
..
BERT Pre-training of Deep Bidirectional Transformers forLanguage Understanding.pdf
New File
2025-10-16 19:18:12 +08:00
Dosovitskiy et al., 2020. An Image is Worth 16x16 Words Transformers for Image Recognition at Scale.pdf
New File
2025-10-16 19:18:12 +08:00
Neural Machine Translation by Jointly Learning to Align and Translate .pdf
New File
2025-10-16 19:18:12 +08:00
NeurIPS-2020-language-models-are-few-shot-learners-Paper.pdf
New File
2025-10-16 19:18:12 +08:00
Vaswani et al., 2017. Attention is All You Need.pdf
New File
2025-10-16 19:18:12 +08:00