Attention is all you need A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez, ... arXiv preprint arXiv:1706.03762, 2017 | 19516 | 2017 |
Advances in neural information processing systems A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez, ... Neural Information Processing Systems Foundation, 5998-6008, 2017 | 391 | 2017 |
Natural questions: a benchmark for question answering research T Kwiatkowski, J Palomaki, O Redfield, M Collins, A Parikh, C Alberti, ... Transactions of the Association for Computational Linguistics 7, 453-466, 2019 | 358 | 2019 |
Tensor2tensor for neural machine translation A Vaswani, S Bengio, E Brevdo, F Chollet, AN Gomez, S Gouws, L Jones, ... arXiv preprint arXiv:1803.07416, 2018 | 347 | 2018 |
The best of both worlds: Combining recent advances in neural machine translation MX Chen, O Firat, A Bapna, M Johnson, W Macherey, G Foster, L Jones, ... arXiv preprint arXiv:1804.09849, 2018 | 263 | 2018 |
One model to learn them all L Kaiser, AN Gomez, N Shazeer, A Vaswani, N Parmar, L Jones, ... arXiv preprint arXiv:1706.05137, 2017 | 238 | 2017 |
Character-level language modeling with deeper self-attention R Al-Rfou, D Choe, N Constant, M Guo, L Jones Proceedings of the AAAI Conference on Artificial Intelligence 33 (01), 3159-3166, 2019 | 159 | 2019 |
Wikireading: A novel large-scale language understanding task over wikipedia D Hewlett, A Lacoste, L Jones, I Polosukhin, A Fandrianto, J Han, ... arXiv preprint arXiv:1608.03542, 2016 | 118 | 2016 |
L. u. Kaiser and I. Polosukhin A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez Advances in Neural Information Processing Systems 30, 5998-6008, 2017 | 82* | 2017 |
Lingvo: a modular and scalable framework for sequence-to-sequence modeling J Shen, P Nguyen, Y Wu, Z Chen, MX Chen, Y Jia, A Kannan, T Sainath, ... arXiv preprint arXiv:1902.08295, 2019 | 79 | 2019 |
Attention is all you need. CoRR abs/1706.03762 (2017) A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez, ... arXiv preprint arXiv:1706.03762, 2017 | 64 | 2017 |
Attention is all you need. arXiv 2017 A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez, ... arXiv preprint arXiv:1706.03762, 2017 | 54 | 2017 |
ProtTrans: Towards Cracking the Language of Life's Code Through Self-Supervised Deep Learning and High Performance Computing A Elnaggar, M Heinzinger, C Dallago, G Rihawi, Y Wang, L Jones, ... arXiv preprint arXiv:2007.06225, 2020 | 31 | 2020 |
Accurate supervised and semi-supervised machine reading for long documents D Hewlett, L Jones, A Lacoste, I Gur Proceedings of the 2017 Conference on Empirical Methods in Natural Language …, 2017 | 19 | 2017 |
Byte-level machine reading across morphologically varied languages T Kenter, L Jones, D Hewlett Proceedings of the AAAI Conference on Artificial Intelligence 32 (1), 2018 | 12 | 2018 |
Kaiser, u. 2017. Attention is all you need A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez arXiv preprint arXiv:1601.03317, 0 | 4 | |
Machine translation using neural network models Z Chen, MR Hughes, Y Wu, M Schuster, X Chen, LO Jones, NJ Parmar, ... US Patent App. 16/521,780, 2020 | 3 | 2020 |
Multi-task multi-modal machine learning system NM Shazeer, AN Gomez, LM Kaiser, JD Uszkoreit, LO Jones, NJ Parmar, ... US Patent 10,789,427, 2020 | 1 | 2020 |
Attention-based sequence transduction neural networks NM Shazeer, AN Gomez, LM Kaiser, JD Uszkoreit, LO Jones, NJ Parmar, ... US Patent 10,452,978, 2019 | 1 | 2019 |
CodeTrans: Towards Cracking the Language of Silicone's Code Through Self-Supervised Deep Learning and High Performance Computing A Elnaggar, W Ding, L Jones, T Gibbs, T Feher, C Angerer, S Severini, ... arXiv preprint arXiv:2104.02443, 2021 | | 2021 |