
Title |
|---|
![]() Qwen Technical Report Jinze Bai Shuai Bai Yunfei Chu Zeyu Cui Kai Dang ...Zhenru Zhang Chang Zhou Jingren Zhou Xiaohuan Zhou Tianhang Zhu |
![]() Llama 2: Open Foundation and Fine-Tuned Chat Models Hugo Touvron Louis Martin Kevin R. Stone Peter Albert Amjad Almahairi ...Sharan Narang Aurelien Rodriguez Robert Stojnic Sergey Edunov Thomas Scialom |
![]() A Survey on Cross-Lingual SummarizationTransactions of the Association for Computational Linguistics (TACL), 2022 |
![]() Training language models to follow instructions with human feedbackNeural Information Processing Systems (NeurIPS), 2022 |
![]() Language Models are Few-Shot LearnersNeural Information Processing Systems (NeurIPS), 2020 |
![]() Multilingual Denoising Pre-training for Neural Machine TranslationTransactions of the Association for Computational Linguistics (TACL), 2020 |
![]() Attention Is All You NeedNeural Information Processing Systems (NeurIPS), 2017 |
![]() Outrageously Large Neural Networks: The Sparsely-Gated
Mixture-of-Experts LayerInternational Conference on Learning Representations (ICLR), 2017 |
![]() Sequence to Sequence Learning with Neural NetworksNeural Information Processing Systems (NeurIPS), 2014 |
![]() Neural Machine Translation by Jointly Learning to Align and TranslateInternational Conference on Learning Representations (ICLR), 2014 |