PaLM: A Hybrid Parser and Language ModelConference on Empirical Methods in Natural Language Processing (EMNLP), 2019 |
Deep Equilibrium ModelsNeural Information Processing Systems (NeurIPS), 2019 |
Language Models as Knowledge Bases?Conference on Empirical Methods in Natural Language Processing (EMNLP), 2019 |
Subword Language Model for Query Auto-CompletionConference on Empirical Methods in Natural Language Processing (EMNLP), 2019 |
Quantity doesn't buy quality syntax with neural language modelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2019 |
Keep Calm and Switch On! Preserving Sentiment and Fluency in Semantic
Text ExchangeConference on Empirical Methods in Natural Language Processing (EMNLP), 2019 |
Transformer Dissection: A Unified Understanding of Transformer's
Attention via the Lens of KernelConference on Empirical Methods in Natural Language Processing (EMNLP), 2019 |
Discourse-Aware Semantic Self-Attention for Narrative Reading
ComprehensionConference on Empirical Methods in Natural Language Processing (EMNLP), 2019 |
Latent Relation Language ModelsAAAI Conference on Artificial Intelligence (AAAI), 2019 |
Beyond English-Only Reading Comprehension: Experiments in Zero-Shot
Multilingual Transfer for BulgarianRecent Advances in Natural Language Processing (RANLP), 2019 |
AutoML: A Survey of the State-of-the-ArtKnowledge-Based Systems (KBS), 2019 |
Leveraging Pre-trained Checkpoints for Sequence Generation TasksTransactions of the Association for Computational Linguistics (TACL), 2019 |
ERNIE 2.0: A Continual Pre-training Framework for Language UnderstandingAAAI Conference on Artificial Intelligence (AAAI), 2019 |
SpanBERT: Improving Pre-training by Representing and Predicting SpansTransactions of the Association for Computational Linguistics (TACL), 2019 |
LakhNES: Improving multi-instrumental music generation with cross-domain
pre-trainingInternational Society for Music Information Retrieval Conference (ISMIR), 2019 |
Large Memory Layers with Product KeysNeural Information Processing Systems (NeurIPS), 2019 |
A Tensorized Transformer for Language ModelingNeural Information Processing Systems (NeurIPS), 2019 |
XLNet: Generalized Autoregressive Pretraining for Language UnderstandingNeural Information Processing Systems (NeurIPS), 2019 |
Pre-Training with Whole Word Masking for Chinese BERTIEEE/ACM Transactions on Audio Speech and Language Processing (TASLP), 2019 |
Theoretical Limitations of Self-Attention in Neural Sequence ModelsTransactions of the Association for Computational Linguistics (TACL), 2019 |
Large-Scale Multi-Label Text Classification on EU LegislationAnnual Meeting of the Association for Computational Linguistics (ACL), 2019 |
Towards Lossless Encoding of SentencesAnnual Meeting of the Association for Computational Linguistics (ACL), 2019 |
Why gradient clipping accelerates training: A theoretical justification
for adaptivityInternational Conference on Learning Representations (ICLR), 2019 |
Interpreting and improving natural-language processing (in machines)
with natural language-processing (in the brain)Neural Information Processing Systems (NeurIPS), 2019 |
Are Sixteen Heads Really Better than One?Neural Information Processing Systems (NeurIPS), 2019 |
Exposure Bias versus Self-Recovery: Are Distortions Really Incremental
for Autoregressive Text Generation?Conference on Empirical Methods in Natural Language Processing (EMNLP), 2019 |
Adaptive Attention Span in TransformersAnnual Meeting of the Association for Computational Linguistics (ACL), 2019 |