
Title |
|---|
![]() Human vs. Muppet: A Conservative Estimate of Human Performance on the
GLUE BenchmarkAnnual Meeting of the Association for Computational Linguistics (ACL), 2019 |
![]() BoolQ: Exploring the Surprising Difficulty of Natural Yes/No QuestionsNorth American Chapter of the Association for Computational Linguistics (NAACL), 2019 |
![]() MCScript2.0: A Machine Comprehension Corpus Focused on Script Events and
ParticipantsInternational Workshop on Semantic Evaluation (SemEval), 2019 |
![]() Performance Analysis of Deep Learning Workloads on Leading-edge SystemsInternational Workshop on Performance Modeling, Benchmarking and Simulation of High Performance Computer Systems (PMBS), 2019 |
![]() HellaSwag: Can a Machine Really Finish Your Sentence?Annual Meeting of the Association for Computational Linguistics (ACL), 2019 |
![]() Story Ending Prediction by Transferable BERTInternational Joint Conference on Artificial Intelligence (IJCAI), 2019 |
![]() ERNIE: Enhanced Language Representation with Informative EntitiesAnnual Meeting of the Association for Computational Linguistics (ACL), 2019 |
![]() SuperGLUE: A Stickier Benchmark for General-Purpose Language
Understanding SystemsNeural Information Processing Systems (NeurIPS), 2019 |
![]() Enabling Robots to Understand Incomplete Natural Language Instructions
Using Commonsense ReasoningIEEE International Conference on Robotics and Automation (ICRA), 2019 |
![]() DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning
Over ParagraphsNorth American Chapter of the Association for Computational Linguistics (NAACL), 2019 |
![]() Still a Pain in the Neck: Evaluating Text Representations on Lexical
CompositionTransactions of the Association for Computational Linguistics (TACL), 2019 |