Papers

Learn more about AI2's Lasting Impact Award
Viewing 701-710 of 991 papers
  • Entity, Relation, and Event Extraction with Contextualized Span Representations

    David Wadden, Ulme Wennberg, Yi Luan, Hannaneh HajishirziEMNLP2019 We examine the capabilities of a unified, multi-task framework for three information extraction tasks: named entity recognition, relation extraction, and event extraction. Our framework (called DyGIE++) accomplishes all tasks by enumerating, refining, and…
  • Everything Happens for a Reason: Discovering the Purpose of Actions in Procedural Text

    Bhavana Dalvi Mishra, Niket Tandon, Antoine Bosselut, Wen-tau Yih, Peter ClarkEMNLP2019 Our goal is to better comprehend procedural text, e.g., a paragraph about photosynthesis, by not only predicting what happens, but why some actions need to happen before others. Our approach builds on a prior process comprehension framework for predicting…
  • Global Reasoning over Database Structures for Text-to-SQL Parsing

    Ben Bogin, Matt Gardner, Jonathan BerantEMNLP2019 State-of-the-art semantic parsers rely on auto-regressive decoding, emitting one symbol at a time. When tested against complex databases that are unobserved at training time (zero-shot), the parser often struggles to select the correct set of database…
  • “Going on a vacation” takes longer than “Going for a walk”: A Study of Temporal Commonsense Understanding

    Ben Zhou, Daniel Khashabi, Qiang Ning, Dan RothEMNLP2019 Understanding time is crucial for understanding events expressed in natural language. Because people rarely say the obvious, it is often necessary to have commonsense knowledge about various temporal aspects of events, such as duration, frequency, and…
  • Knowledge Enhanced Contextual Word Representations

    Matthew E. Peters, Mark Neumann, Robert L. Logan, Roy Schwartz, Vidur Joshi, Sameer Singh, and Noah A. SmithEMNLP2019 Contextual word representations, typically trained on unstructured, unlabeled text, do not contain any explicit grounding to real world entities and are often unable to remember facts about those entities. We propose a general method to embed multiple…
  • Language Modeling for Code-Switching: Evaluation, Integration of Monolingual Data, and Discriminative Training

    Hila Gonen, Yoav GoldbergEMNLP2019 We focus on the problem of language modeling for code-switched language, in the context of automatic speech recognition (ASR). Language modeling for code-switched language is challenging for (at least) three reasons: (1) lack of available large-scale code…
  • Low-Resource Parsing with Crosslingual Contextualized Representations

    Phoebe Mulcaire, Jungo Kasai, Noah A. SmithCoNLL2019 Despite advances in dependency parsing, languages with small treebanks still present challenges. We assess recent approaches to multilingual contextual word representations (CWRs), and compare them for crosslingual transfer from a language with a large…
  • Mixture Content Selection for Diverse Sequence Generation

    Jaemin Cho, Minjoon Seo, Hannaneh HajishirziEMNLP2019 Generating diverse sequences is important in many NLP applications such as question generation or summarization that exhibit semantically one-to-many relationships between source and the target sequences. We present a method to explicitly separate…
  • On the Limits of Learning to Actively Learn Semantic Representations

    Omri Koshorek, Gabriel Stanovsky, Yichu Zhou, Vivek Srikumar and Jonathan BerantCoNLL2019
    Best Paper Honorable Mention
    One of the goals of natural language understanding is to develop models that map sentences into meaning representations. However, training such models requires expensive annotation of complex structures, which hinders their adoption. Learning to actively…
  • PaLM: A Hybrid Parser and Language Model

    Hao Peng, Roy Schwartz, Noah A. SmithEMNLP2019 We present PaLM, a hybrid parser and neural language model. Building on an RNN language model, PaLM adds an attention layer over text spans in the left context. An unsupervised constituency parser can be derived from its attention weights, using a greedy…