Papers

Learn more about AI2's Lasting Impact Award
Viewing 51-60 of 216 papers
  • What Makes Instruction Learning Hard? An Investigation and a New Challenge in a Synthetic Environment

    Matthew Finlayson, Kyle Richardson, Ashish Sabharwal, Peter ClarkEMNLP2022 The instruction learning paradigm—where a model learns to perform new tasks from task descriptions alone—has become popular in general-purpose model research. The capabilities of large transformer models as instruction learners, however, remain poorly under…
  • Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering

    Pan Lu, Swaroop Mishra, Tony Xia, Liang Qiu, Kai-Wei Chang, Song-Chun Zhu, Oyvind Tafjord, Peter Clark, A. KalyanNeurIPS 20222022 When answering a question, humans utilize the information available across different modalities to synthesize a consistent and complete chain of thought (CoT). This process is normally a black box in the case of deep learning models like large-scale language…
  • One Venue, Two Conferences: The Separation of Chinese and American Citation Networks

    Bingchen Zhao, Yuling Gu, Jessica Zosa Forde, Naomi SaphraNeurIPS • AI Cultures Workshop2022 At NeurIPS, American and Chinese institutions cite papers from each other’s regions substantially less than they cite endogamously. We build a citation graph to quantify this divide, compare it to European connectivity, and discuss the causes and consequences…
  • Breakpoint Transformers for Modeling and Tracking Intermediate Beliefs

    Kyle Richardson, Ronen Tamari, Oren Sultan, Reut Tsarfaty, Dafna Shahaf, Ashish SabharwalEMNLP2022 Can we teach natural language understanding models to track their beliefs through intermediate points in text? We propose a representation learning framework called breakpoint modeling that allows for learning of this type. Given any text encoder and data…
  • Learning to Decompose: Hypothetical Question Decomposition Based on Comparable Texts

    Ben Zhou, Kyle Richardson, Xiaodong Yu, Dan RothEMNLP2022 Explicit decomposition modeling, which involves breaking down complex tasks into more straightforward and often more interpretable sub-tasks, has long been a central theme in developing robust and interpretable NLU systems. However, despite the many datasets…
  • Just-DREAM-about-it: Figurative Language Understanding with DREAM-FLUTE

    Yuling Gu, Yao Fu, Valentina Pyatkin, Ian Magnusson, Bhavana Dalvi Mishra, Peter ClarkEMNLP • The Third Workshop on Figurative Language Processing 2022 Figurative language (e.g., “he flew like the wind”) is challenging to understand, as it is hard to tell what implicit information is being conveyed from the surface form alone. We hypothesize that to perform this task well, the reader needs to mentally…
  • Dyna-bAbI: unlocking bAbI’s potential with dynamic synthetic benchmarking

    Ronen Tamari, Kyle Richardson, Aviad Sar-Shalom, Noam Kahlon, Nelson H S Liu, Reut Tsarfaty, Dafna Shahaf SEM2022 While neural language models often perform surprisingly well on natural language understanding (NLU) tasks, their strengths and limitations remain poorly understood. Controlled synthetic tasks are thus an increasingly important resource for diagnosing model…
  • Learning to Repair: Repairing model output errors after deployment using a dynamic memory of feedback

    Niket Tandon, Aman Madaan, Peter Clark, Yiming YangFindings of NAACL 2022 Large language models (LMs), while power-ful, are not immune to mistakes, but can be difficult to retrain. Our goal is for an LM to continue to improve after deployment, without retraining, using feedback from the user. Our approach pairs an LM with (i) a…
  • DeepA2: A Modular Framework for Deep Argument Analysis with Pretrained Neural Text2Text Language Models

    Gregor Betz, Kyle RichardsonSEM2022 In this paper, we present and implement a multi-dimensional, modular framework for performing deep argument analysis (DeepA2) using current pre-trained language models (PTLMs). ArgumentAnalyst – a T5 model (Raffel et al. 2020) set up and trained within DeepA2…
  • Retrieval Data Augmentation Informed by Downstream Question Answering Performance

    James Ferguson, Pradeep Dasigi, Tushar Khot, Hannaneh HajishirziACL • FEVER2022 Training retrieval models to fetch contexts for Question Answering (QA) over large corpora requires labeling relevant passages in those corpora. Since obtaining exhaustive manual annotations of all relevant passages is not feasible, prior work uses text…