Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

Teaching Broad Reasoning Skills via Decomposition-Guided Contexts

Harsh TrivediNiranjan BalasubramanianTushar KhotAshish Sabharwal
2022
EMNLP

Question-answering datasets require a broad set of reasoning skills. We show how to use question decompositions to teach language models these broad reasoning skills in a robust fashion.… 

Towards Teachable Reasoning Systems: Using a Dynamic Memory of User Feedback for Continual System Improvement

Bhavana Dalvi MishraOyvind TafjordPeter Clark
2022
EMNLP

Our goal is a teachable reasoning system for question-answering (QA), where a user can interact with faithful answer explanations, and correct its errors so that the system improves over time. Our… 

Twist Decoding: Diverse Generators Guide Each Other

Jungo KasaiKeisuke SakaguchiRonan Le BrasNoah A. Smith
2022
EMNLP

Natural language generation technology has recently seen remarkable progress with large-scale training, and many natural language applications are now built upon a wide range of generation models.… 

UnifiedSKG: Unifying and Multi-Tasking Structured Knowledge Grounding with Text-to-Text Language Models

Tianbao XieChen Henry WuPeng ShiTao Yu
2022
EMNLP

Structured knowledge grounding (SKG) leverages structured knowledge to complete user requests, such as semantic parsing over databases and question answering over knowledge bases. Since the inputs… 

Unsupervised Learning of Hierarchical Conversation Structure

Bo-Ru LuYushi HuHao ChengMari Ostendorf
2022
EMNLP Findings

Human conversations can evolve in many different ways, creating challenges for automatic understanding and summarization. Goal-oriented conversations often have meaningful sub-dialogue structure,… 

WANLI: Worker and AI Collaboration for Natural Language Inference Dataset Creation

Alisa LiuSwabha SwayamdiptaNoah A. SmithYejin Choi
2022
Findings of EMNLP

A recurring challenge of crowdsourcing NLP datasets at scale is that human writers often rely on repetitive patterns when crafting examples, leading to a lack of linguistic diversity. We introduce a… 

What Makes Instruction Learning Hard? An Investigation and a New Challenge in a Synthetic Environment

Matthew FinlaysonKyle RichardsonAshish SabharwalPeter Clark
2022
EMNLP

The instruction learning paradigm—where a model learns to perform new tasks from task descriptions alone—has become popular in general-purpose model research. The capabilities of large transformer… 

Whose Language Counts as High Quality? Measuring Language Ideologies in Text Data Selection

Suchin GururanganDallas CardSarah K. DrierNoah A. Smith
2022
EMNLP

Language models increasingly rely on massive web dumps for diverse text data. However, these sources are rife with undesirable content. As such, resources like Wikipedia, books, and news often… 

Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering

Pan LuSwaroop MishraTony XiaA. Kalyan
2022
NeurIPS 2022

When answering a question, humans utilize the information available across different modalities to synthesize a consistent and complete chain of thought (CoT). This process is normally a black box… 

Ask4Help: Learning to Leverage an Expert for Embodied Tasks

Kunal Pratap SinghLuca WeihsAlvaro HerrastiRoozbeh Mottaghi
2022
arXiv

Embodied AI agents continue to become more capable every year with the advent of new models, environments, and benchmarks, but are still far away from being performant and reliable enough to be…