Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

Commonsense Knowledge Transfer for Pre-trained Language Models

Wangchunshu ZhouRonan Le BrasYejin Choi
2023
ACL • Findings

Despite serving as the foundation models for a wide range of NLP benchmarks, pre-trained language models have shown limited capabilities of acquiring implicit commonsense knowledge from… 

Modular Transformers: Compressing Transformers into Modularized Layers for Flexible Efficient Inference

Wangchunshu ZhouRonan Le BrasYejin Choi
2023
ACL • Findings

Pre-trained Transformer models like T5 and BART have advanced the state of the art on a wide range of text generation tasks. Compressing these models into smaller ones has become critically… 

Minding Language Models' (Lack of) Theory of Mind: A Plug-and-Play Multi-Character Belief Tracker

Melanie SclarSachin KumarPeter WestYulia Tsvetkov
2023
ACL

Theory of Mind (ToM)$\unicode{x2014}$the ability to reason about the mental states of other people$\unicode{x2014}$is a key element of our social intelligence. Yet, despite their ever more… 

SQuARe: A Large-Scale Dataset of Sensitive Questions and Acceptable Responses Created Through Human-Machine Collaboration

Hwaran LeeSeokhee HongJoonsuk ParkJung-Woo Ha
2023
arXiv.org

The potential social harms that large language models pose, such as generating offensive content and reinforcing biases, are steeply rising. Existing works focus on coping with this concern while… 

Are Machine Rationales (Not) Useful to Humans? Measuring and Improving Human Utility of Free-Text Rationales

Brihi JoshiZiyi LiuSahana RamnathXiang Ren
2023
arXiv.org

Among the remarkable emergent capabilities of large language models (LMs) is free-text rationalization; beyond a certain scale, large LMs are capable of generating seemingly useful rationalizations,… 

ArK: Augmented Reality with Knowledge Interactive Emergent Ability

Qiuyuan HuangJ. ParkAbhinav GuptaJianfeng Gao
2023
arXiv.org

Despite the growing adoption of mixed reality and interactive AI agents, it remains challenging for these systems to generate high quality 2D/3D scenes in unseen environments. The common practice… 

Is Reinforcement Learning (Not) for Natural Language Processing?: Benchmarks, Baselines, and Building Blocks for Natural Language Policy Optimization

Rajkumar RamamurthyPrithviraj AmmanabroluKianté BrantleyYejin Choi
2023
ICLR

We tackle the problem of aligning pre-trained large language models (LMs) with human preferences. If we view text generation as a sequential decision-making problem, reinforcement learning (RL)… 

Queer In AI: A Case Study in Community-Led Participatory AI

Organizers Of Queer in AIAnaelia OvalleArjun SubramonianLuke Stark
2023
FAccT

We present Queer in AI as a case study for community-led participatory design in AI. We examine how participatory design and intersectional tenets started and shaped this community's programs over… 

CHAMPAGNE: Learning Real-world Conversation from Large-Scale Web Videos

Seungju HanJack HesselNouha DziriYoungjae Yu
2023
arXiv.org

Visual information is central to conversation: body gestures and facial expressions, for example, contribute to meaning that transcends words alone. To date, however, most neural conversational… 

Clever Hans or Neural Theory of Mind? Stress Testing Social Reasoning in Large Language Models

Natalie ShapiraMosh LevyS. AlaviVered Shwartz
2023
EACL

The escalating debate on AI’s capabilities warrants developing reliable metrics to assess machine “intelligence.” Recently, many anecdotal examples were used to suggest that newer Large Language…