Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

AppWorld: A Controllable World of Apps and People for Benchmarking Interactive Coding Agents

Harsh TrivediTushar KhotMareike HartmannNiranjan Balasubramanian
2024
ACL

Autonomous agents that address day-to-day digital tasks (e.g., ordering groceries for a household), must not only operate multiple apps (e.g., notes, messaging, shopping app) via APIs, but also… 

Can Language Models Serve as Text-Based World Simulators?

Ruoyao WangGraham ToddZiang XiaoP. Jansen
2024
ACL

Virtual environments play a key role in benchmarking advances in complex planning and decision-making tasks but are expensive and complicated to build by hand. Can current language models themselves… 

CHIME: LLM-Assisted Hierarchical Organization of Scientific Studies for Literature Review Support

Chao-Chun HsuErin BransomJenna SparksAakanksha Naik
2024
ACL

Literature review requires researchers to synthesize a large amount of information and is increasingly challenging as the scientific literature expands. In this work, we investigate the potential of… 

Few-shot Dialogue Strategy Learning for Motivational Interviewing via Inductive Reasoning

Zhouhang XieBodhisattwa Prasad MajumderMengjie ZhaoJulian McAuley
2024
ACL Findings

We consider the task of building a dialogue system that can motivate users to adopt positive lifestyle changes: Motivational Interviewing. Addressing such a task requires a system that can infer… 

Overview of the Context24 Shared Task on Contextualizing Scientific Claims

Joel ChanAakanksha NaikMatthew AkamatsuJenna Sparks
2024
ACL • SDP

To appropriately interpret and use scientific claims for sensemaking and decision-making, it is critical to contextualize them, not just with textual evidence that the claim was in fact asserted,… 

The Unreasonable Effectiveness of Easy Training Data for Hard Tasks

Peter HaseMohit BansalPeter ClarkSarah Wiegreffe
2024
ACL

How can we train models to perform well on hard test data when hard training data is by definition difficult to label correctly? This question has been termed the scalable oversight problem and has… 

Data Contamination Report from the 2024 CONDA Shared Task

Oscar SainzIker Garc'ia-FerreroAlon JacoviJinglin Yang
2024
arXiv

The 1st Workshop on Data Contamination (CONDA 2024) focuses on all relevant aspects of data contamination in natural language processing, where data contamination is understood as situations where… 

The Illusion of State in State-Space Models

William MerrillJackson PettyAshish Sabharwal
2024
ICML

State-space models (SSMs) have emerged as a potential alternative architecture for building large language models (LLMs) compared to the previously ubiquitous transformer architecture. One… 

Data-driven Discovery with Large Generative Models

Bodhisattwa Prasad MajumderHarshit SuranaDhruv AgarwalPeter Clark
2024
ICML

With the accumulation of data at an unprecedented rate, its potential to fuel scientific discovery is growing exponentially. This position paper urges the Machine Learning (ML) community to exploit… 

Skill Set Optimization: Reinforcing Language Model Behavior via Transferable Skills

Kolby NottinghamBodhisattwa Prasad MajumderBhavana DalviRoy Fox
2024
ICML

Large language models (LLMs) have recently been used for sequential decision making in interactive environments. However, leveraging environment reward signals for continual LLM actor improvement is…