See AI2's full collection of videos on our YouTube channel.
All Years
All Videos
Viewing 11-20 of 158 videos
  • Extremely Large Neural Memory of Unstructured Knowledge Thumbnail

    Extremely Large Neural Memory of Unstructured Knowledge

    November 20, 2019  |  Minjoon Seo
    The web is a collection of massive and mostly unstructured knowledge data. My recent research has focused on organizing every piece of information on a web-scale corpus and making it readily accessible so that it can be efficiently utilized for creating a language understanding system that requires world knowledge (e.g. question answering). I approach the problem by creating an extremely large neural memory where the entire text corpus is discretized into billions of atomic information and hashed with key vectors. This allows us to random-access specific word-level information in the corpus very fast and accurately. In the first part of my talk, I will highlight the direct usage of the neural memory in factual question answering (e.g. SQuAD, Natural Questions), where I show that the “memorification” of the corpus leads to at least 100x faster inference with better accuracy. In the second part, I will discuss a more advanced usage as a future research direction that utilizes an interactive memory controller, which could hint how we can approach language understanding tasks that need to jointly consider several different pieces of information spread over the web.
  • Learning Dynamics of LSTM Language Models Thumbnail

    Learning Dynamics of LSTM Language Models

    November 20, 2019  |  Naomi Saphra
    Research has shown that neural models implicitly encode linguistic features, but there has been little work exploring how these encodings arise as the models are trained. I will be presenting work on the learning dynamics of neural language models from a variety of angles. Using Singular Vector Canonical Correlation Analysis to probe the evolution of syntactic, semantic, and topic representations, we find that part-of-speech is learned earlier than topic; that recurrent layers become more similar to those of a tagger during training; and embedding layers less similar. I will also discuss how these results connect to the compositionality of LSTM learning dynamics, with synthetic experimental evidence that language models rely on short memorized segments in learning general long-range relations between words. I will explore links between syntactic relations and mathematical properties of the interactions between words.
  • Medical Question Answering: Dealing with the complexity and specificity of consumer health questions and visual questions Thumbnail

    Medical Question Answering: Dealing with the complexity and specificity of consumer health questions and visual questions

    November 12, 2019  |  Dr. Asma Ben Abacha
    Consumer health questions pose specific challenges to automated answering. Two of the salient aspects are the higher linguistic and semantic complexity when compared to open domain questions, and the more pronounced need for reliable information. In this talk I will present two main approaches to deal with the increased complexity by recognizing question entailment and by question summarization, recently published respectively in BMC Bioinformatics and ACL 2019. In particular, our question entailment approach to question answering (QA) showed that restricting the answer sources to only reliable resources led to an improvement of the QA performance and our summarization experiments showed the relevance of data augmentation methods for abstractive question summarization. I’ll also talk about the MEDIQA shared task on question entailment, textual inference and medical question answering that we recently organized at ACL-BioNLP. In the second part of the talk, I will address more specifically questions about medications and present our last study and dataset on medication QA. Finally, I’ll describe our recent endeavors in visual question answering (VQA) from radiology images and the medical VQA challenge (VQA-Med) editions for 2019 and 2020 that we organize in the scope of ImageCLEF.
  • Boosting innovation and discovery of ideas Thumbnail

    Boosting innovation and discovery of ideas

    August 7, 2019  |  Tom Hope
    The explosion of available idea repositories -- scientific papers, patents, product descriptions -- represents an unprecedented opportunity to accelerate innovation and lead to a wealth of discoveries. Given the scale of the problem and its ever-expanding nature, there is a need for intelligent automation to assist in the process of discovery. In this talk, I will present our work toward addressing this challenging problem. We developed an approach for boosting people’s creativity by helping them discover analogies -- abstract structural connections between ideas. We learn to decompose innovation texts into functional models that describe the components and goals of inventions, and use them to build a search engine supporting expressive inspiration queries. In ideation studies, our inspirations helped people generate better ideas with significant improvement over standard search. We also construct a commonsense ontology of purposes and mechanisms of products, mapping the landscape of ideas. I will also describe a novel machine learning framework we developed in order to identify innovation in patents, where labels are extremely hard to obtain. In our setting, called Ballpark Learning, we are only given groups of instances with coarse constraints over label averages. We demonstrate encouraging results in classification and regression tasks across several domains.
  • Extracting T cell function and differentiation characteristics Thumbnail

    Extracting T cell function and differentiation characteristics

    July 23, 2019  |  Jeff Hammerbacher
    Many promising cancer immunotherapy treatment protocols rely on efficient and increasingly extensive methods for manipulating human immune cells. T cells are a frequent target of the laboratory and clinical research driving the development of such protocols as they are most often the effector of the cytotoxic activity that makes these treatments so potent. However, the cytokine signaling network that drives the differentiation and function of such cells is complex and difficult to replicate on a large scale in model biological systems. Abridged versions of these networks have been established over decades of research but it remains challenging to define their global structure as the classification of T cell subtypes operating in these networks, the mechanics of their formation, and the purpose of the signaling molecules they excrete are all controversial, with a slowly expanding understanding emerging in literature over time. To aid in the quantification of this understanding, we are developing a methodology for identifying references to well known cytokines, transcription factors, and T cell types in literature as well as classifying the relationships between the three in an attempt to determine what cytokines initiate the transcription programs that lead to various cell states in addition to the secretion profiles associated with those states. Entity recognition for this task is performed using SciSpacy and classification of the relations between these entities is based on an LSTM trained using Snorkel, where weak supervision is established through a variety of classification heuristics and distant supervision is provided via previously published immunology databases.
  • Natural Language Understanding for Events and Participants in Text Thumbnail

    Natural Language Understanding for Events and Participants in Text

    May 6, 2019  |  Rachel Rudinger
    Consider the difference between the two sentences “Pat didn’t remember to water the plants” and “Pat didn’t remember that she had watered the plants.” Fluent English speakers recognize that the former sentence implies that Pat did not water the plants, while the latter sentence implies she did. This distinction is crucial to understanding the meaning of these sentences, yet it is one that automated natural language processing (NLP) systems struggle to make. In this talk, I will discuss my work on developing state-of-the-art NLP models that make essential inferences about events (e.g., a “watering” event) and participants (e.g., “Pat” and “the plants”) in natural language sentences. In particular, I will focus on two supervised NLP tasks that serve as core tests of language understanding: Event Factuality Prediction and Semantic Proto-Role Labeling. I will also discuss my work on unsupervised acquisition of common-sense knowledge from large natural language text corpora, and the concomitant challenge of detecting problematic social biases in NLP models trained on such data.
  • Understanding question comprehension, and generalizability Thumbnail

    Understanding question comprehension, and generalizability

    May 2, 2019  |  Pramod Kaushik Mudrakarta
    We present two results: 1) Analysis techniques for state-of-the-art question-answering models on images, tables and passages of text. We show how these networks often ignore important question terms. Leveraging such non-robust behavior, we present a variety of adversarial examples derived by perturbing the questions. Our strongest attacks drop the accuracy of a visual question answering model from 61.1% to 19%, and that of a tabular question answering model from 33.5% to 3.3%. We demonstrate that attributions can augment standard measures of accuracy and empower investigation of model performance. When a model is accurate but for the wrong reasons, attributions can surface erroneous logic in the model that indicates inadequacies in the data. 2) Parameter-efficient transfer learning: We present a novel method for re-purposing pretrained neural networks to new tasks while maintaining most of the weights intact. The basic approach is to learn a model patch - a small set of parameters - that will specialize to each task, instead of fine-tuning the last layer or the entire network. For instance, we show that learning a set of scales and biases is sufficient to convert a pretrained network to perform well on qualitatively different problems (e.g. converting a Single Shot MultiBox Detection (SSD) model into a 1000-class image classification model while reusing 98% of parameters of the SSD feature extractor). Our approach allows both simultaneous (multi-task) as well as sequential transfer learning. In several multi-task learning problems, despite using much fewer parameters than traditional logits-only fine-tuning, we match single-task performance.
  • Augmenting Collective Human Work Thumbnail

    Augmenting Collective Human Work

    May 1, 2019  |  Jonathan Bragg
    A longstanding goal of artificial intelligence (AI) is to develop agents that can assist or augment humans. Such agents have the potential to transform society. While AI agents can excel at well-defined tasks like games, much more limited progress has been made solving real-world problems like interacting with humans, where data collection is costly, objectives are ill-defined, and safety is critical. In this talk, I will discuss how we can design agents to improve the efficiency and success of collective human work ("crowdsourcing"), by leveraging techniques from AI, reinforcement learning, and optimization, together with structured contributions from human workers and task designers. This approach improves on current methods for designing such agents, which typically require large amounts of manual experimentation and costly data collection to get right. I will demonstrate the effectiveness of this approach on several crowdsourcing management problems, and also share recent work on how agents can make shared decisions with humans to achieve better outcomes.
  • Both Sides Now: Generating and Understanding Visually-Grounded Language Thumbnail

    Both Sides Now: Generating and Understanding Visually-Grounded Language

    April 24, 2019  |  Peter Anderson
    From robots to cars, virtual assistants and voice-controlled drones, computing devices are increasingly expected to communicate naturally with people and to understand the visual context in which they operate. In this talk, I will present our latest work on generating and comprehending visually-grounded language. First, we will discuss the challenging task of describing an image (image captioning). I will introduce captioning models that leverage multiple data sources, including object detection datasets and unaligned text corpora, in order to learn about the long-tail of visual concepts found in the real world. To support and encourage further efforts in this area, I will present the 'nocaps' benchmark for novel object captioning. In the second part of the talk, I will describe our recent work on developing agents that follow natural language instructions in reconstructed 3D environments using the R2R dataset for vision-and-language navigation.
  • User-centric Recommendation Models and Systems Thumbnail

    User-centric Recommendation Models and Systems

    April 12, 2019  |  Longqi Yang
    The daily actions and decisions of people are increasingly shaped by recommendation systems, from e-commerce and content platforms to education and wellness applications. These systems selectively suggest and present information items based on their characterization of user preferences. However, existing preference modeling methods are limited due to the incomplete and biased nature of the behavioral data that inform the models. As a result, recommendations can be narrow, skewed, homogeneous, and divergent from users’ aspirations. In this talk, I will introduce user-centric recommendation models and systems that address the incompleteness and bias of existing methods and increase systems’ utility for individuals. Specifically, I will present my work addressing two key research challenges: (1) inferring debiased preferences from biased behavioral data using counterfactual reasoning, and (2) eliciting unobservable current and aspirational preferences from users through interactive machine learning. I will conclude with discussion of field experiments that demonstrate how user-centric systems can promote healthier diets and better content choices.