Viewing 1-20 of 31 data
Clear all
    • Break NEW
      83,978 examples sampled from 10 question answering datasets over text, images and databases.

      Break is a human annotated dataset of natural language questions and their Question Decomposition Meaning Representations (QDMRs). Break consists of 83,978 examples sampled from 10 question answering datasets over text, images and databases.

    • 3864 questions about open domain qualitative relationships

      QuaRTz is a crowdsourced dataset of 3864 multiple-choice questions about open domain qualitative relationships, as described in QuaRTz: An Open-Domain Dataset of Qualitative Relationship Questions (EMNLP 2019). Each question is paired with one of 405 different background sentences (sometimes short paragraphs).

    • 7,787 multiple choice questions annotated with question classification labels

      A dataset of detailed problem domain classification labels for each of the 7,787 multiple-choice science questions found in the AI2 Reasoning Challenge (ARC) dataset, to enable targeted pairing of questions with problem-specific solvers. Also included is a taxonomy of 462 detailed problem domains for grade-school science, organized into 6 levels of specificity.

    • 2771 story questions about qualitative relationships

      QuaRel is a crowdsourced dataset of 2771 multiple-choice story questions, including their logical forms, as described in QuaRel: A Dataset and Models for Answering Questions about Qualitative Relationships (AAAI 2019).

    • 5,957 multiple-choice questions probing a book of 1,326 science facts

      OpenBookQA aims to promote research in advanced question-answering, probing a deeper understanding of both the topic (with salient facts summarized as an open book, also provided with the dataset) and the language it is expressed in. In particular, it contains questions that require multi-step reasoning, use of additional common and commonsense knowledge, and rich text comprehension.

    • Over 39 million published research papers in Computer Science, Neuroscience, and Biomedical

      This is a subset of the full Semantic Scholar corpus which represents papers crawled from the Web and subjected to a number of filters.

    • 488 richly annotated paragraphs about processes (containing 3,300 sentences)

      The ProPara dataset is designed to train and test comprehension of simple paragraphs describing processes (e.g., photosynthesis), designed for the task of predicting, tracking, and answering questions about how entities change during the process.

    • Over 14K paper drafts and over 10K textual peer reviews

      PeerRead is a dataset of scientific peer reviews available to help researchers study this important artifact.

    • 7,787 multiple choice science questions and associated corpora

      A new dataset of 7,787 genuine grade-school level, multiple-choice science questions, assembled to encourage research in advanced question-answering. The dataset is partitioned into a Challenge Set and an Easy Set, where the former contains only questions answered incorrectly by both a retrieval-based algorithm and a word co-occurrence algorithm. We are also including a corpus of over 14 million science sentences relevant to the task, and an implementation of three neural baseline models for this dataset. We pose ARC as a challenge to the community.

    • Explanation graphs for 1,680 questions

      A collection of resources for studying explanation-centered inference, including explanation graphs for 1,680 questions, with 4,950 tablestore rows, and other analyses of the knowledge required to answer elementary and middle-school science questions. ExplanationBank was constructed by Peter Jansen (University of Arizona), in collaboration with AI2.

    • 27,026 statements

      The SciTail dataset is an entailment dataset created from multiple-choice science exams and web sentences. Each question and the correct answer choice are converted into an assertive statement to form the hypothesis.

    • 13,679 science questions with supporting sentences

      The SciQ dataset contains 13,679 crowdsourced science exam questions about Physics, Chemistry and Biology, among others. The questions are in multiple-choice format with 4 answer options each. For the majority of the questions, an additional paragraph with supporting evidence for the correct answer is provided.

    • 156K sentences for 4th grade questions, 107K sentences for 8th grade questions, and derived tuples

      The TupleInf Open IE dataset contains Open IE tuples extracted from 263K sentences that were used by the solver in "Answering Complex Questions Using Open Information Extraction” (referred as Tuple KB, T). These sentences were collected from a large Web corpus using training questions from 4th and 8th grade as queries. This dataset contains 156K sentences collected for 4th grade questions and 107K sentences for 8th grade questions. Each sentence is followed by the Open IE v4 tuples using their simple format.

    • 9,356 science terms and sentences

      This is the dataset for the paper Leveraging Term Banks for Answering Complex Questions: A Case for Sparse Vectors. The dataset contains 9,356 science terms and, for each term, an average of 16,000 sentences that contain the term.

    • 1,076 textbook lessons, 26,260 questions, 6229 images

      The TextbookQuestionAnswering (TQA) dataset is drawn from middle school science curricula as described in Are You Smarter Than A Sixth Grader? Textbook Question Answering for Multimodal Machine Comprehension. It consists of 1,076 lessons from Life Science, Earth Science and Physical Science textbooks downloaded from Each lesson has a set of multiple choice questions that address concepts taught in that lesson. TQA has a total of 26,260 questions including 12,567 that have an accompanying diagram.

    • 294,000 science-relevant tuples

      The Aristo Tuple KB contains 294,000 high-precision, domain-targeted (subject,relation,object) tuples extracted from text using a high-precision extraction pipeline, and guided by domain vocabulary constraints.

    • 1,197,377 science-relevant sentences

      The Aristo Mini corpus contains 1,197,377 (very loosely) science-relevant sentences drawn from public data. It provides simple science-relevant text that may be useful to help answer elementary science questions. It is used in the Aristo Mini system and is also available here as a resource in its own right.

    • 1,363 gold explanation sentences

      This is the dataset for the paper What's in an Explanation? Characterizing Knowledge and Inference Requirements for Elementary Science Exams (COLING'16). The data contains: gold explanation sentences supporting 363 science questions, relation annotation for a subset of those explanations, and a graphical annotation tool with annotation guidelines. This dataset was produced by AI2, the University of Arizona, and Stony Brook University.

    • 4,817 images

      AI2D is a dataset of illustrative diagrams for research on diagram understanding and associated question answering.

    • 1,080 questions

      These questions were created using the "AI2 Elementary School Science Questions (No Diagrams)" data set by changing all of the incorrect answer options of each question with some other related word. This dataset can be a good measure of robustness for QA systems when being testing on modified questions. More details can be found in the paper Question Answering via Integer Programming over Semi-Structured Knowledge.