Allen AI Predoctoral Young Investigators
About the Program
Allen AI Predoctoral Young Investigators is a program that offers predoctoral candidates the opportunity to prepare for graduate-level research through partnership with strong mentors, participation in cutting edge research, and co-authorship of papers at top venues. This program aims to position candidates for successful entry into a top graduate program. For potential candidates who are already enrolled in graduate studies, this program offers the opportunity to work with researchers at AI2 on collaborative projects for the benefit of both AI2’s initiatives and your thesis work.
- Duration: 1-3 years
- Start Date: Flexible (rolling application with no deadline)
- Candidates: PYI Program candidates have graduated with a bachelor’s or master’s degree in a relevant field and are preparing to enter a PhD program in the next 1-2 years. We also welcome applications from candidates already enrolled in a PhD program who wish to align their work with AI2’s endeavors, and we are happy to work on a case-by-case basis with regards to work arrangements if there is a mutual fit.
- Gain strong research experience through direct collaboration with AI2 researchers
- Co-author papers with AI2 researchers and get published in top venues
- Learn about and leverage AI2’s data, engineering infrastructure, and research tools
- Gain industry experience in a supportive, agile, mission-driven environment
- No grant writing, teaching, or administrative responsibilities required
- Assistance with graduate school vetting and application
- Generous travel budget
- Financial and legal support for visa acquisition
PYI Program Alumni
Will graduated from Yale in 2019 with a B.S. in computer science and linguistics. He is broadly interested in the linguistic abilities, robustness, and interpretability of neural networks for NLP, often through the lens of theoretical tools like formal languages. At AI2, Will’s work contributed to formal papers on A Formal Hierarchy of RNN Architectures, On the Linguistic Capacity of Real-Time Counter Automata, Effects of Parameter Norm Growth During Transformer Training: Inductive Bias from Gradient Descent, Provable Limitations of Acquiring Meaning from Ungrounded Form: What Will Future Language Models Understand?, Competency Problems: On Finding and Removing Artifacts in Language Data, and CORD-19: The COVID-19 Open Research Dataset. After leaving AI2, Will will join New York University as a Ph.D. student advised by Tal Linzen.
Sanjay earned a B.S.E. in computer and information science and a B.S. in economics from the University of Pennsylvania in 2019 and has been a predoctoral young investigator in AI2’s AllenNLP team from June 2019 to July 2021. His research interests are in NLP and its intersection with computer vision. Through his work at AI2, he was a co-author of papers about linguistic analysis of pre-trained vision and language models, the faithfulness of neural module networks, a dataset of images and associated text from medical research papers, contrast sets for NLP evaluation, and a latent tree model for visual question answering. Beginning in August 2021, Sanjay will be a PhD student in computer science at UC Berkeley.
Haokun is an NYU graduate. He worked as PYI in the Semantic Scholar team from Jan 2021 to July 2021. Haokun authored around ten papers in natural language understanding evaluation, transfer learning and probing linguistic knowledge of pretrained models. Haokun will start his PhD at UNC-Chapel Hill in fall 2021. He is interested in transfer learning and empirical analysis in NLP.
Suchin graduated from the University of Chicago in 2014 with a bachelor’s in Mathematics, and earned a Master’s in Computational Linguistics at the University of Washington in 2018 after a few years in industry as a data scientist. He is now a PhD student at the University of Washington, advised by Noah Smith and Luke Zettlemoyer. At AI2, Suchin had the opportunity to do research in a number of different areas of NLP, including domain adaptation, safety, efficiency, and reproducibility.
Education: Past: Brown University, BSc Math and Computer Science. Graduated 2018 Current: University of Washington Computer Science PhD program, Fall 2020 Publications: What’s Hidden in a Randomly Weighted Neural Network? CVPR 2020, Soft Threshold Weight Reparameterization for Learnable Sparsity ICML 2020, Supermasks in Superposition ICML 2020 Workshop with spotlight (in submission to NeurIPS 2020), and Parameter Norm Growth during Training Transformers (in submission to NeurIPS 2020, no preprint link
Sarah Pratt completed her undergraduate studies at Brown University in Applied Mathematics and Computer Science before joining the PRIOR team at AI2. While at AI2, she worked on grounded situation recognition, which was published at ECCV 2020. She is now a PhD student at the University of Washington where she works with Prof. Ali Farhadi on computer vision and deep learning.
Chaitanya graduated with a masters degree from Carnegie Mellon University and joined the Mosaic team at AI2. At AI2, he worked on Commonsense knowledge graph construction: link 1 and link 2, Generative data augmentation, and Curation of a dataset for abductive reasoning. He will join the University of Pennsylvania as a PhD student in fall 2020.
Isabel is currently attending Johns Hopkins University as a PhD student in Computer Science, and did her Undergrad at the University of Texas at Austin. Her research interests include Natural Language Processing and Computational Social Science. Her published papers with AI2 include TLDR: Extreme Summarization of Scientific Documents, Citation Text Generation, and Improving the Accessibility of Scientific Documents: Current State, User Needs, and a System Solution to Enhance Scientific PDF Accessibility for Blind and Low Vision Users.
Jaemin graduated from Seoul National University in 2018 with a BS in Industrial Engineering. He will join a PhD program in Computer Science at UNC-Chapel Hill in the fall of 2020. While at AI2 he published Mixture Content Selection for Diverse Sequence Generation for EMNLP 2019 and X-LXMERT: Paint, Caption and Answer Questions with Multi-Modal Transformers for EMNLP 2020.
Mitchell is now a PhD student at the University of Washington working on neural networks with Ali Farhadi. At AI2, Mitchell explored self-adaptive navigation with Kiana Ehsani & Roozbeh Mottahgi and learning neural network wirings with Mohammad Rastegari. Mitchell is from Toronto and completed his undergraduate studies at Brown University in Applied Mathematics and Computer Science before joining AI2.
Kevin Lin joined the AllenNLP team at AI2 after graduating with a BS in Computer Science from Columbia University. At AI2 he worked on structured models for code generation and reading comprehension, as well as benchmarks for situated and qualitative reasoning. He started his PhD at UC Berkeley in 2019.
Noah Siegel was a Research Engineer at AI2 after receiving a BS in Computer Science and Economics from the University of Washington in 2015. He worked on computer vision and machine learning, specifically on applications to information extraction. Noah is currently a research engineer at DeepMind. He is first author on the paper FigureSeer: Parsing Result-Figures in Research Papers (ECCV 2016) and Extracting Scientific Figures with Distantly Supervised Neural Networks (JCDL 2018).
Roie Levin graduated from Brown University in 2015 with a BS in Applied Math/Computer Science and Math. He worked as a Research Engineer on the Euclid project at AI2. Roie accepted an offer of enrollment from CMU beginning in 2017. His publications with AI2 include FigureSeer: Parsing Result-Figures in Research Papers (ECCV 2016) and Beyond Sentential Semantic Parsing: Tackling the Math SAT with a Cascade of Tree Transducers (EMNLP 2017).
Chris Clark received a BS in Computer Science and a BA in Philosophy from the University of Washington, and went on to receive an MS in Artificial Intelligence from the University of Edinburgh where he studied machine learning and completed a thesis on using deep neural networks to play Go. Chris’s research at AI2 focused on information extraction, and his work was deployed as a key feature of the Semantic Scholar project. Chris enrolled in a PhD program at UW in 2015. He is first author on the papers PDFFigures 2.0: Mining Figures from Research Papers (JCDL 2016) and Looking Beyond Text: Extracting Figures, Tables and Captions from Computer Science Papers (AAAI Workshop on Scholarly Big Data 2015).