Mar 22 (ML day)
Spotlights must be recorded by Mar 19 as an MP4 file and shared with the organizers. Please upload your video to the Google Drive folder linked in your paper acceptance email.
09:15am - 09:30am Opening remarks
Keynotes
09:30am - 10:30am Yizhou Sun
10:30am - 11:30am Yuandong Tian
11:30am - 12:30pm Phebe Vayanos
12:30pm - 02:00pm Lunch & Social
Session 1 (paper ID 01-10)
02:00pm - 02:30pm Spotlights (3-min presentations)
02:30pm - 03:00pm Poster Session 1
Session 2 (paper ID 11-20)
03:00pm - 03:30pm Spotlights (3-min presentations)
03:30pm - 04:00pm Poster Session 2
Session 3 (paper ID 21-30)
04:00pm - 04:30pm Spotlights (3-min presentations)
04:30pm - 05:00pm Poster Session 3
Mar 23 (NLP day)
Spotlights must be recorded by Mar 19 as an MP4 file and shared with the organizers. Please upload your video to the Google Drive folder linked in your paper acceptance email.
09:15am - 09:30am Opening remarks
Keynotes
09:30am - 10:30am Nanyun (Violet) Peng
10:30am - 11:30am William Wang
11:30am - 12:30pm Matt Gardner
12:30pm - 12:00pm Lunch
01:00pm - 01:50pm Social (Discussion around networking and career paths for under-represented minorities)
Panelists:
Ana Marasović: Ana Marasović (she/her) is a postdoctoral researcher at the University of Washington and at the AllenNLP team of the Allen Institute for Artificial Intelligence (AI2). Her interests broadly lie in the fields of natural language processing and explainable AI. She is currently working on developing and evaluating models with advanced reasoning abilities that provide readable explanations of their decision process. Dr. Marasović did her PhD in the Heidelberg University NLP Group where she worked on learning with limited labeled data for discourse-oriented tasks. Her advisor was Anette Frank. Prior to receiving her PhD in Janurary 2019, she completed B.Sc. (2013) and M.Sc. (2015) in Mathematics at the University of Zagreb.
Luca Soldaini: Luca Soldaini (he/him) is an applied scientist at Amazon Alexa Search in Manhattan Beach, California. His research efforts are currently focused on building ranking and generative models for natural language understanding tasks, such as open-domain question answering. Dr. Soldaini obtained a B.Eng. in Computer Engineering from the University of Florence in Italy, and he has a Ph.D. in Computer Science from Georgetown University. During his doctoral studies, he investigated natural language processing techniques to improve access to medical literature for both medical professionals and lay people.
Session 1 (paper ID 01-10)
02:00pm - 02:30pm Spotlights (3-min presentations)
02:30pm - 03:30pm Poster Session 1
Session 2 (paper ID 11-20)
03:30pm - 04:00pm Spotlights (3-min presentations)
04:00pm - 05:00pm Poster Session 2
Yizhou Sun
Title: Bringing Additional Symbolic Knowledge for Knowledge Graph Reasoning
Abstract: Knowledge graph has received tremendous attention recently, due to its wide applications, such as search engines and Q&A systems. Knowledge graph embedding, which aims at representing entities as low-dimensional vectors, and relations as operators on these vectors, has been widely studied and successfully applied to many tasks, such as knowledge completion. However, most of the existing knowledge graph embedding approaches treat knowledge graph as a complete, error-free, and flat data structure to store knowledge. In this talk, I will introduce two recent techniques developed in our lab to bring additional knowledge for better knowledge graph embedding. First, external knowledge represented as first-order logic is brought into knowledge graph embedding, which is able to address the uncertainty in knowledge graph and handle missing facts. Second, a unified embedding framework that incorporates ontological view KG into widely studied instance view KG will be introduced, which can seamlessly bring instance world and concept world together. Both techniques can significantly enhance the quality of KG embedding, on different downstream tasks, which also show a promising future direction in better knowledge graph reasoning.
Bio: Yizhou Sun is an associate professor at department of computer science of UCLA. She received her Ph.D. in Computer Science from the University of Illinois at Urbana-Champaign in 2012. Her principal research interest is on mining graphs/networks, and more generally in data mining, machine learning, and network science, with a focus on modeling novel problems and proposing scalable algorithms for large-scale, real-world applications. She is a pioneer researcher in mining heterogeneous information network, with a recent focus on deep learning on graphs/networks. Yizhou has over 100 publications in books, journals, and major conferences. Tutorials of her research have been given in many premier conferences. She received 2012 ACM SIGKDD Best Student Paper Award, 2013 ACM SIGKDD Doctoral Dissertation Award, 2013 Yahoo ACE (Academic Career Enhancement) Award, 2015 NSF CAREER Award, 2016 CS@ILLINOIS Distinguished Educator Award, 2018 Amazon Research Award, and 2019 Okawa Foundation Research Grant.
Matt Gardner
Title: Contrastive evaluation and learning in NLP
Abstract: Any dataset created by humans will almost unavoidably have spurious correlations between inputs and outputs. This means that when we collect data and split it into train and test sets, models that maximize the likelihood of the data will tend to find these spurious correlations, and they will use them to perform better than they should at test time. I will show that this problem is pervasive in natural language processing, extending even to traditional NLP tasks such as dependency parsing, and I will briefly demonstrate one method to partially solve this problem in our evaluations, by generalizing the long-standing notion of a "minimal pair". Solving the problem during training is more challenging. As a start, I will present work that leverages consistency on related examples during training to improve compositional reasoning in neural module networks. This is admittedly a very narrow solution to the problem, but it hints at how we might approach a more general solution.
Bio: Matt is a senior research scientist at the Allen Institute for AI on the AllenNLP team. His research focuses primarily on getting computers to read and answer questions, dealing both with open domain reading comprehension and with understanding question semantics in terms of some formal grounding (semantic parsing). He is particularly interested in cases where these two problems intersect, doing some kind of reasoning over open domain text. He is the original architect of the AllenNLP toolkit, and the instigator of the NLP Highlights podcast.
Yuandong Tian
Title: Understanding and Employing Learned Representation in Supervised, Self-supervised Learning and Decision-Making Process
Abstract: How to learn good latent representations is an important topic in the modern era of machine learning. Deep models excel because from raw data they learn a good representation on which the tasks become easier. Understanding the learned representations leads to strong model interpretability and better algorithms, and using a good representation makes the decision-making process more efficient. In this talk, I will cover our recent works on understanding representation in supervised and self-supervised learning by opening the black-box of the deep network, as well as strong performance in black-box optimization, if a good task-specific representation is learned during the optimization process.
Bio: Yuandong Tian is a Research Scientist and Manager in Facebook AI Research, working on deep reinforcement learning and representation learning. He is the lead scientist and engineer for ELF OpenGo and DarkForest Go projects. Prior to that, he was in Google Self-driving Car team in 2013-2014. He received a Ph.D in Robotics Institute, Carnegie Mellon University in 2013. He is the recipient of 2013 ICCV Marr Prize Honorable Mentions.
Nanyun (Violet) Peng
Title: Controllable Text Generation Beyond Auto-regressive Models
Abstract: Recent advances in large auto-regressive language models have demonstrated appealing results in generating natural languages and significantly improved the performances for applications such as machine translation and summarization. However, when the generation tasks are open-ended and the content is under-specified, existing techniques struggle to generalize to novel scenarios and generate long-term coherent and creative content. This happens because the generation models are trained to capture the surface patterns (i.e. sequences of words) following the left-to-right order, instead of capturing underlying semantics and discourse structures. In this talk, I will present our recent works on controllable text generation that go beyond the prevalent auto-regressive formulation. We explore hierarchical generation and insertion-based generation, with applications to creative story generation and image captioning.
Bio: Nanyun (Violet) Peng is an Assistant Professor of Computer Science at the University of California, Los Angeles. Prior to that, she spent three years at the University of Southern California's Information Sciences Institute as an Assistant Research Professor. She received her Ph.D. in Computer Science from Johns Hopkins University, Center for Language and Speech Processing. Her research focuses on the robustness and generalizability of NLP models, with applications to creative language generation and low-resource information extraction.
William Wang
Title: Learning to Reason with Text and Tables
Abstract: A key challenge for Artificial Intelligence is to design intelligent agents that can reason with heterogeneous representations. In this talk, I will describe our recent work on teaching machines to reason in semi-structured tables and unstructured text data. More specifically, I will introduce: (1) TabFact, a large benchmark dataset for table-based fact-checking; (2) HybridQA and OTT-QA, multi-hop question answering frameworks on tables and text; (3) How one can utilize TabFact to facilitate logical natural language generation with LogicNLG. I will also describe some other work at UCSB's NLP Group on learning to reason with multiple modalities.
Bio: William Wang is the Duncan and Suzanne Mellichamp Chair in Artificial Intelligence and Designs, and an Assistant Professor in the Department of Computer Science at the University of California, Santa Barbara. He is Director of UC Santa Barbara's Natural Language Processing group, and Center for Responsible Machine Learning. He received his PhD from Carnegie Mellon University. He has broad interests in machine learning and natural language processing, including statistical relational learning, information extraction, computational social science, and vision. He has published more than 100 papers at leading NLP/AI/ML/Vision conferences and journals, and received best paper awards (or nominations) at ASRU 2013, CIKM 2013, EMNLP 2015, and CVPR 2019, a DARPA Young Faculty Award (Class of 2018), IEEE Intelligent Systems AI's 10 to Watch (2020), an NSF CAREER Award (2021), and other faculty research awards from Google, Facebook, IBM, Amazon, JP Morgan Chase, Adobe, and Intel. His work and opinions appear at major tech media outlets such as Wired, VICE, Scientific American, Fortune, Fast Company, NPR, etc.
Phebe Vayanos
Title: Integer optimization for predictive and prescriptive analytics in high stakes domains
Abstract: Data-driven predictive and prescriptive analytics tools are increasingly being used to assist decision-making in high stakes domains (e.g., to prioritize people experiencing homelessness for scarce housing resources, to identify individuals at risk of suicide, and to design public health interventions). The deployment of such algorithms in these domains that can impact people’s lives and societal outcomes creates an urgent need for algorithms that are fair and interpretable and that leverage the available data to its full extent to yield the most accurate decisions. In this presentation, we discuss our recent works that leverage tools from integer optimization and causal inference to design optimal, interpretable, and fair decision-support tools that are suitable to deploy in high stakes settings.
Bio: Phebe Vayanos is an Assistant Professor of Industrial & Systems Engineering and Computer Science at the University of Southern California. She is also an Associate Director of CAIS, the Center for Artificial Intelligence in Society, an interdisciplinary research initiative between the schools of Engineering and Social Work at USC. Her research is focused on Operations Research and Artificial Intelligence and in particular on optimization and machine learning. Her work is motivated by problems that are important for social good, such as those arising in public housing allocation, public health, and biodiversity conservation. Prior to joining USC, she was lecturer in the Operations Research and Statistics Group at the MIT Sloan School of Management, and a postdoctoral research associate in the Operations Research Center at MIT. She holds a PhD degree in Operations Research and an MEng degree in Electrical & Electronic Engineering, both from Imperial College London. She served as a member of the ad hoc INFORMS AI Strategy Advisory Committee and is an elected member of the Committee on Stochastic Programming (COSP). She is a recipient of the INFORMS Diversity, Equity, and Inclusion Ambassador Program Award.