Date: 9/19/19, 9 am to 5 pm
Location: MIT Samberg Conference Center
Room: Salon M (7th Floor)
Workshop Report
The workshop report is now available for download!
Detailed Schedule
Note: Apparantly there are issues with the links. Please access the video folder directly if you are not successful. We are working on this!
Event | Time |
---|---|
(Workshop Welcome) Kartik Talamadupula, video | |
(Keynote) Andrew McCallum (UMass Amherst), video | 09:00 - 09:45 |
L+R Graph Representations | 09:45 - 10:30 |
- Kelsey Allen (MIT), video | |
- Ferran Alet, (MIT), slides, video | |
- Lee Martie (IBM), slides, video | |
Coffee + KR2ML@IBM Posters | 10:30 - 11:00 |
(Keynote) Milind Tambe (Harvard), video | 11:00 - 11:45 |
L+R New Combinations | 11:45 - 12:30 |
- Masataro Asai (IBM), video | |
- Achille Fokoue (IBM), video | |
- Lucian Popa (IBM), slides, video | |
Lunch | 12:30 - 14:00 |
(Keynote) Tim Finin (UMBC), slides, video | 14:00 - 14:45 |
L+R Over Knowledge Graphs | 14:45 - 15:30 |
- Alfio Gliozzo (IBM), slides, video | |
- Ganesh Ramakrishnan (IIT Bombay), video | |
- Lingfei Wu (IBM), slides, video | |
Coffee + KR2ML@IBM Posters | 15:30 - 16:15 |
QA Challenge Session | 16:15 - 17:00 |
- Sumit Bhatia (IBM), video | |
- Challenge Participants | |
- Tengfei Ma (IBM), video | |
Workshop Closing |
Please also check out the rest of the AI Week program.
Keynotes
Andrew McCallum, UMass Amherst
Title: Embeddings of Language, Knowledge Representation, and Reasoning
Milind Tambe, Harvard
Title: AI for Social Good - Learning and Planning in the End-to-End, Data-to-Deployment Pipeline
Abstract: With the maturing of AI and multiagent systems research, we have a tremendous opportunity to direct these advances towards addressing complex societal problems. I will focus on the problems of public safety and security, wildlife conservation and public health in low-resource communities, and present research advances in multiagent systems to address one key cross-cutting challenge: how to strategically deploy our limited intervention resources in these problem domains. I will discuss the importance of conducting this research via building the full data to field deployment end-to-end pipeline rather than just building machine learning or planning components in isolation. Results from our deployments from around the world show concrete improvements over the state of the art. In pushing this research agenda, we believe AI can indeed play an important role in fighting social injustice and improving society.
Tim Finin, UMBC
Title: Fifty Years of Using Knowledge Graphs for Language Understanding
Abstract: There is a long history of using structured knowledge of one kind or another to support AI tasks and especially ones involving natural language understanding. Over the years the names and details of changed, from semantic networks to frames to logic programs to databases to expert systems to knowledge bases to semantic web and currently to knowledge graphs. But a common tread is that an organized representation of knowledge that can be queried and evolved is a core component of many intelligent systems. Current research is exploiting today’s two big technology trends: the ready availability of enormous amounts of content in the form of text, data, images and videos and powerful new machine learning technologies that can take advantage of it. After a brief review of this history, I’ll give some examples of combining KR and ML from our recent work.
Topical Sessions
L+R Graph Representations
Brandy Freitas, Pitney Bowes
Title: Graph Native Machine Learning
Abstract: Graph databases have become much more widely popularized in the recent year. By representing highly connected data in a graph, we have access to a host of graph native algorithms that depend on and exploit the relationships between our data. Computing and storing graph metrics can add strong new features to nodes, creating innovative predictors for machine learning. Using algorithms designed for path finding, centrality, community detection, and graph pattern matching, we can begin to rely less on inflexible, subject-driven feature engineering. In this session, Brandy Freitas will cover the interplay between graph analytics and machine learning, improved feature engineering with graph native algorithms, and outline the current use of graph technology in industry.
Kelsey Allen, MIT
Title: Embodied graph networks - learning graph-based models and policies for physical interaction
Abstract: Many tasks we care about in the physical environment can be represented as graphs, such as towers of blocks or even more general physical scenes. In this talk, I will describe our ongoing work to learn models and policies which enable artificial agents to build block towers, stabilize block towers with artificial “glue”, and perform long-range predictions of complex 2D and 3D shapes in artificial physical environments.
Ferran Alet, MIT
Title: Graph Element Networks
Abstract: We explore the use of graph neural networks (GNNs) to model spatial processes in which there is no a priori graphical structure. Similar to finite element analysis, we assign nodes of a GNN to spatial locations and use a computational process defined on the graph to model the relationship between an initial function defined over a space and a resulting function in the same space. We use GNNs as a computational substrate, and show that the locations of the nodes in space as well as their connectivity can be optimized to focus on the most complex parts of the space. Moreover, this representational strategy allows the learned input-output relationship to generalize over the size of the underlying space and run the same model at different levels of precision, trading computation for accuracy. We demonstrate this method on a traditional PDE problem, a physical prediction problem from robotics, and learning to predict scene images from novel viewpoints.
L+R New Combinations
Masataro Asai, IBM
Title: Classical Planning in Deep Latent Space: Bridging the Subsymbolic-Symbolic Boundary
Abstract: Current domain-independent, classical planners require symbolic models of the problem domain and instance as input, resulting in a knowledge acquisition bottleneck. Meanwhile, although deep learning has achieved significant success in many fields, the knowledge is encoded in a subsymbolic representation which is incompatible with symbolic systems such as planners. We introduce LatPlan, an unsupervised architecture combining deep learning and classical planning. Given only an unlabeled set of image pairs showing a subset of transitions allowed in the environment (training inputs), and a pair of images representing the initial and the goal states (planning inputs), LatPlan finds a plan to the goal state in a symbolic latent space and returns a visualized plan execution.
Lee Martie, IBM
Title: Reflecting After Learning for Understanding
Abstract: Today, image classification is a common way for systems to describe images. While neural network approaches to classification have seen great progress in reducing error rates, it is not clear what this means for a cognitive system needing to make sense of the multiple and competing labels from its own classifiers for image understanding. As a step to address this, we present a novel approach and framework using meta-reasoning and meta-operations to unify labels into higher-level knowledge for the cognitive system. Using our framework on images from ImageNet, we demonstrate systems capable of unifying 41% to 46% of labels and 67% to 75% of labels when these systems can explain their differences. Further, we demonstrate such systems in “the wild” by feeding live video images through one and show the system unifying 51% of labels and 69% of labels when the system can explain their differences. In a survey given to 24 participants, it was found that 87% of the unified labels made with our system describe their corresponding images.
Lucian Popa, IBM
Title: Human-Machine Co-Creation of Explainable Models
Abstract: While the role of humans is increasingly recognized in machine learning community, representation of and interaction with models in current human-in-the-loop machine learning (HITL-ML) approaches are too low-level and far-removed from human’s conceptual models. I will present ongoing work in my team to support human-machine co-creation with learning and human-in-the-loop techniques. In particular, I will focus on HEIDL, a system supports human-machine co-creation by exposing the machine-learned model through high-level, explainable linguistic expressions. In HEIDL, human’s role is elevated from simply evaluating model predictions to interpreting and even updating the model logic directly by enabling interaction with rule predicates themselves. Raising the currency of interaction to such semantic levels calls for new interaction paradigms between humans and machines that result in improved productivity for model development process. Moreover, by involving humans in the process, the human-machine co-created models generalize better to unseen data as domain experts are able to instill their expertise by extrapolating from what has been learned by automated algorithms from few labelled data.
L+R Over Graphs
Alfio Gliozzo, IBM
Title: Knowledge Induction @ IBM Research AI
Ganesh Ramakrishnan, IIT Bombay
Title: Knowledge Understanding and Representation for Automatic Question Generation
Lingfei Wu, IBM
Title: Exploiting Graph Neural Networks with Context Information for RDF- to-Text Generation
Abstract: The task of RDF-to-Text generation is to generate a corresponding descriptive text given a set of RDF triplets. Most of the previous approaches either cast this task as a sequence- to-sequence problem or employ graph-based encoders for modeling RDF triplets and decode a text sequence. However, none of these methods can explicitly model both global and local structure information within and between the triplets. Furthermore, they did not exploit the target text as an important additional context for modeling complex RDF triplets. To address these issues, we propose to jointly learn local information and global structure information via combining graph encoder and graph-based triple encoder for the input triplets. Furthermore, we also exploit Seq2Seq-based auto-encoder to leverage target text as the context to supervise the combination of input encoders. Experimental results on the WebNLG dataset show that our proposed model outperforms the state-of-the-art baselines.
Achille Fokoue, IBM
Title: TRAIL: A Deep Reinforcement Learning Approach to First-Order Logic Reasoning
Abstract: Traditional first-order logic (FOL) reasoning systems usually rely upon manual heuristics to aid in their search for proofs or solutions. We propose TRAIL: a flexible, deep reinforcement learning based FOL reasoner that circumvents the need for customized heuristics by learning to improve its search and decision-making process. TRAIL relies upon a traditional reasoner and a learning agent that are agnostic of one another, an architecture that allows for developing neural reasoners for any logic. We demonstrate that TRAIL is comparable to state of the art FOL reasoners.
Tengfei Ma, IBM
Title: Graph Enhanced Cross-Domain Text-to-SQL Generation
Abstract: Semantic parsing is a fundamental problem in natural language understanding, as it involves the mapping of natural language to structured forms such as executable queries or logic-like knowledge representations. Existing deep learning approaches for semantic parsing have shown promise on a variety of benchmark data sets, particularly on text-to-SQL parsing. However, most text-to-SQL parsers do not generalize to unseen data sets in different domains. In this paper, we propose a new cross-domain learning scheme to perform text-to-SQL translation and demonstrate its use on Spider, a large-scale cross-domain text-to-SQL data set. We improve upon the SyntaxSQLNet, by constructing a graph of column names for all databases and using graph neural networks to compute their embeddings. The resulting embeddings offer better cross-domain representations and SQL queries, as evidenced by substantial improvement on the Spider data set.