Differentiable Learning of Logical Rules for Knowledge Base Reasoning
About
We study the problem of learning probabilistic first-order logical rules for knowledge base reasoning. This learning problem is difficult because it requires learning the parameters in a continuous space as well as the structure in a discrete space. We propose a framework, Neural Logic Programming, that combines the parameter and structure learning of first-order logical rules in an end-to-end differentiable model. This approach is inspired by a recently-developed differentiable logic called TensorLog, where inference tasks can be compiled into sequences of differentiable operations. We design a neural controller system that learns to compose these operations. Empirically, our method outperforms prior work on multiple knowledge base benchmark datasets, including Freebase and WikiMovies.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Link Prediction | FB15k-237 (test) | Hits@1040.8 | 419 | |
| Link Prediction | WN18RR (test) | Hits@1056.6 | 380 | |
| Link Prediction | FB15k-237 | MRR25 | 280 | |
| Knowledge Graph Completion | FB15k-237 (test) | MRR0.24 | 179 | |
| Knowledge Graph Completion | WN18RR (test) | MRR0.435 | 177 | |
| Knowledge Graph Completion | WN18RR | Hits@136.8 | 165 | |
| Link Prediction | FB15K (test) | Hits@100.837 | 164 | |
| Link Prediction | WN18 (test) | Hits@100.945 | 142 | |
| Knowledge Graph Completion | FB15k-237 | Hits@100.361 | 108 | |
| Knowledge Graph Completion | WN18 (test) | Hits@100.945 | 80 |