Natural Language to Structured Query Generation via Meta-Learning
About
In conventional supervised training, a model is trained to fit all the training examples. However, having a monolithic model may not always be the best strategy, as examples could vary widely. In this work, we explore a different learning protocol that treats each example as a unique pseudo-task, by reducing the original learning problem to a few-shot meta-learning scenario with the help of a domain-dependent relevance function. When evaluated on the WikiSQL dataset, our approach leads to faster convergence and achieves 1.1%-5.4% absolute accuracy gains over the non-meta-learning counterparts.
Po-Sen Huang, Chenglong Wang, Rishabh Singh, Wen-tau Yih, Xiaodong He• 2018
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Semantic Parsing | WikiSQL (test) | Execution Accuracy68 | 27 | |
| Natural Language to SQL | WikiSQL (test) | -- | 17 | |
| Semantic Parsing | WikiSQL (dev) | Accuracy68.3 | 13 | |
| Text-to-SQL | WikiSQL (dev) | Logic Form Acc63.1 | 8 | |
| Text-to-SQL | WikiSQL (test) | Logic Form Accuracy62.8 | 8 |
Showing 5 of 5 rows