Author:
Zhang Xu,Hu Xiaoyu,Liu Zejie,Xiang Yanzheng,Zhou Deyu
Abstract
Text-to-SQL, a computational linguistics task, seeks to facilitate the conversion of natural language queries into SQL queries. Recent methodologies have leveraged the concept of slot-filling in conjunction with predetermined SQL templates to effectively bridge the semantic gap between natural language questions and structured database queries, achieving commendable performance by harnessing the power of multi-task learning. However, employing identical features across diverse tasks is an ill-suited practice, fraught with inherent drawbacks. Firstly, based on our observation, there are clear boundaries in the natural language corresponding to SELECT and WHERE clauses. Secondly, the exclusive features integral to each subtask are inadequately emphasized and underutilized, thereby hampering the acquisition of discriminative features for each specific subtask. In an endeavor to rectify these issues, the present work introduces an innovative approach: the hierarchical feature decoupling model for SQL query generation from natural language. This novel approach involves the deliberate separation of features pertaining to subtasks within both SELECT and WHERE clauses, further dissociating these features at the subtask level to foster better model performance. Empirical results derived from experiments conducted on the WikiSQL benchmark dataset reveal the superiority of the proposed approach over several state-of-the-art baseline methods in the context of text-to-SQL query generation.
Reference34 articles.
1. Multitask learning;Caruana;Machine Learning,1997
2. S. Chang, P. Liu, Y. Tang, J. Huang, X. He and B. Zhou, Zero-shot text-to-sql learning with auxiliary task, in: Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 34, 2020, pp. 7488–7495.
3. Y. Chen, X. Guo, C. Wang, J. Qiu, G. Qi, M. Wang and H. Li, Leveraging table content for zero-shot text-to-sql with meta-learning, in: Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35, 2021, pp. 3992–4000.
4. Z. Chen, V. Badrinarayanan, C.-Y. Lee and A. Rabinovich, Gradnorm: Gradient normalization for adaptive loss balancing in deep multitask networks, in: International Conference on Machine Learning, PMLR, 2018, pp. 794–803.
5. L. Dong and M. Lapata, Coarse-to-fine decoding for neural semantic parsing, in: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2018, pp. 731–742.