Natural Language to SQL(nl2sql)

Slides:



Advertisements
Similar presentations
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
Advertisements

PDDL: A Language with a Purpose? Lee McCluskey Department of Computing and Mathematical Sciences, The University of Huddersfield.
Distributed Representations of Sentences and Documents
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Eric H. Huang, Richard Socher, Christopher D. Manning, Andrew Y. Ng Computer Science Department, Stanford University, Stanford, CA 94305, USA ImprovingWord.
Describing Images using Inferred Visual Dependency Representations Authors : Desmond Elliot & Arjen P. de Vries Presentation of Paper by : Jantre Sanket.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Neural Networks in Computer Science n CS/PY 231 Lab Presentation # 1 n January 14, 2005 n Mount Union College.
Addressing the Rare Word Problem in Neural Machine Translation
The Database and Info. Systems Lab. University of Illinois at Urbana-Champaign Understanding Web Query Interfaces: Best-Efforts Parsing with Hidden Syntax.
Convolutional LSTM Networks for Subcellular Localization of Proteins
BOOTSTRAPPING INFORMATION EXTRACTION FROM SEMI-STRUCTURED WEB PAGES Andrew Carson and Charles Schafer.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
Short Text Similarity with Word Embedding Date: 2016/03/28 Author: Tom Kenter, Maarten de Rijke Source: CIKM’15 Advisor: Jia-Ling Koh Speaker: Chih-Hsuan.
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation EMNLP’14 paper by Kyunghyun Cho, et al.
Naifan Zhuang, Jun Ye, Kien A. Hua
Attention Model in NLP Jichuan ZENG.
Olivier Siohan David Rybach
Neural Machine Translation
Convolutional Sequence to Sequence Learning
Unsupervised Learning of Video Representations using LSTMs
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
Deep Learning Amin Sobhani.
Sentence Modeling Representation of sentences is the heart of Natural Language Processing A sentence model is a representation and analysis of semantic.
Recursive Neural Networks
Adversarial Learning for Neural Dialogue Generation
Neural Machine Translation by Jointly Learning to Align and Translate
Intro to NLP and Deep Learning
Natural Language Processing of Knee MRI Reports
Neural networks (3) Regularization Autoencoder
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
Neural Machine Translation By Learning to Jointly Align and Translate
Attention Is All You Need
Attention-based Caption Description Mun Jonghwan.
Chinese Poetry Generation with Planning based Neural Network
Final Presentation: Neural Network Doc Summarization
Goodfellow: Chapter 14 Autoencoders
Code Completion with Neural Attention and Pointer Networks
MATERIAL Resources for Cross-Lingual Information Retrieval
Introduction to Natural Language Processing
AST Based Sequence to Sequence Natural Language Question to SQL Method
Ying Dai Faculty of software and information science,
Machine Translation(MT)
Domain Mixing for Chinese-English Neural Machine Translation
Dialogue State Tracking & Dialogue Corpus Survey
Unsupervised Pretraining for Semantic Parsing
Attention.
Independent Project Natural Language to SQL
Using Multilingual Neural Re-ranking Models for Low Resource Target Languages in Cross-lingual Document Detection Using Multilingual Neural Re-ranking.
Dennis Zhao,1 Dragomir Radev PhD1 LILY Lab
Sequential Question Answering Using Neural Coreference Resolution
S.N.U. EECS Jeong-Jin Lee Eui-Taik Na
Deep Learning for the Soft Cutoff Problem
Neural Modular Networks
Attention for translation
Automatic Handwriting Generation
A unified extension of lstm to deep network
Question Answering System
Neural Machine Translation
Presenter: Yu Chen Computer Science Department
Week 3 Presentation Ngoc Ta Aidean Sharghi.
Sequence-to-Sequence Models
Week 7 Presentation Ngoc Ta Aidean Sharghi
Neural Machine Translation by Jointly Learning to Align and Translate
Listen Attend and Spell – a brief introduction
Visual Grounding.
Goodfellow: Chapter 14 Autoencoders
Presentation transcript:

Natural Language to SQL(nl2sql) Zilin Zhang LILY Lab Department of Computer Science, Yale University Introduction Results The project of Natural Language to SQL(nl2sql) focuses on extracting SQL query from natural language problem with database schema. In this project, I helped annotated databases with questions and SQL labels. I also wrote SQL parser which generates training and dev data for the project. The problem of nl2sql can be issued as semantic parsing problem. Semantic parsing aims at mapping natural language to machine interpretable meaning representations. Traditional approaches rely on high-quality lexicons, manually-built templates, and linguistic features which are either domainor representation-specific. In this project, we borrow the idea of seq2seq and seq2tree and present a general method based on an attention-enhanced encoder-decoder model. We encode input utterances into vector representations, and generate their sql by conditioning the output sequences or trees on the encoding vectors. Figure 1. Input utterances and their logical forms are encoded and decoded with neural networks. An attention layer is used to learn soft alignments. Table 1. Results of models in Geoquery dataset The above table shows that seq2tree model is superior to seq2seq model. This is expeced since seq2tree explicitly models compositional structure. In Geoquery(split column), seq2tree outperforms seq2seq by 30% and seq2seq+attention by 8%. In Geoquery(split question), seq2tree outperforms seq2seq by 25% and seq2seq+attention by 6%. We find that adding attention can substantially imrpoves performance. Figure 3. Sequence-to-tree (SEQ2TREE) model with a hierarchical tree decoder. Materials and Methods Conclusion We aim to learn a model which maps natural language input to a SQL query representation. Our model consists of an encoder which encodes natural language into a vector representation and a decoder which generates SQL query based on the encoding vector. In this project, we come up with two models, sequence-to-sequence (seq2seq) model and sequence-to-tree (seq2tree) model respectively. The encoder and decoder of seq2seq are two different L-layer recurrent neural networks with long short-term memory(LSTM) until which recursively process tokens one by one. However, the seq2seq model ignores the hierarchical structure of SQL query. In order to solve this question, we come up with seq2tree model, which consists of a hierarchical tree decoder. We also adopt attention mechanism to better predict the current token. In this project we presented an encoder-decoder neural netowork model for mapping natural language to SQL query. We first encode the input natural language utterances into vectors and then use decoder to generate their corresponding SQL query. Considering the drawback of basic sequence decoder, we introduce a new kind decoder which takes hiearacture into consideration. And we also use an attention mechanism to improve performance. The result shows that the approach performs competitively wihout domain or representation specific features. Figure 2. Sequence-to-sequence (SEQ2SEQ) model with two- layer recurrent neural networks. Acknowledgement Figure 5. Attention scores are computed by the current hidden vector and all the hidden vectors of encoder. Then, the encoder-side context vector ct is obtained in the form of a weighted sum, which is further used to predict yt. I sincerely appreciate the instruction from professor Drago and the help from Tao Yu and Rui Zhang. Figure 4. A SEQ2TREE decoding example for the logical form “A B (C)”.