2019
DOI: 10.48550/arxiv.1911.04942
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

RAT-SQL: Relation-Aware Schema Encoding and Linking for Text-to-SQL Parsers

Abstract: When translating natural language questions into SQL queries to answer questions from a database, contemporary semantic parsing models struggle to generalize to unseen database schemas. The generalization challenge lies in (a) encoding the database relations in an accessible way for the semantic parser, and (b) modeling alignment between database columns and their mentions in a given query. We present a unified framework, based on the relation-aware self-attention mechanism, to address schema encoding, schema … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
65
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 30 publications
(66 citation statements)
references
References 15 publications
0
65
0
1
Order By: Relevance
“…P e (y|x t , i t ) = ∏ j=1..m P dec (y j |E(x t ), i, y <j ), but additionally uses i, the position of the non-terminal being expanded. We implement P dec (y j |E(x t ), i t , y <j ) with a (causal) relational transformer decoder, similar to Wang et al [36]. Relational transformers augment the attention mechanism by incorporating predefined relationships among elements.…”
Section: Grammformermentioning
confidence: 99%
“…P e (y|x t , i t ) = ∏ j=1..m P dec (y j |E(x t ), i, y <j ), but additionally uses i, the position of the non-terminal being expanded. We implement P dec (y j |E(x t ), i t , y <j ) with a (causal) relational transformer decoder, similar to Wang et al [36]. Relational transformers augment the attention mechanism by incorporating predefined relationships among elements.…”
Section: Grammformermentioning
confidence: 99%
“…For a multi-head attention layer, we replicate the update in equation 2 across all heads. With the aid of ground truth relations, (b ) has been used to modify the attention in the GREAT [29] and RAT-SQL [73] models, whereas the combination of (b ) and (d ) has been used in the Code Transformer model [84]. Since the edges we model are sparse, the additional term in equation 2 can be computed and backpropagated through with sparse primitives in standard automatic differentiation libraries.…”
Section: Using Predicted Relationsmentioning
confidence: 99%
“…Many machine learning models for code take into account relational structure. These models include variants of graph neural networks [3] and Transformers using relative positions [65,29,73,84]. Our focus is specifically on models that modify the standard attention computation [65,19] by altering q or k using edge embeddings.…”
Section: Introductionmentioning
confidence: 99%
“…Global gated graph neural network (Bogin et al, 2019) is designed to train the structure of database patterns and apply it in the encoding and decoding stages. Recently RAT-SQL (Wang et al, 2019) uses a relation-aware self-attention mechanism for schema encoding, feature representation and schema linking. It obtains the state-of-art accuracy of 65.6 on Spider test set.…”
Section: Related Workmentioning
confidence: 99%