Syntax_and_Relation_Enhanced_Query_Generation_for_
Syntax_and_Relation_Enhanced_Query_Generation_for_
ISSN: 2710-0588
Abstract
Text-to-SQL parsing is the task of converting natural language questions into executable
SQL queries, a significant branch of semantic parsing, which has gained increasing
attention in recent years. This technology lowers the barrier for people to access
databases, enhancing the convenience and availability of data. However, the primary
challenge for text-to-SQL parsing lies in domain adaptation, which concerns whether the
model can be applied to new databases and effectively align natural language questions
with the corresponding tables or columns within the database. To address these issues,
research has introduced SRSQL (Syntax and Relation-Augmented Query Generation),
which incorporates syntax information and predefined relationships into the model,
effectively utilizing syntactic dependencies and pattern linking to improve performance.
Using a Transformer-based decoder, SRSQL generates SQL queries in the form of
Abstract Syntax Trees (AST), significantly boosting prediction accuracy. Experimental
results show that SRSQL outperforms comparative models, particularly on challenging
benchmarks like Spider and Spider-SYN.
Keywords
Text-to-SQL; Syntactic Dependencies; Transformer; Abstract Syntax Trees.
1. Introduction
As the era of information technology advances, vast amounts of data are generated daily from
individuals and businesses. To facilitate management and operations, these data are commonly
stored in relational databases. To retrieve the needed data from these databases, Structured
Query Language (SQL) or similar structured query languages are employed. Despite the
growing popularity of relational databases, non-expert users often face limitations in accessing
information due to the need to understand complex query languages, creating a certain level of
usability barrier. Consequently, Text-to-SQL [1] parsing has garnered significant attention,
aiming to directly convert natural language questions into corresponding SQL query statements.
This task alleviates the challenges faced by non-expert users when interacting with relational
databases, to a certain extent.
Developing large-scale annotated question datasets along with corresponding SQL queries has
propelled advancements in the field. In contrast to previous efforts on parsed datasets [2], the
new datasets such as WikiSQL [3] and Spider [4] heavily test models' ability to generalize to
unseen database schemas. Each query in these new tasks is based on a multi-table database
architecture, and database schemas do not repeat between training and test sets.
A key issue in achieving domain generalization [5] is the need for complex reasoning to generate
structurally rich SQL queries. To accurately contextualize user queries with specific databases,
this involves explicit relationships (such as those defined by the database schema for tables and
82
Frontiers in Science and Engineering Volume 4 Issue 4, 2024
ISSN: 2710-0588
columns) and implicit relationships (such as determining whether phrases in the query
correspond to specific columns or tables).
However, this implies that models need to predict queries in database contexts unseen during
training and accurately express query intentions through SQL logic. Therefore, text-to-SQL
parsers across databases cannot solely rely on observed SQL schema; they must also accurately
model natural language questions, database structures, and the context between them.
Current research typically adopts the following strategies to enhance the cross-domain
generalization ability of models. First, by learning schema-based embedding functions,
questions and database schemas are contextualized mutually [6]. Secondly, pretrained language
models such as BERT [7] and RoBERTa [8] have been shown to improve prediction accuracy by
acquiring semantic relationships in different contexts and capturing distant dependencies.
Methods that incorporate syntax-enhanced synthetic examples into the BERT pretraining
framework, alongside a basic semantic parser, have shown promising results.
This study introduces SRSQL, which integrates joint encoding of questions and schemas into a
novel Transformer [9] variant for text-to-SQL parsing. By representing each natural language
question as a graph with multiple relations, including syntactic dependencies and part-of-
speech, and a database schema as a graph composed of tables, columns, and their relationships,
SRSQL employs a relation-aware Transformer to learn the connections between the schema
and the question. A Transformer-based tree decoder is then proposed to generate SQL queries
in Abstract Syntax Tree (AST) form. Experiments on the Spider benchmark show that SRSQL
attains a 74.5% Exact Match (EM) accuracy on the test set, outperforming the baseline model
significantly.
2. Related Work
In earlier research, a commonly used approach was sketch-based slot filling, which employs
different modules to predict various parts of the generated SQL query. This method
decomposes the SQL generation task into several independent sketches and utilizes different
classifiers to predict each part, such as SQLNet [10], TypeSQL [11], RYANSQL [12], X-SQL [13], among
others. Most of these methods only handle simple queries, making it challenging to apply them
to more complex scenarios.
There are various methods to address the challenges posed by complex SQL tasks. One common
approach is to use generic neural network-enhanced encoders for global reasoning over natural
language questions and database schemas. For instance, IRNet [14] employs LSTM and self-
attention mechanisms to encode questions and schemas separately, while BRIDGE [15] serializes
questions and schemas into token sequences and maximizes the utilization of BERT and
database content to capture the linking relationships between questions and schemas.
On the other hand, many studies utilize graph structures to represent a series of complex
relationships. For example, Global-GNN [16] employs graph neural networks (GNN) to output
queries selecting subsets of tables or columns, while ShadowGNN [17] introduces a graph
projection neural network to abstract representations of questions and schemas. Further
developments include SADGA [18], which separately encodes question graphs and schema
graphs based on dependency relationship structures and database schemas, and SDSQL [19],
which improves structured reasoning by modeling relationships between schemas and
questions.
Recent work has demonstrated the effectiveness of fine-tuning pretrained models. For instance,
Shwa et al. [20] showed that fine-tuning the pretrained T5-3B model can yield highly competitive
results. Building upon this, PICARD [21] was introduced, a technique that restricts the
autoregressive decoder by applying incremental parsing during inference. It real-time filters
83
Frontiers in Science and Engineering Volume 4 Issue 4, 2024
ISSN: 2710-0588
out syntactically incorrect sequences during beam search, significantly enhancing the quality
of generated SQL.
In summary, due to the complexity of semantic understanding and data dependency
relationships, generating complex query statements can lead to syntax errors, inappropriate
semantics, and database connection errors. The SRSQL model proposed in this paper effectively
alleviates these issues.
3. Question Definition
The task aims to transform the input natural language question Q into the corresponding
Abstract Syntax Tree (AST) representation of the SQL query y, given a database schema S= (T,
C).
Specifically, Let Q = (q1, q2, …, q|Q|) be a sequence of natural language tokens. |Q| is the length of
the natural language problem. The database schema S includes multiple tables T = (t1, t2, …, t|T|)
and column C = (c1, c2, …, c|C|), |T| and |C| respectively represent the number of tables and
columns in the database. The table name for each table is denoted as 𝑡 ∈ 𝑇 , it can be
represented as a sequence of tokens ti = ti1 ,ti2 ,⋯,ti|ti | , |ti| represents the number of tokens for
table names, Similarly, each column name in table t is represented as ci ={ci1 ,ci2 ,⋯,ci|ci | }.|ci|
represents the number of tokens for column names.
4. Model
4.1. Model Framework
The overall structure of the SRSQL model is illustrated in Figure 1. Model inherits the end-to-
end capabilities of Transformer and treats the text-to-SQL problem as a translation task. The
model consists of L layers of encoders and decoders, with the encoding of the question and the
database schema concatenated as the model's input. We utilize a relation-aware Transformer
[22] as the encoder (see Section 4.3), which employs relation-aware self-attention [23] to replace
the original self-attention mechanism. Additionally, we extend the Transformer decoder (see
Section 4.4), integrating node types and embeddings of previous actions to autoregressively
generate SQL queries. These queries are a series of actions derived according to SQL syntax.
84
Frontiers in Science and Engineering Volume 4 Issue 4, 2024
ISSN: 2710-0588
Database Schema Graph Construction: The database schema graph is represented as a graph GS
= (S, R), where Node Set S = (T, C) representing tables T and columns C in the database schema.
The text is written in Chinese. Here's the translation into English: R={r1 ,…,r|R| }represents the
structural relationship between entities and attributes in a schema. We employ classic
database-specific relationships, such as whether a column belongs to a table, whether it is a
primary key of that table, and whether it is a foreign key referencing another column. Figure 3
illustrates an example of a database schema diagram. We also encode the database schema
diagrams using GAT [24] and obtain vector representations for each database schema through
global average pooling.
85
Frontiers in Science and Engineering Volume 4 Issue 4, 2024
ISSN: 2710-0588
elements within the input and predefined relations. We utilize it to encode the linkages between
questions and database schemas, which will be elaborated on in Section 4.4.
Suppose that the input to each encoder layer in the original Transformer model is an n-
dimensional word vector sequence X=(x1 ,…,xn ), where 𝑥 ∈ 𝑅 represents the input, and the
output is a new sequence y where 𝑦 ∈ 𝑅 . The model consists of multiple stacked self-
attention layers, each containing H attention heads. The output element y is calculated by first
linearly transforming the input elements x and then performing a weighted sum using self-
attention, following this computational process:
( ) ( )
( ) 𝑥𝑊 (𝑥 𝑊 ) ( ) ( )
𝑒 = ,𝛼 = 𝑠𝑜𝑓𝑡𝑚𝑎𝑥 𝑒 (1)
𝑑 /𝐻
( ) ( ) ( ) ( ) ( )
𝑧 = 𝛼 (𝑥 𝑊 ) , 𝑧 = 𝐶𝑜𝑛𝑐𝑎𝑡(𝑧 ,⋯,𝑧 ) (2)
𝑦 = 𝐿𝑎𝑦𝑒𝑟𝑁𝑜𝑟𝑚(𝑥 + 𝑧 ) (3)
( ) ( )
( ) 𝑥𝑊 (𝑥 𝑊 +𝑟 )
𝑒 = (5)
𝑑 /𝐻
( ) ( ) ( )
𝑧 = 𝛼 (𝑥 𝑊 +𝑟 ) (6)
86
Frontiers in Science and Engineering Volume 4 Issue 4, 2024
ISSN: 2710-0588
87
Frontiers in Science and Engineering Volume 4 Issue 4, 2024
ISSN: 2710-0588
In the SRSQL decoder, each node possesses two attributes: a node type and the action from the
previous time step. We denote the vector representation of the current node type as 𝑚 , and
the vector representation of the action from the previous time step as 𝑎 .
Specifically, node types include SQL keywords, table names, and column names from the
database. Decoding actions are divided into two categories: (1) applying the generated rules to
the current syntax tree, which is the APPLYRULE action, and (2) selecting a table or column
from the database schema, which are the SELECTTABLE and SELECTCOLUMN actions. The
process of generating the Abstract Syntax Tree (AST) involves sequential application of these
actions, with a depth-first traversal order for constructing a SQL query y. As shown in Figure 5,
SQL statements based on the AST are generated using context-free grammar.
When incorporating 𝑚 as a bias term in the calculation of attention, the formula takes the
following form:
𝑄𝐾
𝐴= +𝑈 (7)
𝑑
𝑈 =𝑊 ×𝑚 (8)
Here, the query vector Q is derived from the attention vector 𝑎 , while the key vector K and
value vector V come from the hidden states of the encoder. The vector U is derived from the
node type vector 𝑚 , with 𝑊 being a learnable weight matrix. In the l-th layer, the residual
update for the node vector 𝑛 can be represented as follows, where || denotes concatenation,
and H is the number of attention heads:
𝑛 =𝑛 + 𝑂 || 𝑠𝑜𝑓𝑡𝑚𝑎𝑥 𝐴 𝑉 (9)
88
Frontiers in Science and Engineering Volume 4 Issue 4, 2024
ISSN: 2710-0588
Consequently, after passing through L layers of the decoder, the decoder state 𝑛 at the current
time step is fed into an action output MLP (Multi Layer Perceptron), which computes the
probability distribution 𝑃(𝑎 ) for the APPLYRULE action at that step, with the specific
calculation formula being:
Here, g (⋅)represents an MLP with a tanh activation function, and 𝑎 denotes all action pairs up
to time step t for the SELECTTABLE action. We calculate using the following formula:
The calculation method for the "SELECT COLUMN" action is similar, and the prediction of the
AST form of the final SQL query can be decoupled into a sequence of actions 𝑎 = (𝑎 , ⋯ , 𝑎| | ),
with the resulting training objective being:
| |
5. Experiment
5.1. Dataset
Our experiment employed the Spider and Spider-Syn datasets [25]. The Spider dataset is a large,
complex, and cross-domain semantic parsing and text-to-SQL dataset, consisting of nine classic
datasets like Scholar [26], WikiSQL, GeoQuery [27], etc. It contains 8,659 training samples, 1,034
development samples, and 2,147 test samples, spanning 138 domains across 200 complex
databases. The Spider-Syn dataset, derived from the Spider benchmark, is a manually curated
dataset where NL problems are modified from Spider by replacing words related to patterns
with synonyms chosen to reflect real-world problem interpretations. It consists of 7,000
training samples and 1,034 development samples.
5.2. Assessment Metrics
Following Yu et al.'s [3] metrics, we compute the Exact Match accuracy (EM) for all examples,
grouped by difficulty. This is done by dividing the predicted SQL and the actual SQL into distinct
subsets based on keywords, and then checking if the predicted set matches the actual one. EM
assesses whether the predicted SQL query matches the ground truth exactly. Like previous
work on Spider, these metrics do not consider the model's performance in generating values
within the SQL.
5.3. Model Configuration
We utilized stanza [28] for tokenization, word segmentation, part-of-speech tagging, and
dependency parsing. For training, we set the maximum input length to 1024, the maximum
number of generated nodes in the AST to 200, the batch size to 32, and the maximum training
steps to 40,000. The encoder and decoder had 6 layers with a dimension of 512 and 8 attention
heads. The vector dimensions for tables and columns were set to 512, and the embeddings for
node types and actions were of size 512. We employed Adafactor as the optimizer, with a
learning rate of 1e-4 and a dropout rate of 0.1.
89
Frontiers in Science and Engineering Volume 4 Issue 4, 2024
ISSN: 2710-0588
90
Frontiers in Science and Engineering Volume 4 Issue 4, 2024
ISSN: 2710-0588
Table 4. Ablation study of EM accuracy on the development set for SRSQL (±95% confidence
interval)
Method Scheme 1
SRSQL 77.2 ± 0.76
SRSQL w/o syntactic dependency 76.1 ± 0.73
SRSQL w/o Part-of-Speech tagging 76.3 ± 0.56
SRSQL w/o schema linking relations 73.8 ± 0.80
SRSQL encoder + LSTM-based decoder 72.9 ± 0.38
As shown in Table 4, the most significant impact among these designs came from the choice of
decoder. After switching to an LSTM-based decoder, the EM accuracy dropped from 77.2% to
72.9%, resulting in a 4.3% decrease in performance, highlighting the superiority of a
Transformer-based decoder. Second, the removal of pattern linking had a considerable effect
on the model, with EM accuracy dropping by 3.4%. This is because the task of matching
questions to database schemas became more challenging, and previous research has already
confirmed the importance of this component for text-to-SQL parsing. Lastly, removing syntax
dependencies and part-of-speech tagging had a smaller effect on performance, with decreases
of 1.1% and 0.9% respectively.
6. Conclusion
In this paper, we present SRSQL, a syntax and relation-enhanced text-to-SQL parser that stands
out with its autoregressive SQL query prediction based on Transformers. By incorporating
relation-aware self-attention, SRSQL integrates pattern linking relationships into its encoder.
The Transformer-based tree decoder, grounded in joint encoding, integrates node types and
prior actions to generate SQL queries during the learning process. Notably, SRSQL
demonstrates state-of-the-art performance on the Spider and Spider-SYN datasets. However,
our model still has some limitations, with this study primarily focusing on the grammatical
aspects of text-to-SQL conversion. Future work could explore incorporating large pre-trained
language models or leveraging techniques from large-scale prompt-based models.
References
[1] John M. Zelle and Raymond J. Mooney: Learning to parse database queries using inductive logic
programming. Proceedings of the thirteenth national conference on Artificial intelligence (Portland,
Oregon, 1996). Vol.2, p1050–1055.
[2] Catherine Finegan-Dollak, Jonathan K. Kummerfeld, Li Zhang, Karthik Ramanathan, Sesh Sadasivam,
Rui Zhang, and Dragomir Radev: Improving Text-to-SQL Evaluation Methodology. Proceedings of
the 56th Annual Meeting of the Association for Computational Linguistics (Melbourne, Australia,
July ,2018). Vol.1, p351–360.
[3] Victor Zhong, Caiming Xiong, and Richard Socher: Seq2SQL: Generating Structured Queries from
Natural Language using Reinforcement Learning.arXiv:1709.00103, 2017.
91
Frontiers in Science and Engineering Volume 4 Issue 4, 2024
ISSN: 2710-0588
[4] Tao Yu, Rui Zhang, Kai Yang, Michihiro Yasunaga, Dongxu Wang, Zifan Li, James Ma, Irene Li,
Qingning Yao, Shanelle Roman, Zilin Zhang, and Dragomir Radev: Spider: A Large-Scale Human-
Labeled Dataset for Complex and Cross-Domain Semantic Parsing and Text-to-SQL Task.
Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing
(Brussels, Belgium, October-November ,2018), p3911–3921.
[5] Binyuan Hui, Ruiying Geng, Lihan Wang, Bowen Qin, Bowen Li, Jian Sun, and Yongbin Li: S2 SQL:
Injecting syntax to question-schema interaction graph encoder for text-to-sql parsers.2022.
arXiv:2203.06958.
[6] Wonseok Hwang, Jinyeong Yim, Seunghyun Park, and Minjoon Seo: A Comprehensive Exploration
on WikiSQL with Table-Aware Word Contextualization. arXiv:1902.01069, 2019.
[7] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova: BERT: Pre-training of Deep
Bidirectional Transformers for Language Understanding. Proceedings of the 2019 Conference of the
North American Chapter of the Association for Computational Linguistics: Human Language
Technologies (Minneapolis, Minnesota, June,2019).
[8] Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis,
Luke Zettlemoyer, and Veselin Stoyanov: RoBERTa: A Robustly Optimized BERT Pretraining
Approach. arXiv:1907.11692, 2019.
[9] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz
Kaiser, and Illia Polosukhin: Attention Is All You Need. arXiv:1706.03762, 2017.
[10] Xiaojun Xu, Chang Liu, and Dawn Song: SQLNet: Generating Structured Queries From Natural
Language Without Reinforcement Learning.arXiv:1711.04436, 2017.
[11] Tao Yu, Zifan Li, Zilin Zhang, Rui Zhang, and Dragomir Radev: TypeSQL: Knowledge-based Type-
Aware Neural Text-to-SQL Generation.ArXiv:1804.09769, 2018.
[12] DongHyun Choi, Myeong Cheol Shin, EungGyun Kim, and Dong Ryeol Shin: RYANSQL: Recursively
Applying Sketch-based Slot Fillings for Complex Text-to-SQL in Cross-Domain Databases.ArXiv:
2004.03125,2020.
[13] Pengcheng He, Yi Mao, Kaushik Chakrabarti, and Weizhu Chen: X-SQL: reinforce schema
representation with context. ArXiv:1908.08113, 2019.
[14] Jiaqi Guo, Zecheng Zhan, Yan Gao, Yan Xiao,Jian-Guang Lou, Ting Liu, and Dongmei Zhang: Towards
Complex Text-to-SQL in Cross-Domain Database with Intermediate Representation. ArXiv:1905.
08205, 2019.
[15] Xi Victoria Lin, Richard Socher, and Caiming Xiong: Bridging textual and tabular data for
crossdomain text-to-SQL semantic parsing. Findings of the Association for Computational
Linguistics: EMNLP 2020(Online, November,2020), p4870–4888.
[16] Ben Bogin, Matt Gardner, and Jonathan Berant: Global reasoning over database structures for text-
to-SQL parsing. Proceedings of the 2019 Conference on Empirical Methods in Natural Language
Processing and the 9th International Joint Conference on Natural Language Processing (Hong Kong,
China, November,2019), p3657–3662.
[17] Zhi Chen, Lu Chen, Yanbin Zhao, Ruisheng Cao, Zihan Xu, Su Zhu, and Kai Yu: Shadowgnn: Graph
projection neural network for text-to-sql parser. Proceedings of the 2021 Conference of the North
American Chapter of the Association for Computational Linguistics: Human Language
Technologies(Online, June,2021), p 5567–5577.
[18] Ruichu Cai, Jinjie Yuan, Boyan Xu, and Zhifeng Hao: SADGA: Structure-aware dual graph aggregation
network for text-to-sql. Advances in Neural Information Processing Systems (2021). Vol.34, p7664–
7676.
[19] Binyuan Hui, Xiang Shi, Ruiying Geng, Binhua Li, Yongbin Li, Jian Sun, Xiaodan Zhu: Improving Text-
to-SQL with Schema Dependency Learning. arXiv:2103.04399,2021.
[20] Peter Shaw, Ming-Wei Chang, Panupong Pasupat, and Kristina Toutanova: Compositional
generalization and natural language variation: Can a semantic parsing approach handle both? In
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the
92
Frontiers in Science and Engineering Volume 4 Issue 4, 2024
ISSN: 2710-0588
11th International Joint Conference on Natural Language Processing (Online, August,2021). Vol.1,
p922–938, Online.
[21] Torsten Scholak, Nathan Schucher, and Dzmitry Bahdanau: PICARD: Parsing incrementally for
constrained auto-regressive decoding from language models. Proceedings of the 2021 Conference
on Empirical Methods in Natural Language Processing (Online and Punta Cana, Dominican
Republic.November,2021), p9895–9901.
[22] Bailin Wang, Richard Shin, Xiaodong Liu, Oleksandr Polozov, and Matthew Richardson: RAT-SQL:
Relation-Aware Schema Encoding and Linking for Text-to-SQL Parsers. Proceedings of the 58th
Annual Meeting of the Association for Computational Linguistics (Online, July,2020), p7567-7578.
[23] Peter Shaw, Jakob Uszkoreit, and Ashish Vaswani: Self-Attention with Relative Position
Representations. Proceedings of the 2018 Conference of the North American Chapter of the
Association for Computational Linguistics: Human Language Technologies (New Orleans, Louisiana,
June,2018). Vol.2, p464–468.
[24] Petar Velickovi ˇ c, Guillem Cucurull, Arantxa ´Casanova, Adriana Romero, Pietro Liò, and Yoshua
Bengio: Graph Attention Networks.arXiv:1710.10903,2018.
[25] Yujian Gan, Xinyun Chen, Qiuping Huang, Matthew Purver, John R. Woodward, Jinxia Xie, and
Pengsheng Huang: Towards robustness of text-to-SQL models against synonym substitution.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the
11th International Joint Conference on Natural Language Processing (Online, August,2021).
Vol.1,p2505-2515.
[26] Srinivasan Iyer, Ioannis Konstas, Alvin Cheung, Jayant Krishnamurthy, and Luke Zettlemoyer:
Learning a neural semantic parser from user feedback. Proceedings of the 55th Annual Meeting of
the Association for Computational Linguistics (Vancouver, Canada, July,2017). Vol .1, p963–973.
[27] John M. Zelle and Raymond J. Mooney: Learning to parse database queries using inductive logic
programming. Proceedings of the Thirteenth National Conference on Artificial Intelligence (1996).
Vol. 2, p1050–1055.
[28] Peng Qi, Yuhao Zhang, Yuhui Zhang, Jason Bolton, and Christopher D. Manning: Stanza: A Python
Natural Language Processing Toolkit for Many Human Languages. Proceedings of the 58th Annual
Meeting of the Association for Computational Linguistics: System Demonstrations (Online,
July,2020), p101–108.
[29] Binyuan Hui, Ruiying Geng, Lihan Wang, Bowen Qin, Bowen Li, Jian Sun, and Yongbin Li: S2 SQL:
Injecting Syntax to Question-Schema Interaction Graph Encoder for Text-to-SQL Parsers.arXiv:
2203.06958,2022.
[30] Ruisheng Cao, Lu Chen, Zhi Chen, Yanbin Zhao, Su Zhu, and Kai Yu.: LGESQL: Line Graph Enhanced
Text-to-SQL Model with Mixed Local and Non-Local Relations. ArXiv:2106.01093,2021.
[31] Jiexing Qi, Jingyao Tang, Ziwei He, Xiangpeng Wan, Yu Cheng, Chenghu Zhou, Xinbing Wang, Quanshi
Zhang, and Zhouhan Lin: RASAT: Integrating Relational Structures into Pretrained Seq2Seq Model
for Text-to-SQL. arXiv:2205.06983,2022.
[32] Jinyang Li, Binyuan Hui, Reynold Cheng, Bowen Qin, Chenhao Ma, Nan Huo, Fei Huang, Wenyu Du,
Luo Si, and Yongbin Li: Graphix-t5: Mixing pretrained transformers with graph-aware layers for
text-to-sql parsing. arXiv:2301.07507,2023.
93