文字问题转sql语句。

waynesun f01fa37d85 update 5 years ago
data 9a17477635 update 5 years ago
img 61a61e330e update readme 5 years ago
saved_model 9a9c2a4c7f init 5 years ago
sqlnet f01fa37d85 update 5 years ago
LICENSE 9a9c2a4c7f init 5 years ago
README.md f01fa37d85 update 5 years ago
start_test.sh 9a9c2a4c7f init 5 years ago
start_train.sh 9a9c2a4c7f init 5 years ago
test.py 9a9c2a4c7f init 5 years ago
train.py 9a9c2a4c7f init 5 years ago

README.md

Introduction

This baseline method is developed and refined based on code of SQLNet, which is a baseline model in WikiSQL.

The model decouples the task of generating a whole SQL into several sub-tasks, including select-number, select-column, select-aggregation, condition-number, condition-column and so on.

Simple model structure shows here, implementation details could refer to the origin paper.

The difference between SQLNet and this baseline model is, Select-Number and Where-Relationship sub-tasks are added to adapt this Chinese NL2SQL dataset better.

Dependencies

  • Python 2.7
  • torch 1.0.1
  • tqdm

Start to train

Firstly, download the provided datasets at ~/data_nl2sql/, which should include train.json, train.tables.json, dev.json, dev.tables.json and char_embedding, and divide them in following structure.

├── data
│ ├── train
│ │ ├── train.json
│ │ ├── train.tables.json
│ ├── dev
│ │ ├── dev.json
│ │ ├── dev.tables.json
├── char_embedding

and then

mkdir ~/nl2sql
cd ~/nl2sql/
git clone https://github.com/ZhuiyiTechnology/nl2sql_baseline.git

cp ~/data_nl2sql/* ~/nl2sql/nl2sql_baseline/data
cd ~/nl2sql/nl2sql_baseline/

sh ./start_train.py 0 128

while the first parameter 0 means gpu number, the second parameter means batch size.

Start to evaluate

To evaluate on dev.json or test.json, make sure trained model is ready, then run

cd ~/nl2sql/nl2sql_baseline/
sh ./start_test.py 0 pred_example

while the first parameter 0 means gpu number, the second parameter means the output path of prediction.

Experiment result

We have run experiments several times, achiving avegrage 27.5% logic form accuracy on the dev dataset.

Experiment analysis

We found the main challenges of this datasets containing poor condition value prediction, select column and condition column not mentioned in NL question, inconsistent condition relationship representation between NL question and SQL, etc. All these challenges could not be solved by existing baseline and SOTA models.

Correspondingly, this baseline model achieves only 77% accuracy on condition column and 62% accuracy on condition value respectively even on the training set, and the overall logic form is only around 50% as well, indicating these problems are challenging for contestants to solve.

Related resources:

https://github.com/salesforce/WikiSQL

https://yale-lily.github.io/spider

Semantic Parsing with Syntax- and Table-Aware SQL Generation