Skip to content

HKUSTDial/NL2SQL-Bugs-Benchmark

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

10 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

NL2SQL-BUGs: A Benchmark for Detecting Semantic Errors in NL2SQL Translation

πŸ” Overview

NL2SQL-BUGs is the benchmark dedicated to detecting and categorizing semantic errors in Natural Language to SQL (NL2SQL) translation. While state-of-the-art NL2SQL models have made significant progress in translating natural language queries to SQL, they still frequently generate semantically incorrect queries that may execute successfully but produce incorrect results. This benchmark aims to support research in semantic error detection, which is a prerequisite for any subsequent error correction.

  • Two-Level Taxonomy: A comprehensive classification system for semantic errors with 9 main categories and 31 subcategories
  • Expert-Annotated Dataset: 2,018 instances including 1,019 correct examples and 999 semantically incorrect examples
  • Detailed Error Annotations: Each incorrect example is meticulously annotated with specific error types

πŸ“Š Data Statistics

Our taxonomy classifies semantic errors in NL2SQL translation into 9 main categories and 31 subcategories:

  • Total instances: 2,018
  • Correct examples: 1,019
  • Incorrect examples (with semantic errors): 999
  • Main error categories: 9
  • Error subcategories: 31

πŸš€ Dataset

Our database is consistent with BIRD benchmark. You can download the database from either:

You can find NL2SQL-BUGs label and error type in ./data/ dir

πŸ“ˆ Evaluation

We evaluate semantic error detection performance using the following metrics:

  • Overall Accuracy: Percentage of correctly identified instances (correct or incorrect)
  • Negative Precision (NP): Proportion of correctly predicted incorrect cases out of all predicted incorrect cases
  • Negative Recall (NR): Proportion of correctly predicted incorrect cases out of all actual incorrect cases
  • Positive Precision (PP): Proportion of correctly predicted correct cases out of all predicted correct cases
  • Positive Recall (PR): Proportion of correctly predicted correct cases out of all actual correct cases
  • Type-Specific Accuracy (TSA): Accuracy for each specific error type

πŸ“ Citation

If you use NL2SQL-BUGs in your research, please cite our paper:

@inproceedings{10.1145/3711896.3737427,
author = {Liu, Xinyu and Shen, Shuyu and Li, Boyan and Tang, Nan and Luo, Yuyu},
title = {NL2SQL-BUGs: A Benchmark for Detecting Semantic Errors in NL2SQL Translation},
year = {2025},
isbn = {9798400714542},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3711896.3737427},
doi = {10.1145/3711896.3737427},
booktitle = {Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V.2},
pages = {5662–5673},
numpages = {12},
keywords = {interface for databases, large language model, text-to-SQL},
location = {Toronto ON, Canada},
series = {KDD '25}
}

πŸ“¬ Contact

If you have any questions or need further support, please feel free to reach out: xliu371[at]connect.hkust-gz.edu.cn.

About

πŸ”₯[SIGKDD'25] NL2SQL-BUGs: A Benchmark for Detecting Semantic Errors in NL2SQL Translation.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •