Automated NLIDB Testing Solutions
Automated NLIDB Testing Solutions
■ NLIDB: Natural language interface to database takes input a natural language question and outputs a structured query e.g. SQL
e.g. PRECISE, NALIR, ATHENA
– user does not need to know SQL and exact schema
■ Problem with current testing/evaluation of NLIDB systems
– Testing of such NLIDB systems have been performed in ad-hoc fashion. This hampers the production usage of NLIDB
systems
– The benchmarks used contain a set of natural language questions and their gold standard SQL. They lack in
■ Natural language variation
■ Query syntax and semantics coverage
■ They do not cater towards exploiting the choices an NLIDB system takes during the translation
■ Goal of the internship
– To create automated test case (NLQ, OQL pairs) generation for NLIDB systems for a domain given ontology and other
linguistic resources
– To define metrics of good test case for NLIDB systems
– To experimentally demonstrate that existing benchmarks lack in these metrics
– To create static test suite (with limitation of number of test cases)
– To create dynamic test suite (the test question is dependent on the response of previous question)
– Evaluation of existing NLIDB systems PRECISE, NALIR, ATHENA using both static and dynamic test suite generation
technique