refactor: standard naming (#278)

* refactor: rename flinksql to flink

* refactor: rename pgsql to postgresql

* refactor: rename trinosql to trino

* refactor: replace all default exports with named export

* refactor: rename basicParser to basicSQL

* refactor: rename basic-parser-types to types

* refactor: replace arrow func with plain func
This commit is contained in:
Hayden
2024-03-27 10:33:25 +08:00
committed by GitHub
parent a99721162b
commit bb0fad1dbe
325 changed files with 33161 additions and 33202 deletions

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
add: readSQL(__dirname, 'add.sql'),
@ -10,7 +10,7 @@ const features = {
describe('Spark Add Syntax Tests', () => {
features.add.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
alterDatabase: readSQL(__dirname, 'alterDatabase.sql'),
@ -14,7 +14,7 @@ describe('SparkSQL Alter Syntax Tests', () => {
Object.keys(features).forEach((key) => {
features[key].forEach((sql) => {
it(sql, () => {
expect(parser.validate(sql).length).toBe(0);
expect(spark.validate(sql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
analyzeTable: readSQL(__dirname, 'analyzeTable.sql'),
@ -10,7 +10,7 @@ const features = {
describe('Spark Analyze Table Syntax Tests', () => {
features.analyzeTable.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
cache: readSQL(__dirname, 'cache.sql'),
@ -10,7 +10,7 @@ const features = {
describe('Spark Cache Syntax Tests', () => {
features.cache.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
createDatabase: readSQL(__dirname, 'createDatabase.sql'),
@ -14,7 +14,7 @@ describe('SparkSQL Create Syntax Tests', () => {
Object.keys(features).forEach((key) => {
features[key].forEach((sql) => {
it(sql, () => {
expect(parser.validate(sql).length).toBe(0);
expect(spark.validate(sql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
describe: readSQL(__dirname, 'describe.sql'),
@ -10,7 +10,7 @@ const features = {
describe('Spark Describe Syntax Tests', () => {
features.describe.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
dropDatabase: readSQL(__dirname, 'dropDatabase.sql'),
@ -15,7 +15,7 @@ describe('SparkSQL Drop Syntax Tests', () => {
Object.keys(features).forEach((key) => {
features[key].forEach((sql) => {
it(sql, () => {
expect(parser.validate(sql).length).toBe(0);
expect(spark.validate(sql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
insertIntoTable: readSQL(__dirname, 'insertIntoTable.sql'),
@ -14,7 +14,7 @@ describe('SparkSQL Insert Syntax Tests', () => {
Object.keys(features).forEach((key) => {
features[key].forEach((sql) => {
it(sql, () => {
expect(parser.validate(sql).length).toBe(0);
expect(spark.validate(sql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
/**
* 关键词有多个值
@ -16,7 +16,7 @@ describe('SparkSQL Keyword Has Multiple Values Syntax Tests', () => {
Object.keys(features).forEach((key) => {
features[key].forEach((sql) => {
it(sql, () => {
expect(parser.validate(sql).length).toBe(0);
expect(spark.validate(sql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
list: readSQL(__dirname, 'list.sql'),
@ -10,7 +10,7 @@ const features = {
describe('Spark List Syntax Tests', () => {
features.list.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
loadData: readSQL(__dirname, 'loadData.sql'),
@ -11,7 +11,7 @@ describe('SparkSQL Load Syntax Tests', () => {
Object.keys(features).forEach((key) => {
features[key].forEach((sql) => {
it(sql, () => {
expect(parser.validate(sql).length).toBe(0);
expect(spark.validate(sql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
set: readSQL(__dirname, 'optimize.sql'),
@ -10,7 +10,7 @@ const features = {
describe('Spark Optimize Syntax Tests', () => {
features.set.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
refresh: readSQL(__dirname, 'refresh.sql'),
@ -10,7 +10,7 @@ const features = {
describe('Spark Refresh Syntax Tests', () => {
features.refresh.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
reset: readSQL(__dirname, 'reset.sql'),
@ -10,7 +10,7 @@ const features = {
describe('Spark Reset Syntax Tests', () => {
features.reset.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
selectAggregateFn: readSQL(__dirname, 'selectAggregateFn.sql'),
@ -34,132 +34,132 @@ const features = {
describe('Spark Select Syntax Tests', () => {
features.selectAggregateFn.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectCase.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectCET.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectWindowFn.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectWhere.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectUnPivot.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectTVF.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectTransform.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectTableSample.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectSortBy.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectPivot.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectOrderBy.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectOffset.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectLimit.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectLike.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectLateralView.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectLateralSubQuery.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectJoin.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectInlineTable.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectHiving.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectHint.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectGroupBy.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectFile.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectExplain.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectDistributeBy.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
features.selectClusterBy.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
set: readSQL(__dirname, 'set.sql'),
@ -10,7 +10,7 @@ const features = {
describe('Spark Set Syntax Tests', () => {
features.set.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
show: readSQL(__dirname, 'show.sql'),
@ -10,7 +10,7 @@ const features = {
describe('Spark Show Syntax Tests', () => {
features.show.forEach((itemSql) => {
it(itemSql, () => {
expect(parser.validate(itemSql).length).toBe(0);
expect(spark.validate(itemSql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
createDataSourceTable: readSQL(__dirname, 'createDataSourceTable.sql'),
@ -16,7 +16,7 @@ describe('SparkSQL About Table Syntax Tests', () => {
Object.keys(features).forEach((key) => {
features[key].forEach((sql) => {
it(sql, () => {
expect(parser.validate(sql).length).toBe(0);
expect(spark.validate(sql).length).toBe(0);
});
});
});

View File

@ -1,7 +1,7 @@
import SparkSQL from 'src/parser/spark';
import { SparkSQL } from 'src/parser/spark';
import { readSQL } from 'test/helper';
const parser = new SparkSQL();
const spark = new SparkSQL();
const features = {
useDatabase: readSQL(__dirname, 'useDatabase.sql'),
@ -11,7 +11,7 @@ describe('SparkSQL Use Database Syntax Tests', () => {
Object.keys(features).forEach((key) => {
features[key].forEach((sql) => {
it(sql, () => {
expect(parser.validate(sql).length).toBe(0);
expect(spark.validate(sql).length).toBe(0);
});
});
});