lava-oushudb-dt-sql-parser/test/parser/spark/contextCollect/entityCollector.test.ts

363 lines
14 KiB
TypeScript
Raw Normal View History

feat: collect entity (#265) * feat: add text and word utils * feat: add entity collector class * refactor: rename SyntaxContextType to EntityContextType * refactor: improve EntityCollector * feat: improve mysql parser grammar * feat: add mysql entity collector * test: mysql entity collector tests * feat: remove useless method * feat: improve spark grammar file * feat: add spark entity collector * test: spark entity collector unit tests * feat: remove useless code * feat: add queryStatement label * feat: add crateDatabaseStmt * feat: add trino entity collector * feat: rename trinosql to trino * test: trino collect entity unit tests * test: fix spark test * feat(impala): support impale entity collector (#256) * Feat/collect entity hive (#263) * feat(hive): support hive collect entity * feat(hive): update tableAllColumns * feat: replace antlr4ts with antlr4ng * feat(pgsql): pgsql collect entity (#268) * feat(pgsql): pgsql collect entity * feat(pgsql): optimize some name --------- Co-authored-by: zhaoge <> * feat: get word text by token.text * feat: supprt collect db/function and add splitListener (#270) * feat: supprt collect db/function and add splitListner * feat: remove SplitListener interface in baseParser to use SplitListener in root * fix(mysql): fix show create xxx not celloct as createXXXEntity type * test: fix pgsql unit tests * Feat/error recover predicate (#274) * feat: optimize pgsql grammar * feat: add sql parser base * feat: apply SQLParserBase * feat: add geAllEntities method * test: test collect table when missing column * feat: compose collect and suggestion (#276) * feat: mark stmt which contain caret * test: correct name of getAllEntities * test: remove misscolumn unit tests * test: add suggestionWithEntity tests * feat: flink collect entity (#277) * feat: improve flink sql parser * feat: support flink entity collector * test: flink entity collect unit test * feat: move combine entities to parent class --------- Co-authored-by: 霜序 <976060700@qq.com> Co-authored-by: XCynthia <942884029@qq.com>
2024-03-26 14:28:27 +08:00
import fs from 'fs';
import path from 'path';
import SparkSQL from 'src/parser/spark';
import { SparkSqlSplitListener, SparkEntityCollector } from 'src/parser/spark';
import { ParseTreeListener } from 'antlr4ng';
import { SparkSqlParserListener } from 'src/lib/spark/SparkSqlParserListener';
import { EntityContextType } from 'src/parser/common/basic-parser-types';
import { StmtContextType } from 'src/parser/common/entityCollector';
const commonSql = fs.readFileSync(path.join(__dirname, 'fixtures', 'common.sql'), 'utf-8');
describe('SparkSQL entity collector tests', () => {
const spark = new SparkSQL();
const parseTree = spark.parse(commonSql);
const splitListener = new SparkSqlSplitListener();
spark.listen(splitListener as SparkSqlParserListener, parseTree);
test('validate common sql', () => {
expect(spark.validate(commonSql).length).toBe(0);
});
test('split results', () => {
expect(splitListener.statementsContext.length).toBe(12);
});
test('create table like', () => {
const columnCreateTableContext = splitListener.statementsContext[0];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, columnCreateTableContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(2);
const tableCreateEntity = allEntities[0];
expect(tableCreateEntity.entityContextType).toBe(EntityContextType.TABLE_CREATE);
expect(tableCreateEntity.text).toBe('new_tb1');
expect(tableCreateEntity.position).toEqual({
startIndex: 27,
endIndex: 33,
line: 1,
startColumn: 28,
endColumn: 35,
});
expect(tableCreateEntity.belongStmt.stmtContextType).toBe(
StmtContextType.CREATE_TABLE_STMT
);
expect(tableCreateEntity.belongStmt.position).toEqual({
startIndex: 0,
endIndex: 50,
startLine: 1,
endLine: 1,
startColumn: 1,
endColumn: 52,
});
expect(tableCreateEntity.relatedEntities.length).toBe(1);
const beLikedEntity = allEntities[1];
expect(tableCreateEntity.relatedEntities[0]).toBe(beLikedEntity);
expect(beLikedEntity.text).toBe('like_old_tb');
expect(beLikedEntity.entityContextType).toBe(EntityContextType.TABLE);
expect(beLikedEntity.belongStmt).toBe(tableCreateEntity.belongStmt);
});
test('create hive format table', () => {
const columnCreateTableContext = splitListener.statementsContext[1];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, columnCreateTableContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(1);
const tableCreateEntity = allEntities[0];
expect(tableCreateEntity.entityContextType).toBe(EntityContextType.TABLE_CREATE);
expect(tableCreateEntity.text).toBe('new_tb2');
expect(tableCreateEntity.position).toEqual({
startIndex: 67,
endIndex: 73,
line: 3,
startColumn: 14,
endColumn: 21,
});
expect(tableCreateEntity.belongStmt.stmtContextType).toBe(
StmtContextType.CREATE_TABLE_STMT
);
expect(tableCreateEntity.belongStmt.position).toEqual({
startIndex: 54,
endIndex: 242,
startLine: 3,
endLine: 8,
startColumn: 1,
endColumn: 22,
});
expect(tableCreateEntity.relatedEntities).toBeNull();
expect(tableCreateEntity.columns.length).toBe(2);
tableCreateEntity.columns.forEach((columEntity) => {
expect(columEntity.entityContextType).toBe(EntityContextType.COLUMN_CREATE);
expect(columEntity.belongStmt).toBe(tableCreateEntity.belongStmt);
expect(columEntity.text).toBe(
commonSql.slice(columEntity.position.startIndex, columEntity.position.endIndex + 1)
);
});
});
test('create data source table', () => {
const testingContext = splitListener.statementsContext[2];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, testingContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(2);
const tableCreateEntity = allEntities[0];
const originTableEntity = allEntities[1];
expect(tableCreateEntity.entityContextType).toBe(EntityContextType.TABLE_CREATE);
expect(tableCreateEntity.text).toBe('student_copy');
expect(tableCreateEntity.belongStmt.stmtContextType).toBe(
StmtContextType.CREATE_TABLE_STMT
);
expect(tableCreateEntity.columns).toBeNull();
expect(tableCreateEntity.relatedEntities.length).toBe(1);
expect(tableCreateEntity.relatedEntities[0]).toBe(originTableEntity);
expect(originTableEntity.entityContextType).toBe(EntityContextType.TABLE);
expect(originTableEntity.text).toBe('student');
expect(originTableEntity.belongStmt.rootStmt).toBe(tableCreateEntity.belongStmt);
});
test('create view', () => {
const testingContext = splitListener.statementsContext[3];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, testingContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(2);
const viewEntity = allEntities[0];
const tableEntity = allEntities[1];
expect(viewEntity.entityContextType).toBe(EntityContextType.VIEW_CREATE);
expect(viewEntity.belongStmt.stmtContextType).toBe(StmtContextType.CREATE_VIEW_STMT);
expect(viewEntity.text).toBe('new_view1');
expect(viewEntity.columns.length).toBe(2);
viewEntity.columns.forEach((columEntity) => {
expect(columEntity.entityContextType).toBe(EntityContextType.COLUMN_CREATE);
expect(columEntity.belongStmt).toBe(viewEntity.belongStmt);
expect(columEntity.text).toBe(
commonSql.slice(columEntity.position.startIndex, columEntity.position.endIndex + 1)
);
});
expect(tableEntity.entityContextType).toBe(EntityContextType.TABLE);
expect(tableEntity.belongStmt.stmtContextType).toBe(StmtContextType.SELECT_STMT);
expect(tableEntity.belongStmt.rootStmt).toBe(viewEntity.belongStmt);
expect(tableEntity.text).toBe('old_tb_1');
});
test('select from table', () => {
const testingContext = splitListener.statementsContext[4];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, testingContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(2);
const tableEntity1 = allEntities[0];
const tableEntity2 = allEntities[1];
expect(tableEntity1.entityContextType).toBe(EntityContextType.TABLE);
expect(tableEntity1.belongStmt.stmtContextType).toBe(StmtContextType.SELECT_STMT);
expect(tableEntity1.text).toBe('employee');
expect(tableEntity2.entityContextType).toBe(EntityContextType.TABLE);
expect(tableEntity2.belongStmt.stmtContextType).toBe(StmtContextType.SELECT_STMT);
expect(tableEntity2.text).toBe('department');
});
test('insert into table values', () => {
const testingContext = splitListener.statementsContext[5];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, testingContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(1);
const tableEntity = allEntities[0];
expect(tableEntity.entityContextType).toBe(EntityContextType.TABLE);
expect(tableEntity.belongStmt.stmtContextType).toBe(StmtContextType.INSERT_STMT);
expect(tableEntity.text).toBe('insert_tb');
});
test('insert overwrite table', () => {
const testingContext = splitListener.statementsContext[6];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, testingContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(2);
const targetTableEntity = allEntities[0];
const sourceTableEntity = allEntities[1];
expect(targetTableEntity.entityContextType).toBe(EntityContextType.TABLE);
expect(targetTableEntity.belongStmt.stmtContextType).toBe(StmtContextType.INSERT_STMT);
expect(targetTableEntity.text).toBe('target_tb');
expect(sourceTableEntity.entityContextType).toBe(EntityContextType.TABLE);
expect(sourceTableEntity.belongStmt.stmtContextType).toBe(StmtContextType.SELECT_STMT);
expect(sourceTableEntity.belongStmt.rootStmt).toBe(targetTableEntity.belongStmt);
expect(sourceTableEntity.text).toBe('source_tb');
});
test('insert overwrite dir', () => {
const testingContext = splitListener.statementsContext[7];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, testingContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(1);
const sourceTableEntity = allEntities[0];
expect(sourceTableEntity.entityContextType).toBe(EntityContextType.TABLE);
expect(sourceTableEntity.belongStmt.stmtContextType).toBe(StmtContextType.SELECT_STMT);
expect(sourceTableEntity.text).toBe('from_tb');
});
test('create database', () => {
const testingContext = splitListener.statementsContext[8];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, testingContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(1);
const sourceTableEntity = allEntities[0];
expect(sourceTableEntity.entityContextType).toBe(EntityContextType.DATABASE_CREATE);
expect(sourceTableEntity.belongStmt.stmtContextType).toBe(
StmtContextType.CREATE_DATABASE_STMT
);
expect(sourceTableEntity.text).toBe('customer_db');
});
test('use namespace', () => {
const testingContext = splitListener.statementsContext[9];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, testingContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(1);
const sourceTableEntity = allEntities[0];
expect(sourceTableEntity.entityContextType).toBe(EntityContextType.DATABASE);
expect(sourceTableEntity.belongStmt.stmtContextType).toBe(StmtContextType.COMMON_STMT);
expect(sourceTableEntity.text).toBe('ns1');
});
test('create function', () => {
const functionContext = splitListener.statementsContext[10];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, functionContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(1);
const functionEntity = allEntities[0];
expect(functionEntity.entityContextType).toBe(EntityContextType.FUNCTION_CREATE);
expect(functionEntity.text).toBe('simple_udf');
expect(functionEntity.position).toEqual({
endColumn: 38,
endIndex: 905,
line: 28,
startColumn: 28,
startIndex: 896,
});
expect(functionEntity.belongStmt.stmtContextType).toBe(
StmtContextType.CREATE_FUNCTION_STMT
);
expect(functionEntity.belongStmt.position).toEqual({
endColumn: 54,
endIndex: 921,
endLine: 28,
startColumn: 1,
startIndex: 869,
startLine: 28,
});
expect(functionEntity.columns).toBeNull();
expect(functionEntity.relatedEntities).toBeNull();
});
test('create xxx function', () => {
const functionContext = splitListener.statementsContext[11];
const collectListener = new SparkEntityCollector(commonSql);
spark.listen(collectListener as ParseTreeListener, functionContext);
const allEntities = collectListener.getEntities();
expect(allEntities.length).toBe(1);
const functionEntity = allEntities[0];
expect(functionEntity.entityContextType).toBe(EntityContextType.FUNCTION_CREATE);
expect(functionEntity.text).toBe('simple_udf');
expect(functionEntity.position).toEqual({
endColumn: 27,
endIndex: 950,
line: 30,
startColumn: 17,
startIndex: 941,
});
expect(functionEntity.belongStmt.stmtContextType).toBe(
StmtContextType.CREATE_FUNCTION_STMT
);
expect(functionEntity.belongStmt.position).toEqual({
endColumn: 43,
endIndex: 966,
endLine: 30,
startColumn: 1,
startIndex: 925,
startLine: 30,
});
expect(functionEntity.columns).toBeNull();
expect(functionEntity.relatedEntities).toBeNull();
});
});