[ https://issues.apache.org/jira/browse/FLINK-35560?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
dongwoo.kim updated FLINK-35560: -------------------------------- Description: h3. Summary Hello I'd like to suggest custom pluggable query operation validator support in flink sql gateway. As an sql gateway operator, there is need for query operation validation to only execute safe queries and drop unsafe queries. To address this need, I propose adding a {{OperationValidator}} interface in flink sql gateway api package. This interface will allow users to implement their own query operation validation logic, providing benefits to flink sql gateway operators. h3. Interface Below is a draft for the interface. It takes Operation and check whether the query is valid or not. {code:java} package org.apache.flink.table.gateway.api.validator; import org.apache.flink.annotation.Public; import org.apache.flink.table.operations.Operation; public interface OperationValidator extends Plugin { /** * Validate a given operation and return an optional error string. * * @param op the operation to be validated. * @return Optional error string, should be present only if validation resulted in an error. */ Optional<String> validate(Operation op); } {code} h3. Example implementation Below is an example implementation that inspects Kafka table options, specifically {{{}scan.startup.timestamp-millis{}}}, and reject when the value is too small, which can cause high disk I/O load. {code:java} package org.apache.flink.table.gateway.api.validator; import org.apache.flink.table.gateway.api.validator.OperationValidator; import org.apache.flink.table.operations.Operation; import org.apache.flink.table.operations.ddl.CreateTableOperation; public class KafkaTableValidator implements OperationValidator { private static final long ONE_DAY = 24 * 60 * 60 * 1000L; @Override public Optional<String> validate(Operation op) { if (op instanceof CreateTableOperation) { final CreateTableOperation createTableOp = (CreateTableOperation) op; final String connector = createTableOp.getCatalogTable().getOptions().get("connector"); if ("kafka".equals(connector)) { try { final long startupTimestampValue = Long.parseLong( createTableOp .getCatalogTable() .getOptions() .get("scan.startup.timestamp-millis")); if (startupTimestampValue < System.currentTimeMillis() - ONE_DAY) { return Optional.of( String.format( "Validation failed in %s: 'scan.startup.timestamp-millis' is too old. Given value: %d. It must be within one day from the current time.", KafkaTableValidator.class.getName(), startupTimestampValue)); } } catch (NumberFormatException e) { return Optional.empty(); } } } return Optional.empty(); } }{code} I'm looking forward to get feedback from the community. Thanks was: h3. Summary Hello I'd like to suggest query validator support in flink sql gateway via spi pattern. As an sql gateway operator, there is need for query validation to only execute safe queries and drop unsafe queries. To address this need, I propose adding a {{QueryValidator}} interface in flink sql gateway api package. This interface will allow users to implement their own query validation logic, providing benefits to flink sql gateway operators. h3. Interface Below is a draft for the interface. It takes Operation and check whether the query is valid or not. {code:java} package org.apache.flink.table.gateway.api.validator; import org.apache.flink.annotation.Public; import org.apache.flink.table.operations.Operation; /** * Interface for implementing a validator that checks the safety of executing queries. */ @Public public interface QueryValidator { boolean validateQuery(Operation op); } {code} h3. Example implementation Below is an example implementation that inspects Kafka table options, specifically {{{}scan.startup.timestamp-millis{}}}, and reject when the value is too small, which can cause high disk I/O load. {code:java} package org.apache.flink.table.gateway.api.validator; import org.apache.flink.table.gateway.api.validator.QueryValidator; import org.apache.flink.table.operations.Operation; import org.apache.flink.table.operations.ddl.CreateTableOperation; public class KafkaTimestampValidator implements QueryValidator { private static final long ONE_DAY = 24 * 60 * 60 * 1000L; @Override public boolean validateQuery(Operation op) { if (op instanceof CreateTableOperation) { CreateTableOperation createTableOp = (CreateTableOperation) op; String connector = createTableOp.getCatalogTable().getOptions().get("connector"); if ("kafka".equals(connector)) { String startupTimestamp = createTableOp.getCatalogTable().getOptions().get("scan.startup.timestamp-millis"); if (startupTimestamp != null && Long.parseLong(startupTimestamp) < System.currentTimeMillis() - ONE_DAY) { return false; } } } return true; } }{code} I'd be happy to implement this feature, if we can reach on agreement. Thanks. > Support custom query operation validator for sql gateway > -------------------------------------------------------- > > Key: FLINK-35560 > URL: https://issues.apache.org/jira/browse/FLINK-35560 > Project: Flink > Issue Type: Improvement > Components: Table SQL / Gateway > Reporter: dongwoo.kim > Priority: Major > Labels: pull-request-available > > h3. Summary > Hello I'd like to suggest custom pluggable query operation validator support > in flink sql gateway. > As an sql gateway operator, there is need for query operation validation to > only execute safe queries and drop unsafe queries. > To address this need, I propose adding a {{OperationValidator}} interface in > flink sql gateway api package. > This interface will allow users to implement their own query operation > validation logic, providing benefits to flink sql gateway operators. > h3. Interface > Below is a draft for the interface. > It takes Operation and check whether the query is valid or not. > {code:java} > package org.apache.flink.table.gateway.api.validator; > import org.apache.flink.annotation.Public; > import org.apache.flink.table.operations.Operation; > public interface OperationValidator extends Plugin { > /** > * Validate a given operation and return an optional error string. > * > * @param op the operation to be validated. > * @return Optional error string, should be present only if validation > resulted in an error. > */ > Optional<String> validate(Operation op); > } > {code} > h3. Example implementation > Below is an example implementation that inspects Kafka table options, > specifically {{{}scan.startup.timestamp-millis{}}}, and reject when the value > is too small, which can cause high disk I/O load. > {code:java} > package org.apache.flink.table.gateway.api.validator; > import org.apache.flink.table.gateway.api.validator.OperationValidator; > import org.apache.flink.table.operations.Operation; > import org.apache.flink.table.operations.ddl.CreateTableOperation; > public class KafkaTableValidator implements OperationValidator { > private static final long ONE_DAY = 24 * 60 * 60 * 1000L; > @Override > public Optional<String> validate(Operation op) { > if (op instanceof CreateTableOperation) { > final CreateTableOperation createTableOp = (CreateTableOperation) > op; > final String connector = > createTableOp.getCatalogTable().getOptions().get("connector"); > if ("kafka".equals(connector)) { > try { > final long startupTimestampValue = > Long.parseLong( > createTableOp > .getCatalogTable() > .getOptions() > > .get("scan.startup.timestamp-millis")); > if (startupTimestampValue < System.currentTimeMillis() - > ONE_DAY) { > return Optional.of( > String.format( > "Validation failed in %s: > 'scan.startup.timestamp-millis' is too old. Given value: %d. It must be > within one day from the current time.", > KafkaTableValidator.class.getName(), > startupTimestampValue)); > } > } catch (NumberFormatException e) { > return Optional.empty(); > } > } > } > return Optional.empty(); > } > }{code} > I'm looking forward to get feedback from the community. > Thanks -- This message was sent by Atlassian Jira (v8.20.10#820010)