wuguihu created FLINK-23859: ------------------------------- Summary: [typo][flink-core][flink-connectors]fix typo for code Key: FLINK-23859 URL: https://issues.apache.org/jira/browse/FLINK-23859 Project: Flink Issue Type: Bug Reporter: wuguihu
There are some typo issues in these modules. {code:java} # Use the Codespell tool to check typo issue. pip install codespell codespell -h {code} 1、 codespell flink-java/src {code:java} flink-java/src/main/java/org/apache/flink/api/java/operators/PartitionOperator.java:125: partioning ==> partitioning flink-java/src/main/java/org/apache/flink/api/java/operators/PartitionOperator.java:128: neccessary ==> necessary {code} 2、 codespell flink-clients/ {code:java} flink-clients/src/test/java/org/apache/flink/client/program/DefaultPackagedProgramRetrieverTest.java:545: acessible ==> accessible {code} 3、codespell flink-connectors/ -S '*.xml' -S '*.iml' -S '*.txt' {code:java} flink-connectors/flink-connector-base/src/main/java/org/apache/flink/connector/base/source/reader/SourceReaderOptions.java:25: tht ==> that flink-connectors/flink-connector-jdbc/src/test/java/org/apache/flink/connector/jdbc/catalog/PostgresCatalogTestBase.java:192: doens't ==> doesn't flink-connectors/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/dialect/JdbcDialect.java:96: PostgresSQL ==> postgresql flink-connectors/flink-connector-cassandra/src/test/java/org/apache/flink/batch/connectors/cassandra/CustomCassandraAnnotatedPojo.java:38: instanciation ==> instantiation flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/planner/delegation/hive/HiveParserCalcitePlanner.java:822: partion ==> partition flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/planner/delegation/hive/HiveParserTypeCheckProcFactory.java:943: funtion ==> function flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/planner/delegation/hive/copy/HiveASTParseDriver.java:55: funtion ==> function flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/planner/delegation/hive/copy/HiveASTParseDriver.java:51: characteres ==> characters flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/util/KinesisConfigUtil.java:436: paremeters ==> parameters flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/planner/delegation/hive/copy/HiveParserBaseSemanticAnalyzer.java:2369: Unkown ==> Unknown flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/config/ConsumerConfigConstants.java:75: reprsents ==> represents flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/functions/hive/HiveFunctionWrapper.java:28: functino ==> function flink-connectors/flink-connector-hbase-2.2/src/main/java/org/apache/flink/connector/hbase2/source/HBaseRowDataAsyncLookupFunction.java:62: implemenation flink-connectors/flink-connector-pulsar/src/main/java/org/apache/flink/connector/pulsar/source/enumerator/cursor/StartCursor.java:70: ture ==> true flink-connectors/flink-connector-pulsar/src/test/resources/containers/txnStandalone.conf:907: partions ==> partitions flink-connectors/flink-connector-pulsar/src/test/resources/containers/txnStandalone.conf:468: implementatation ==> implementation flink-connectors/flink-connector-files/src/main/java/org/apache/flink/connector/file/src/enumerate/BlockSplittingRecursiveEnumerator.java:141: bloc ==> block flink-connectors/flink-connector-files/src/main/java/org/apache/flink/connector/file/src/reader/SimpleStreamFormat.java:37: te ==> the flink-connectors/flink-connector-kafka/src/main/java/org/apache/flink/connector/kafka/source/reader/deserializer/KafkaRecordDeserializationSchema.java:70: determin ==> determine flink-connectors/flink-connector-kafka/src/main/java/org/apache/flink/connector/kafka/source/enumerator/subscriber/TopicListSubscriber.java:36: hav ==> have flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/planner/delegation/hive/copy/HiveParserQBSubQuery.java:555: correlatd ==> correlated flink-connectors/flink-connector-kafka/src/test/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaConsumerBaseTest.java:263: intial ==> initial flink-connectors/flink-connector-kafka/src/test/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaConsumerBaseTest.java:302: intial ==> initial flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/internals/KinesisDataFetcher.java:249: wth ==> with flink-connectors/flink-connector-files/src/test/java/org/apache/flink/connector/file/sink/writer/FileWriterBucketStateSerializerMigrationTest.java:232: comitted ==> committed flink-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQDeserializationSchema.java:96: invokation ==> invocation flink-connectors/flink-connector-kafka/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaConsumerTestBase.java:171: doesnt ==> doesn't {code} 4、codespell flink-core/ -S '*.xml' -S '*.iml' -S '*.txt' {code:bash} flink-core/src/main/java/org/apache/flink/api/common/typeutils/TypeSerializerConfigSnapshot.java:194: preform ==> perform flink-core/src/main/java/org/apache/flink/configuration/CheckpointingOptions.java:150: aways flink-core/src/main/java/org/apache/flink/configuration/ConfigOption.java:164: documention ==> documentation flink-core/src/main/java/org/apache/flink/configuration/ConfigOption.java:175: documention ==> documentation flink-core/src/main/java/org/apache/flink/core/fs/RecoverableFsDataOutputStream.java:54: retured ==> returned flink-core/src/test/java/org/apache/flink/api/common/io/DelimitedInputFormatTest.java:352: skipp ==> skip flink-core/src/main/java/org/apache/flink/util/InstantiationUtil.java:216: occurences ==> occurrences {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)