This is an automated email from the ASF dual-hosted git repository.
orpiske pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/camel.git
The following commit(s) were added to refs/heads/main by this push:
new 642e185ad5b CAMEL-20161: fix usages of restricted identifiers (#12213)
642e185ad5b is described below
commit 642e185ad5ba8474d2c54a81c18622836b1a73bd
Author: Otavio Rodolfo Piske <[email protected]>
AuthorDate: Mon Nov 27 15:18:40 2023 +0100
CAMEL-20161: fix usages of restricted identifiers (#12213)
---
.../camel/dataformat/asn1/ASN1DataFormat.java | 10 ++---
.../aws2/ddbstream/Ddb2StreamConsumer.java | 4 +-
.../aws2/firehose/KinesisFirehose2Producer.java | 6 +--
.../timestream/write/Timestream2WriteProducer.java | 4 +-
.../component/azure/cosmosdb/CosmosDbConsumer.java | 8 ++--
.../camel/dataformat/bindy/BindyCsvFactory.java | 40 ++++++++++----------
.../dataformat/bindy/BindyFixedLengthFactory.java | 44 +++++++++++-----------
.../braintree/internal/BraintreeLogHandler.java | 32 ++++++++--------
.../camel/dataformat/csv/CsvRecordConverter.java | 4 +-
.../camel/dataformat/csv/CsvRecordConverters.java | 16 ++++----
.../camel/component/debezium/DebeziumEndpoint.java | 10 ++---
.../component/dns/cloud/DnsServiceDiscovery.java | 12 +++---
.../camel/component/dns/types/DnsConverter.java | 4 +-
.../component/flatpack/FlatpackConverter.java | 24 ++++++------
.../component/influxdb2/InfluxDb2Producer.java | 6 +--
.../camel/component/influxdb2/data/Record.java | 12 +++---
.../camel/component/influxdb2/data/Records.java | 16 ++++----
.../kafka/consumer/AbstractCommitManager.java | 4 +-
.../kafka/consumer/AsyncCommitManager.java | 4 +-
.../component/kafka/consumer/CommitManager.java | 2 +-
.../consumer/support/KafkaRecordProcessor.java | 8 ++--
.../support/KafkaRecordProcessorFacade.java | 20 +++++-----
.../support/KafkaProducerMetadataCallBack.java | 8 ++--
.../kafka/SingleNodeKafkaResumeStrategy.java | 14 +++----
.../cloud/KubernetesDnsSrvServiceDiscovery.java | 4 +-
.../apache/camel/openapi/RestOpenApiSupport.java | 8 ++--
.../rest/openapi/RestOpenApiEndpoint.java | 8 ++--
.../component/salesforce/PubSubApiConsumer.java | 4 +-
.../internal/client/PubSubApiClient.java | 12 +++---
.../org/apache/camel/impl/engine/RouteService.java | 8 ++--
.../camel/main/download/DownloadThreadPool.java | 8 ++--
31 files changed, 182 insertions(+), 182 deletions(-)
diff --git
a/components/camel-asn1/src/main/java/org/apache/camel/dataformat/asn1/ASN1DataFormat.java
b/components/camel-asn1/src/main/java/org/apache/camel/dataformat/asn1/ASN1DataFormat.java
index 88d8d1abfdf..3d8630f2944 100644
---
a/components/camel-asn1/src/main/java/org/apache/camel/dataformat/asn1/ASN1DataFormat.java
+++
b/components/camel-asn1/src/main/java/org/apache/camel/dataformat/asn1/ASN1DataFormat.java
@@ -62,12 +62,12 @@ public class ASN1DataFormat extends ServiceSupport
implements DataFormat, DataFo
encodeGenericTypeObject(exchange, stream);
return;
}
- Object record = exchange.getIn().getBody();
- if (record instanceof ASN1Primitive) {
- ASN1Primitive asn1Primitive =
ObjectHelper.cast(ASN1Primitive.class, record);
+ Object body = exchange.getIn().getBody();
+ if (body instanceof ASN1Primitive) {
+ ASN1Primitive asn1Primitive =
ObjectHelper.cast(ASN1Primitive.class, body);
berOut = new ByteArrayInputStream(asn1Primitive.getEncoded());
- } else if (record instanceof byte[]) {
- berOut = new
ByteArrayInputStream(ObjectHelper.cast(byte[].class, record));
+ } else if (body instanceof byte[]) {
+ berOut = new
ByteArrayInputStream(ObjectHelper.cast(byte[].class, body));
}
} else {
byte[] byteInput =
exchange.getContext().getTypeConverter().mandatoryConvertTo(byte[].class,
exchange, graph);
diff --git
a/components/camel-aws/camel-aws2-ddb/src/main/java/org/apache/camel/component/aws2/ddbstream/Ddb2StreamConsumer.java
b/components/camel-aws/camel-aws2-ddb/src/main/java/org/apache/camel/component/aws2/ddbstream/Ddb2StreamConsumer.java
index a51741329c7..62491771907 100644
---
a/components/camel-aws/camel-aws2-ddb/src/main/java/org/apache/camel/component/aws2/ddbstream/Ddb2StreamConsumer.java
+++
b/components/camel-aws/camel-aws2-ddb/src/main/java/org/apache/camel/component/aws2/ddbstream/Ddb2StreamConsumer.java
@@ -79,8 +79,8 @@ public class Ddb2StreamConsumer extends
ScheduledBatchPollingConsumer {
}
List<Record> records = result.records();
Queue<Exchange> exchanges = new ArrayDeque<>();
- for (Record record : records) {
- exchanges.add(createExchange(record));
+ for (Record polledRecord : records) {
+ exchanges.add(createExchange(polledRecord));
}
processedExchangeCount += processBatch(CastUtils.cast(exchanges));
diff --git
a/components/camel-aws/camel-aws2-kinesis/src/main/java/org/apache/camel/component/aws2/firehose/KinesisFirehose2Producer.java
b/components/camel-aws/camel-aws2-kinesis/src/main/java/org/apache/camel/component/aws2/firehose/KinesisFirehose2Producer.java
index 0e8f4635dee..04a55dd7bee 100644
---
a/components/camel-aws/camel-aws2-kinesis/src/main/java/org/apache/camel/component/aws2/firehose/KinesisFirehose2Producer.java
+++
b/components/camel-aws/camel-aws2-kinesis/src/main/java/org/apache/camel/component/aws2/firehose/KinesisFirehose2Producer.java
@@ -178,12 +178,12 @@ public class KinesisFirehose2Producer extends
DefaultProducer {
private PutRecordRequest createRequest(Exchange exchange) {
ByteBuffer body = exchange.getIn().getBody(ByteBuffer.class);
- Record.Builder record = Record.builder();
- record.data(SdkBytes.fromByteBuffer(body));
+ Record.Builder builder = Record.builder();
+ builder.data(SdkBytes.fromByteBuffer(body));
PutRecordRequest.Builder putRecordRequest = PutRecordRequest.builder();
putRecordRequest.deliveryStreamName(getEndpoint().getConfiguration().getStreamName());
- putRecordRequest.record(record.build());
+ putRecordRequest.record(builder.build());
return putRecordRequest.build();
}
diff --git
a/components/camel-aws/camel-aws2-timestream/src/main/java/org/apache/camel/component/aws2/timestream/write/Timestream2WriteProducer.java
b/components/camel-aws/camel-aws2-timestream/src/main/java/org/apache/camel/component/aws2/timestream/write/Timestream2WriteProducer.java
index 18c346ab71a..4f0bb640917 100644
---
a/components/camel-aws/camel-aws2-timestream/src/main/java/org/apache/camel/component/aws2/timestream/write/Timestream2WriteProducer.java
+++
b/components/camel-aws/camel-aws2-timestream/src/main/java/org/apache/camel/component/aws2/timestream/write/Timestream2WriteProducer.java
@@ -714,8 +714,8 @@ public class Timestream2WriteProducer extends
DefaultProducer {
builder.tableName(tableName);
}
if
(ObjectHelper.isNotEmpty(exchange.getIn().getHeader(Timestream2Constants.RECORD)))
{
- Record record =
exchange.getIn().getHeader(Timestream2Constants.RECORD, Record.class);
- builder.commonAttributes(record);
+ Record recordObject =
exchange.getIn().getHeader(Timestream2Constants.RECORD, Record.class);
+ builder.commonAttributes(recordObject);
}
if
(ObjectHelper.isNotEmpty(exchange.getIn().getHeader(Timestream2Constants.RECORD_LIST)))
{
List<Record> records =
exchange.getIn().getHeader(Timestream2Constants.RECORD_LIST, List.class);
diff --git
a/components/camel-azure/camel-azure-cosmosdb/src/main/java/org/apache/camel/component/azure/cosmosdb/CosmosDbConsumer.java
b/components/camel-azure/camel-azure-cosmosdb/src/main/java/org/apache/camel/component/azure/cosmosdb/CosmosDbConsumer.java
index 91e67820f67..054e7ede630 100644
---
a/components/camel-azure/camel-azure-cosmosdb/src/main/java/org/apache/camel/component/azure/cosmosdb/CosmosDbConsumer.java
+++
b/components/camel-azure/camel-azure-cosmosdb/src/main/java/org/apache/camel/component/azure/cosmosdb/CosmosDbConsumer.java
@@ -82,8 +82,8 @@ public class CosmosDbConsumer extends DefaultConsumer {
return (CosmosDbEndpoint) super.getEndpoint();
}
- private void onEventListener(final List<Map<String, ?>> record) {
- final Exchange exchange = createAzureCosmosDbExchange(record);
+ private void onEventListener(final List<Map<String, ?>> recordList) {
+ final Exchange exchange = createAzureCosmosDbExchange(recordList);
// add exchange callback
exchange.getExchangeExtension().addOnCompletion(onCompletion);
@@ -91,11 +91,11 @@ public class CosmosDbConsumer extends DefaultConsumer {
getAsyncProcessor().process(exchange, EmptyAsyncCallback.get());
}
- private Exchange createAzureCosmosDbExchange(final List<Map<String, ?>>
record) {
+ private Exchange createAzureCosmosDbExchange(final List<Map<String, ?>>
recordList) {
final Exchange exchange = createExchange(true);
final Message message = exchange.getIn();
- message.setBody(record);
+ message.setBody(recordList);
return exchange;
}
diff --git
a/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyCsvFactory.java
b/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyCsvFactory.java
index a24adf46cda..5d12684e2f7 100644
---
a/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyCsvFactory.java
+++
b/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyCsvFactory.java
@@ -627,71 +627,71 @@ public class BindyCsvFactory extends BindyAbstractFactory
implements BindyFactor
for (Class<?> cl : models) {
// Get annotation @CsvRecord from the class
- CsvRecord record = cl.getAnnotation(CsvRecord.class);
+ CsvRecord csvRecord = cl.getAnnotation(CsvRecord.class);
// Get annotation @Section from the class
Section section = cl.getAnnotation(Section.class);
- if (record != null) {
- LOG.debug("Csv record: {}", record);
+ if (csvRecord != null) {
+ LOG.debug("Csv record: {}", csvRecord);
// Get skipFirstLine parameter
- skipFirstLine = record.skipFirstLine();
+ skipFirstLine = csvRecord.skipFirstLine();
LOG.debug("Skip First Line parameter of the CSV: {}",
skipFirstLine);
// Get skipFirstLine parameter
- skipField = record.skipField();
+ skipField = csvRecord.skipField();
LOG.debug("Skip Field parameter of the CSV: {}",
skipField);
// Get generateHeaderColumnNames parameter
- generateHeaderColumnNames = record.generateHeaderColumns();
+ generateHeaderColumnNames =
csvRecord.generateHeaderColumns();
LOG.debug("Generate header column names parameter of the
CSV: {}", generateHeaderColumnNames);
// Get Separator parameter
-
org.apache.camel.util.ObjectHelper.notNull(record.separator(),
+
org.apache.camel.util.ObjectHelper.notNull(csvRecord.separator(),
"No separator has been defined in the @Record
annotation");
- separator = record.separator();
+ separator = csvRecord.separator();
LOG.debug("Separator defined for the CSV: {}", separator);
// Get carriage return parameter
- crlf = record.crlf();
+ crlf = csvRecord.crlf();
LOG.debug("Carriage return defined for the CSV: {}", crlf);
// Get isOrdered parameter
- messageOrdered = record.isOrdered();
+ messageOrdered = csvRecord.isOrdered();
LOG.debug("Must CSV record be ordered: {}",
messageOrdered);
- if
(org.apache.camel.util.ObjectHelper.isNotEmpty(record.quote())) {
- quote = record.quote();
+ if
(org.apache.camel.util.ObjectHelper.isNotEmpty(csvRecord.quote())) {
+ quote = csvRecord.quote();
LOG.debug("Quoting columns with: {}", quote);
}
- quoting = record.quoting();
+ quoting = csvRecord.quoting();
LOG.debug("CSV will be quoted: {}", quoting);
- autospanLine = record.autospanLine();
+ autospanLine = csvRecord.autospanLine();
LOG.debug("Autospan line in last record: {}",
autospanLine);
// Get allowEmptyStream parameter
- allowEmptyStream = record.allowEmptyStream();
+ allowEmptyStream = csvRecord.allowEmptyStream();
LOG.debug("Allow empty stream parameter of the CSV: {}",
allowEmptyStream);
// Get quotingEscaped parameter
- quotingEscaped = record.quotingEscaped();
+ quotingEscaped = csvRecord.quotingEscaped();
LOG.debug("Escape quote character flag of the CSV: {}",
quotingEscaped);
// Get quotingOnlyWhenNeeded parameter
- quotingOnlyWhenNeeded = record.quotingOnlyWhenNeeded();
+ quotingOnlyWhenNeeded = csvRecord.quotingOnlyWhenNeeded();
LOG.debug("Quoting only when needed: {}",
quotingOnlyWhenNeeded);
// Get endWithLineBreak parameter
- endWithLineBreak = record.endWithLineBreak();
+ endWithLineBreak = csvRecord.endWithLineBreak();
LOG.debug("End with line break: {}", endWithLineBreak);
- removeQuotes = record.removeQuotes();
+ removeQuotes = csvRecord.removeQuotes();
LOG.debug("Remove quotes: {}", removeQuotes);
- trimLine = record.trimLine();
+ trimLine = csvRecord.trimLine();
LOG.debug("Trim line: {}", trimLine);
}
diff --git
a/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyFixedLengthFactory.java
b/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyFixedLengthFactory.java
index 88ce8298311..20f0048b1eb 100644
---
a/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyFixedLengthFactory.java
+++
b/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyFixedLengthFactory.java
@@ -172,7 +172,7 @@ public class BindyFixedLengthFactory extends
BindyAbstractFactory implements Bin
String delimiter;
Field field;
- final UnicodeHelper record = new UnicodeHelper(
+ final UnicodeHelper unicodeHelper = new UnicodeHelper(
recordStr, (this.countGrapheme) ?
UnicodeHelper.Method.GRAPHEME : UnicodeHelper.Method.CODEPOINTS);
// Iterate through the list of positions
@@ -211,19 +211,19 @@ public class BindyFixedLengthFactory extends
BindyAbstractFactory implements Bin
}
if (length > 0) {
- if (record.length() < offset) {
+ if (unicodeHelper.length() < offset) {
token = "";
} else {
int endIndex = offset + length - 1;
- if (endIndex > record.length()) {
- endIndex = record.length();
+ if (endIndex > unicodeHelper.length()) {
+ endIndex = unicodeHelper.length();
}
- token = record.substring(offset - 1, endIndex);
+ token = unicodeHelper.substring(offset - 1, endIndex);
}
offset += length;
} else if (!delimiter.isEmpty()) {
final UnicodeHelper tempToken = new UnicodeHelper(
- record.substring(offset - 1, record.length()),
+ unicodeHelper.substring(offset - 1,
unicodeHelper.length()),
(this.countGrapheme) ? UnicodeHelper.Method.GRAPHEME :
UnicodeHelper.Method.CODEPOINTS);
token = tempToken.substring(0, tempToken.indexOf(delimiter));
// include the delimiter in the offset calculation
@@ -324,7 +324,7 @@ public class BindyFixedLengthFactory extends
BindyAbstractFactory implements Bin
}
// check for unmapped non-whitespace data at the end of the line
- if (offset <= record.length() && !(record.substring(offset - 1,
record.length())).isBlank()
+ if (offset <= unicodeHelper.length() &&
!(unicodeHelper.substring(offset - 1, unicodeHelper.length())).isBlank()
&& !isIgnoreTrailingChars()) {
throw new IllegalArgumentException(
"Unexpected / unmapped characters found at the end of the
fixed-length record at line : " + line);
@@ -566,36 +566,36 @@ public class BindyFixedLengthFactory extends
BindyAbstractFactory implements Bin
for (Class<?> cl : models) {
// Get annotation @FixedLengthRecord from the class
- FixedLengthRecord record =
cl.getAnnotation(FixedLengthRecord.class);
+ FixedLengthRecord fixedLengthRecord =
cl.getAnnotation(FixedLengthRecord.class);
- if (record != null) {
- LOG.debug("Fixed length record: {}", record);
+ if (fixedLengthRecord != null) {
+ LOG.debug("Fixed length record: {}", fixedLengthRecord);
// Get carriage return parameter
- crlf = record.crlf();
+ crlf = fixedLengthRecord.crlf();
LOG.debug("Carriage return defined for the CSV: {}", crlf);
- eol = record.eol();
+ eol = fixedLengthRecord.eol();
LOG.debug("EOL(end-of-line) defined for the CSV: {}", eol);
// Get header parameter
- header = record.header();
+ header = fixedLengthRecord.header();
LOG.debug("Header: {}", header);
hasHeader = header != void.class;
LOG.debug("Has Header: {}", hasHeader);
// Get skipHeader parameter
- skipHeader = record.skipHeader();
+ skipHeader = fixedLengthRecord.skipHeader();
LOG.debug("Skip Header: {}", skipHeader);
// Get footer parameter
- footer = record.footer();
+ footer = fixedLengthRecord.footer();
LOG.debug("Footer: {}", footer);
- hasFooter = record.footer() != void.class;
+ hasFooter = fixedLengthRecord.footer() != void.class;
LOG.debug("Has Footer: {}", hasFooter);
// Get skipFooter parameter
- skipFooter = record.skipFooter();
+ skipFooter = fixedLengthRecord.skipFooter();
LOG.debug("Skip Footer: {}", skipFooter);
// Get isHeader parameter
@@ -607,21 +607,21 @@ public class BindyFixedLengthFactory extends
BindyAbstractFactory implements Bin
LOG.debug("Is Footer: {}", isFooter);
// Get padding character
- paddingChar = record.paddingChar();
+ paddingChar = fixedLengthRecord.paddingChar();
LOG.debug("Padding char: {}", paddingChar);
// Get length of the record
- recordLength = record.length();
+ recordLength = fixedLengthRecord.length();
LOG.debug("Length of the record: {}", recordLength);
// Get flag for ignore trailing characters
- ignoreTrailingChars = record.ignoreTrailingChars();
+ ignoreTrailingChars = fixedLengthRecord.ignoreTrailingChars();
LOG.debug("Ignore trailing chars: {}", ignoreTrailingChars);
- ignoreMissingChars = record.ignoreMissingChars();
+ ignoreMissingChars = fixedLengthRecord.ignoreMissingChars();
LOG.debug("Enable ignore missing chars: {}",
ignoreMissingChars);
- countGrapheme = record.countGrapheme();
+ countGrapheme = fixedLengthRecord.countGrapheme();
LOG.debug("Enable grapheme counting instead of codepoints:
{}", countGrapheme);
}
}
diff --git
a/components/camel-braintree/src/main/java/org/apache/camel/component/braintree/internal/BraintreeLogHandler.java
b/components/camel-braintree/src/main/java/org/apache/camel/component/braintree/internal/BraintreeLogHandler.java
index b0b3fdec554..5fcedcd2fbd 100644
---
a/components/camel-braintree/src/main/java/org/apache/camel/component/braintree/internal/BraintreeLogHandler.java
+++
b/components/camel-braintree/src/main/java/org/apache/camel/component/braintree/internal/BraintreeLogHandler.java
@@ -50,46 +50,46 @@ public final class BraintreeLogHandler extends Handler {
}
@Override
- public void publish(LogRecord record) {
- if (record != null) {
- final Logger logger = getLogger(record);
- final String message = formatMessage(record);
+ public void publish(LogRecord logRecord) {
+ if (logRecord != null) {
+ final Logger logger = getLogger(logRecord);
+ final String message = formatMessage(logRecord);
- int level = record.getLevel().intValue();
+ int level = logRecord.getLevel().intValue();
if (level <= TRACE_LEVEL_THRESHOLD) {
- logger.trace(message, record.getThrown());
+ logger.trace(message, logRecord.getThrown());
} else if (level <= DEBUG_LEVEL_THRESHOLD) {
- logger.debug(message, record.getThrown());
+ logger.debug(message, logRecord.getThrown());
} else if (level <= INFO_LEVEL_THRESHOLD) {
- logger.info(message, record.getThrown());
+ logger.info(message, logRecord.getThrown());
} else if (level <= WARN_LEVEL_THRESHOLD) {
- logger.warn(message, record.getThrown());
+ logger.warn(message, logRecord.getThrown());
} else {
- logger.error(message, record.getThrown());
+ logger.error(message, logRecord.getThrown());
}
}
}
- private Logger getLogger(LogRecord record) {
- String name = record.getLoggerName();
+ private Logger getLogger(LogRecord logRecord) {
+ String name = logRecord.getLoggerName();
if (name == null) {
name = DEFAULT_LOGGER_NAME;
}
return LoggerFactory.getLogger(name);
}
- private String formatMessage(LogRecord record) {
- String message = record.getMessage();
+ private String formatMessage(LogRecord logRecord) {
+ String message = logRecord.getMessage();
if (message != null) {
- ResourceBundle bundle = record.getResourceBundle();
+ ResourceBundle bundle = logRecord.getResourceBundle();
if (bundle != null) {
try {
message = bundle.getString(message);
} catch (MissingResourceException e) {
}
}
- Object[] params = record.getParameters();
+ Object[] params = logRecord.getParameters();
// avoid formatting when there are no or 0 parameters. see also
// http://jira.qos.ch/browse/SLF4J-203
if (params != null && params.length > 0) {
diff --git
a/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverter.java
b/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverter.java
index acaf80d9d90..6f03d48da20 100644
---
a/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverter.java
+++
b/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverter.java
@@ -31,8 +31,8 @@ public interface CsvRecordConverter<T> {
/**
* Converts the CSV record into another type.
*
- * @param record CSV record to convert
+ * @param csvRecord CSV record to convert
* @return converted CSV record
*/
- T convertRecord(CSVRecord record);
+ T convertRecord(CSVRecord csvRecord);
}
diff --git
a/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverters.java
b/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverters.java
index cb7cc41b127..4c9ec4d5b27 100644
---
a/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverters.java
+++
b/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverters.java
@@ -45,10 +45,10 @@ final class CsvRecordConverters {
private static final ListCsvRecordConverter SINGLETON = new
ListCsvRecordConverter();
@Override
- public List<String> convertRecord(CSVRecord record) {
- List<String> answer = new ArrayList<>(record.size());
- for (int i = 0; i < record.size(); i++) {
- answer.add(record.get(i));
+ public List<String> convertRecord(CSVRecord csvRecord) {
+ List<String> answer = new ArrayList<>(csvRecord.size());
+ for (int i = 0; i < csvRecord.size(); i++) {
+ answer.add(csvRecord.get(i));
}
return answer;
}
@@ -67,8 +67,8 @@ final class CsvRecordConverters {
private static final MapCsvRecordConverter SINGLETON = new
MapCsvRecordConverter();
@Override
- public Map<String, String> convertRecord(CSVRecord record) {
- return record.toMap();
+ public Map<String, String> convertRecord(CSVRecord csvRecord) {
+ return csvRecord.toMap();
}
}
@@ -86,8 +86,8 @@ final class CsvRecordConverters {
@Override
@SuppressWarnings("unchecked")
- public Map<String, String> convertRecord(CSVRecord record) {
- return record.toMap();
+ public Map<String, String> convertRecord(CSVRecord csvRecord) {
+ return csvRecord.toMap();
}
}
}
diff --git
a/components/camel-debezium/camel-debezium-common/camel-debezium-common-component/src/main/java/org/apache/camel/component/debezium/DebeziumEndpoint.java
b/components/camel-debezium/camel-debezium-common/camel-debezium-common-component/src/main/java/org/apache/camel/component/debezium/DebeziumEndpoint.java
index 311944f0d4a..9ca582f2ab2 100644
---
a/components/camel-debezium/camel-debezium-common/camel-debezium-common-component/src/main/java/org/apache/camel/component/debezium/DebeziumEndpoint.java
+++
b/components/camel-debezium/camel-debezium-common/camel-debezium-common-component/src/main/java/org/apache/camel/component/debezium/DebeziumEndpoint.java
@@ -64,7 +64,7 @@ public abstract class DebeziumEndpoint<C extends
EmbeddedDebeziumConfiguration>
"DebeziumConsumer");
}
- public Exchange createDbzExchange(DebeziumConsumer consumer, final
SourceRecord record) {
+ public Exchange createDbzExchange(DebeziumConsumer consumer, final
SourceRecord sourceRecord) {
final Exchange exchange;
if (consumer != null) {
exchange = consumer.createExchange(false);
@@ -74,8 +74,8 @@ public abstract class DebeziumEndpoint<C extends
EmbeddedDebeziumConfiguration>
final Message message = exchange.getIn();
- final Schema valueSchema = record.valueSchema();
- final Object value = record.value();
+ final Schema valueSchema = sourceRecord.valueSchema();
+ final Object value = sourceRecord.value();
// extract values from SourceRecord
final Map<String, Object> sourceMetadata =
extractSourceMetadataValueFromValueStruct(valueSchema, value);
@@ -85,8 +85,8 @@ public abstract class DebeziumEndpoint<C extends
EmbeddedDebeziumConfiguration>
final Object timestamp = extractValueFromValueStruct(valueSchema,
value, Envelope.FieldName.TIMESTAMP);
final Object ddl = extractValueFromValueStruct(valueSchema, value,
HistoryRecord.Fields.DDL_STATEMENTS);
// set message headers
- message.setHeader(DebeziumConstants.HEADER_IDENTIFIER, record.topic());
- message.setHeader(DebeziumConstants.HEADER_KEY, record.key());
+ message.setHeader(DebeziumConstants.HEADER_IDENTIFIER,
sourceRecord.topic());
+ message.setHeader(DebeziumConstants.HEADER_KEY, sourceRecord.key());
message.setHeader(DebeziumConstants.HEADER_SOURCE_METADATA,
sourceMetadata);
message.setHeader(DebeziumConstants.HEADER_OPERATION, operation);
message.setHeader(DebeziumConstants.HEADER_BEFORE, before);
diff --git
a/components/camel-dns/src/main/java/org/apache/camel/component/dns/cloud/DnsServiceDiscovery.java
b/components/camel-dns/src/main/java/org/apache/camel/component/dns/cloud/DnsServiceDiscovery.java
index 735a8e4b972..022bfe4e2a3 100644
---
a/components/camel-dns/src/main/java/org/apache/camel/component/dns/cloud/DnsServiceDiscovery.java
+++
b/components/camel-dns/src/main/java/org/apache/camel/component/dns/cloud/DnsServiceDiscovery.java
@@ -59,7 +59,7 @@ public final class DnsServiceDiscovery extends
DefaultServiceDiscovery {
.filter(SRVRecord.class::isInstance)
.map(SRVRecord.class::cast)
.sorted(COMPARATOR)
- .map(record -> asService(name, record))
+ .map(srvRecord -> asService(name, srvRecord))
.collect(Collectors.toList());
} else {
services = Collections.emptyList();
@@ -85,15 +85,15 @@ public final class DnsServiceDiscovery extends
DefaultServiceDiscovery {
return byPriority.thenComparing(byWeight);
}
- private static ServiceDefinition asService(String serviceName, SRVRecord
record) {
+ private static ServiceDefinition asService(String serviceName, SRVRecord
srvRecord) {
Map<String, String> meta = new HashMap<>();
- ObjectHelper.ifNotEmpty(record.getPriority(), val ->
meta.put("priority", Integer.toString(val)));
- ObjectHelper.ifNotEmpty(record.getWeight(), val -> meta.put("weight",
Integer.toString(val)));
+ ObjectHelper.ifNotEmpty(srvRecord.getPriority(), val ->
meta.put("priority", Integer.toString(val)));
+ ObjectHelper.ifNotEmpty(srvRecord.getWeight(), val ->
meta.put("weight", Integer.toString(val)));
return new DefaultServiceDefinition(
serviceName,
- record.getTarget().toString(true),
- record.getPort(),
+ srvRecord.getTarget().toString(true),
+ srvRecord.getPort(),
meta);
}
}
diff --git
a/components/camel-dns/src/main/java/org/apache/camel/component/dns/types/DnsConverter.java
b/components/camel-dns/src/main/java/org/apache/camel/component/dns/types/DnsConverter.java
index 4876a68cac9..4cc0f9f0883 100644
---
a/components/camel-dns/src/main/java/org/apache/camel/component/dns/types/DnsConverter.java
+++
b/components/camel-dns/src/main/java/org/apache/camel/component/dns/types/DnsConverter.java
@@ -36,8 +36,8 @@ public final class DnsConverter {
}
@Converter
- public static String toString(Record record) {
- return record.toString();
+ public static String toString(Record recordObject) {
+ return recordObject.toString();
}
@Converter
diff --git
a/components/camel-flatpack/src/main/java/org/apache/camel/component/flatpack/FlatpackConverter.java
b/components/camel-flatpack/src/main/java/org/apache/camel/component/flatpack/FlatpackConverter.java
index 12ff3c4b7d9..6942677ee2b 100644
---
a/components/camel-flatpack/src/main/java/org/apache/camel/component/flatpack/FlatpackConverter.java
+++
b/components/camel-flatpack/src/main/java/org/apache/camel/component/flatpack/FlatpackConverter.java
@@ -40,12 +40,12 @@ public final class FlatpackConverter {
}
@Converter
- public static Map<String, Object> toMap(Record record) {
+ public static Map<String, Object> toMap(Record recordObj) {
Map<String, Object> map = new HashMap<>();
- if (record instanceof DataSet dataSet) {
+ if (recordObj instanceof DataSet dataSet) {
putValues(map, dataSet);
} else {
- putValues(map, record);
+ putValues(map, recordObj);
}
return map;
@@ -110,23 +110,23 @@ public final class FlatpackConverter {
/**
* Puts the values of the record into the map
*/
- private static void putValues(Map<String, Object> map, Record record) {
- String[] columns = record.getColumns();
+ private static void putValues(Map<String, Object> map, Record recordObj) {
+ String[] columns = recordObj.getColumns();
for (String column : columns) {
- String value = record.getString(column);
+ String value = recordObj.getString(column);
map.put(column, value);
}
}
private static Element createDatasetRecord(DataSet dataSet, Document doc) {
- Element record;
+ Element element;
if (dataSet.isRecordID(FlatpackComponent.HEADER_ID)) {
- record = doc.createElement("DatasetHeader");
+ element = doc.createElement("DatasetHeader");
} else if (dataSet.isRecordID(FlatpackComponent.TRAILER_ID)) {
- record = doc.createElement("DatasetTrailer");
+ element = doc.createElement("DatasetTrailer");
} else {
- record = doc.createElement("DatasetRecord");
+ element = doc.createElement("DatasetRecord");
}
String[] columns = getColumns(dataSet);
@@ -138,10 +138,10 @@ public final class FlatpackConverter {
columnElement.setAttribute("name", column);
columnElement.setTextContent(value);
- record.appendChild(columnElement);
+ element.appendChild(columnElement);
}
- return record;
+ return element;
}
private static String[] getColumns(DataSet dataSet) {
diff --git
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/InfluxDb2Producer.java
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/InfluxDb2Producer.java
index 7a8abfc75f0..417414a552d 100644
---
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/InfluxDb2Producer.java
+++
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/InfluxDb2Producer.java
@@ -120,12 +120,12 @@ public class InfluxDb2Producer extends DefaultProducer {
private void insertRecord(Exchange exchange, String orgName, String
bucketName, WritePrecision writePrecision)
throws InvalidPayloadException {
- Record record = exchange.getIn().getMandatoryBody(Record.class);
+ Record recordObj = exchange.getIn().getMandatoryBody(Record.class);
try {
if (LOG.isDebugEnabled()) {
- LOG.debug("Writing record {}", record);
+ LOG.debug("Writing record {}", recordObj);
}
- writeApi.writeRecord(bucketName, orgName, writePrecision,
record.getInfluxRecord());
+ writeApi.writeRecord(bucketName, orgName, writePrecision,
recordObj.getInfluxRecord());
} catch (Exception ex) {
exchange.setException(new CamelInfluxDb2Exception(ex));
}
diff --git
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Record.java
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Record.java
index 89ef913655f..ab4589c9990 100644
---
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Record.java
+++
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Record.java
@@ -21,10 +21,10 @@ package org.apache.camel.component.influxdb2.data;
* Influxdb write Record
*/
public class Record {
- private String record;
+ private String recordObj;
- public Record(String record) {
- this.record = record;
+ public Record(String recordObj) {
+ this.recordObj = recordObj;
}
public static Record fromString(String record) {
@@ -32,10 +32,10 @@ public class Record {
}
public String getInfluxRecord() {
- return record;
+ return recordObj;
}
- public void setRecord(String record) {
- this.record = record;
+ public void setRecord(String recordObj) {
+ this.recordObj = recordObj;
}
}
diff --git
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Records.java
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Records.java
index 8f4ffb2380b..dc4f6ee6d4b 100644
---
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Records.java
+++
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Records.java
@@ -53,21 +53,21 @@ public class Records {
return new Records(records);
}
- public static Records create(Record record) {
- return new Records().addRecord(record);
+ public static Records create(Record recordObj) {
+ return new Records().addRecord(recordObj);
}
- public static Records create(String record) {
- return create(Record.fromString(record));
+ public static Records create(String recordObj) {
+ return create(Record.fromString(recordObj));
}
- public Records addRecord(String record) {
- this.records.add(Record.fromString(record));
+ public Records addRecord(String recordObj) {
+ this.records.add(Record.fromString(recordObj));
return this;
}
- public Records addRecord(Record record) {
- this.records.add(record);
+ public Records addRecord(Record recordObj) {
+ this.records.add(recordObj);
return this;
}
diff --git
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AbstractCommitManager.java
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AbstractCommitManager.java
index d35ac3d65c3..1e7638a80fa 100644
---
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AbstractCommitManager.java
+++
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AbstractCommitManager.java
@@ -71,14 +71,14 @@ public abstract class AbstractCommitManager implements
CommitManager {
@Override
public KafkaManualCommit getManualCommit(
- Exchange exchange, TopicPartition partition,
ConsumerRecord<Object, Object> record) {
+ Exchange exchange, TopicPartition partition,
ConsumerRecord<Object, Object> consumerRecord) {
KafkaManualCommitFactory manualCommitFactory =
kafkaConsumer.getEndpoint().getKafkaManualCommitFactory();
if (manualCommitFactory == null) {
manualCommitFactory = new DefaultKafkaManualCommitFactory();
}
- return getManualCommit(exchange, partition, record,
manualCommitFactory);
+ return getManualCommit(exchange, partition, consumerRecord,
manualCommitFactory);
}
@Override
diff --git
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AsyncCommitManager.java
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AsyncCommitManager.java
index 7e117ee4d61..b7f776ac56b 100644
---
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AsyncCommitManager.java
+++
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AsyncCommitManager.java
@@ -75,14 +75,14 @@ public class AsyncCommitManager extends
AbstractCommitManager {
@Override
public KafkaManualCommit getManualCommit(
- Exchange exchange, TopicPartition partition,
ConsumerRecord<Object, Object> record) {
+ Exchange exchange, TopicPartition partition,
ConsumerRecord<Object, Object> consumerRecord) {
KafkaManualCommitFactory manualCommitFactory =
kafkaConsumer.getEndpoint().getKafkaManualCommitFactory();
if (manualCommitFactory == null) {
manualCommitFactory = new DefaultKafkaManualAsyncCommitFactory();
}
- return getManualCommit(exchange, partition, record,
manualCommitFactory);
+ return getManualCommit(exchange, partition, consumerRecord,
manualCommitFactory);
}
@Override
diff --git
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/CommitManager.java
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/CommitManager.java
index 999fde26563..c852f6a5fd2 100644
---
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/CommitManager.java
+++
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/CommitManager.java
@@ -23,7 +23,7 @@ import org.apache.kafka.common.TopicPartition;
public interface CommitManager {
- KafkaManualCommit getManualCommit(Exchange exchange, TopicPartition
partition, ConsumerRecord<Object, Object> record);
+ KafkaManualCommit getManualCommit(Exchange exchange, TopicPartition
partition, ConsumerRecord<Object, Object> consumerRecord);
/**
* Commits everything that has been cached
diff --git
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessor.java
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessor.java
index 86af7396d87..2954836f5f6 100644
---
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessor.java
+++
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessor.java
@@ -128,17 +128,17 @@ public class KafkaRecordProcessor {
private boolean processException(
Exchange exchange, TopicPartition topicPartition,
- ConsumerRecord<Object, Object> record, ProcessingResult lastResult,
+ ConsumerRecord<Object, Object> consumerRecord, ProcessingResult
lastResult,
ExceptionHandler exceptionHandler) {
// processing failed due to an unhandled exception, what should we do
if (configuration.isBreakOnFirstError()) {
if (lastResult.getPartition() != -1 &&
- lastResult.getPartition() != record.partition()) {
+ lastResult.getPartition() != consumerRecord.partition()) {
LOG.error("About to process an exception with UNEXPECTED
partition & offset. Got topic partition {}. " +
" The last result was on partition {} with offset {}
but was expecting partition {} with offset {}",
topicPartition.partition(), lastResult.getPartition(),
lastResult.getPartitionLastOffset(),
- record.partition(), record.offset());
+ consumerRecord.partition(), consumerRecord.offset());
}
// we are failing and we should break out
@@ -147,7 +147,7 @@ public class KafkaRecordProcessor {
LOG.warn("Error during processing {} from topic: {} due to
{}", exchange, topicPartition.topic(),
exc.getMessage());
LOG.warn("Will seek consumer to offset {} on partition {} and
start polling again.",
- record.offset(), record.partition());
+ consumerRecord.offset(), consumerRecord.partition());
}
// force commit, so we resume on next poll where we failed
diff --git
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessorFacade.java
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessorFacade.java
index 44573daa60d..5c7060400d7 100644
---
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessorFacade.java
+++
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessorFacade.java
@@ -77,16 +77,16 @@ public class KafkaRecordProcessorFacade {
logRecordsInPartition(partitionRecords, partition);
while (!lastResult.isBreakOnErrorHit() && recordIterator.hasNext()
&& !isStopping()) {
- ConsumerRecord<Object, Object> record = recordIterator.next();
+ ConsumerRecord<Object, Object> consumerRecord =
recordIterator.next();
- LOG.debug("Processing record on partition {} with offset {}",
record.partition(), record.offset());
+ LOG.debug("Processing record on partition {} with offset {}",
consumerRecord.partition(), consumerRecord.offset());
lastResult = processRecord(partition,
partitionIterator.hasNext(), recordIterator.hasNext(), lastResult,
- kafkaRecordProcessor, record);
+ kafkaRecordProcessor, consumerRecord);
LOG.debug(
"Processed record on partition {} with offset {} and
got ProcessingResult for partition {} and offset {}",
- record.partition(), record.offset(),
lastResult.getPartition(), lastResult.getPartitionLastOffset());
+ consumerRecord.partition(), consumerRecord.offset(),
lastResult.getPartition(), lastResult.getPartitionLastOffset());
if (consumerListener != null) {
if (!consumerListener.afterProcess(lastResult)) {
@@ -125,15 +125,15 @@ public class KafkaRecordProcessorFacade {
boolean recordHasNext,
final ProcessingResult lastResult,
KafkaRecordProcessor kafkaRecordProcessor,
- ConsumerRecord<Object, Object> record) {
+ ConsumerRecord<Object, Object> consumerRecord) {
- logRecord(record);
+ logRecord(consumerRecord);
Exchange exchange = camelKafkaConsumer.createExchange(false);
ProcessingResult currentResult
= kafkaRecordProcessor.processExchange(exchange, partition,
partitionHasNext,
- recordHasNext, record, lastResult,
camelKafkaConsumer.getExceptionHandler());
+ recordHasNext, consumerRecord, lastResult,
camelKafkaConsumer.getExceptionHandler());
if (!currentResult.isBreakOnErrorHit()) {
commitManager.recordOffset(partition,
currentResult.getPartitionLastOffset());
@@ -145,10 +145,10 @@ public class KafkaRecordProcessorFacade {
return currentResult;
}
- private void logRecord(ConsumerRecord<Object, Object> record) {
+ private void logRecord(ConsumerRecord<Object, Object> consumerRecord) {
if (LOG.isTraceEnabled()) {
- LOG.trace("Partition = {}, offset = {}, key = {}, value = {}",
record.partition(),
- record.offset(), record.key(), record.value());
+ LOG.trace("Partition = {}, offset = {}, key = {}, value = {}",
consumerRecord.partition(),
+ consumerRecord.offset(), consumerRecord.key(),
consumerRecord.value());
}
}
diff --git
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/producer/support/KafkaProducerMetadataCallBack.java
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/producer/support/KafkaProducerMetadataCallBack.java
index 37db4907049..053307a96b7 100644
---
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/producer/support/KafkaProducerMetadataCallBack.java
+++
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/producer/support/KafkaProducerMetadataCallBack.java
@@ -24,18 +24,18 @@ import static
org.apache.camel.component.kafka.producer.support.ProducerUtil.set
public class KafkaProducerMetadataCallBack implements Callback {
private final Object body;
- private final boolean record;
+ private final boolean recordMetadata;
- public KafkaProducerMetadataCallBack(Object body, boolean record) {
+ public KafkaProducerMetadataCallBack(Object body, boolean recordMetadata) {
this.body = body;
- this.record = record;
+ this.recordMetadata = recordMetadata;
}
@Override
public void onCompletion(RecordMetadata recordMetadata, Exception e) {
setException(body, e);
- if (record) {
+ if (this.recordMetadata) {
setRecordMetadata(body, recordMetadata);
}
}
diff --git
a/components/camel-kafka/src/main/java/org/apache/camel/processor/resume/kafka/SingleNodeKafkaResumeStrategy.java
b/components/camel-kafka/src/main/java/org/apache/camel/processor/resume/kafka/SingleNodeKafkaResumeStrategy.java
index 2f9113f2423..94be16f6258 100644
---
a/components/camel-kafka/src/main/java/org/apache/camel/processor/resume/kafka/SingleNodeKafkaResumeStrategy.java
+++
b/components/camel-kafka/src/main/java/org/apache/camel/processor/resume/kafka/SingleNodeKafkaResumeStrategy.java
@@ -108,9 +108,9 @@ public class SingleNodeKafkaResumeStrategy implements
KafkaResumeStrategy, Camel
*
*/
protected void produce(byte[] key, byte[] message, UpdateCallBack
updateCallBack) {
- ProducerRecord<byte[], byte[]> record = new
ProducerRecord<>(resumeStrategyConfiguration.getTopic(), key, message);
+ ProducerRecord<byte[], byte[]> producerRecord = new
ProducerRecord<>(resumeStrategyConfiguration.getTopic(), key, message);
- producer.send(record, (recordMetadata, e) -> {
+ producer.send(producerRecord, (recordMetadata, e) -> {
if (e != null) {
LOG.error("Failed to send message {}", e.getMessage(), e);
}
@@ -235,15 +235,15 @@ public class SingleNodeKafkaResumeStrategy implements
KafkaResumeStrategy, Camel
do {
ConsumerRecords<byte[], byte[]> records = consume(consumer);
- for (ConsumerRecord<byte[], byte[]> record : records) {
- byte[] value = record.value();
+ for (ConsumerRecord<byte[], byte[]> consumerRecord : records) {
+ byte[] value = consumerRecord.value();
if (LOG.isTraceEnabled()) {
- LOG.trace("Read from Kafka at {} ({}): {}",
Instant.ofEpochMilli(record.timestamp()),
- record.timestampType(), value);
+ LOG.trace("Read from Kafka at {} ({}): {}",
Instant.ofEpochMilli(consumerRecord.timestamp()),
+ consumerRecord.timestampType(), value);
}
- if (!deserializable.deserialize(ByteBuffer.wrap(record.key()),
ByteBuffer.wrap(record.value()))) {
+ if
(!deserializable.deserialize(ByteBuffer.wrap(consumerRecord.key()),
ByteBuffer.wrap(consumerRecord.value()))) {
LOG.warn("Deserializer indicates that this is the last
record to deserialize");
}
}
diff --git
a/components/camel-kubernetes/src/main/java/org/apache/camel/component/kubernetes/cloud/KubernetesDnsSrvServiceDiscovery.java
b/components/camel-kubernetes/src/main/java/org/apache/camel/component/kubernetes/cloud/KubernetesDnsSrvServiceDiscovery.java
index c3727547588..a4c203e9442 100644
---
a/components/camel-kubernetes/src/main/java/org/apache/camel/component/kubernetes/cloud/KubernetesDnsSrvServiceDiscovery.java
+++
b/components/camel-kubernetes/src/main/java/org/apache/camel/component/kubernetes/cloud/KubernetesDnsSrvServiceDiscovery.java
@@ -100,8 +100,8 @@ public class KubernetesDnsSrvServiceDiscovery extends
KubernetesServiceDiscovery
List<ServiceDefinition> servers = new LinkedList<>();
while (resolved.hasMore()) {
- String record = (String) resolved.next();
- String[] items = record.split(" ", -1);
+ String recordObj = (String) resolved.next();
+ String[] items = recordObj.split(" ", -1);
String host = items[3].trim();
String port = items[2].trim();
diff --git
a/components/camel-openapi-java/src/main/java/org/apache/camel/openapi/RestOpenApiSupport.java
b/components/camel-openapi-java/src/main/java/org/apache/camel/openapi/RestOpenApiSupport.java
index 764b5a12e65..6ec28f1f065 100644
---
a/components/camel-openapi-java/src/main/java/org/apache/camel/openapi/RestOpenApiSupport.java
+++
b/components/camel-openapi-java/src/main/java/org/apache/camel/openapi/RestOpenApiSupport.java
@@ -182,10 +182,10 @@ public class RestOpenApiSupport {
Pattern p = Pattern.compile("\\{(.*?)}");
Matcher m = p.matcher(url);
while (m.find()) {
- String var = m.group(1);
- if (server != null && server.getVariables() != null &&
server.getVariables().get(var) != null) {
- String varValue = server.getVariables().get(var).getDefault();
- url = url.replace("{" + var + "}", varValue);
+ String variable = m.group(1);
+ if (server != null && server.getVariables() != null &&
server.getVariables().get(variable) != null) {
+ String varValue =
server.getVariables().get(variable).getDefault();
+ url = url.replace("{" + variable + "}", varValue);
}
}
return url;
diff --git
a/components/camel-rest-openapi/src/main/java/org/apache/camel/component/rest/openapi/RestOpenApiEndpoint.java
b/components/camel-rest-openapi/src/main/java/org/apache/camel/component/rest/openapi/RestOpenApiEndpoint.java
index 67bfdbf219d..b6b20d75891 100644
---
a/components/camel-rest-openapi/src/main/java/org/apache/camel/component/rest/openapi/RestOpenApiEndpoint.java
+++
b/components/camel-rest-openapi/src/main/java/org/apache/camel/component/rest/openapi/RestOpenApiEndpoint.java
@@ -421,10 +421,10 @@ public final class RestOpenApiEndpoint extends
DefaultEndpoint {
Matcher m = p.matcher(url);
while (m.find()) {
- String var = m.group(1);
- if (server != null && server.getVariables() != null &&
server.getVariables().get(var) != null) {
- String varValue = server.getVariables().get(var).getDefault();
- url = url.replace("{" + var + "}", varValue);
+ String variable = m.group(1);
+ if (server != null && server.getVariables() != null &&
server.getVariables().get(variable) != null) {
+ String varValue =
server.getVariables().get(variable).getDefault();
+ url = url.replace("{" + variable + "}", varValue);
}
}
return url;
diff --git
a/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/PubSubApiConsumer.java
b/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/PubSubApiConsumer.java
index 66b9b15e6bf..e71d3eec75c 100644
---
a/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/PubSubApiConsumer.java
+++
b/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/PubSubApiConsumer.java
@@ -66,10 +66,10 @@ public class PubSubApiConsumer extends DefaultConsumer {
}
}
- public void processEvent(Object record, String replayId) throws
IOException {
+ public void processEvent(Object recordObj, String replayId) throws
IOException {
final Exchange exchange = createExchange(true);
final Message in = exchange.getIn();
- in.setBody(record);
+ in.setBody(recordObj);
in.setHeader(HEADER_SALESFORCE_PUBSUB_REPLAY_ID, replayId);
AsyncCallback cb = defaultConsumerCallback(exchange, true);
getAsyncProcessor().process(exchange, cb);
diff --git
a/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/internal/client/PubSubApiClient.java
b/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/internal/client/PubSubApiClient.java
index cb05734e191..d6caba9b317 100644
---
a/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/internal/client/PubSubApiClient.java
+++
b/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/internal/client/PubSubApiClient.java
@@ -240,8 +240,8 @@ public class PubSubApiClient extends ServiceSupport {
}
byte[] bytes;
if (body instanceof IndexedRecord indexedRecord) {
- if (body instanceof GenericRecord record) {
- bytes = getBytes(body, new
GenericDatumWriter<>(record.getSchema()));
+ if (body instanceof GenericRecord genericRecord) {
+ bytes = getBytes(body, new
GenericDatumWriter<>(genericRecord.getSchema()));
} else if (body instanceof SpecificRecord) {
bytes = getBytes(body, new SpecificDatumWriter<>());
} else {
@@ -367,7 +367,7 @@ public class PubSubApiClient extends ServiceSupport {
private void processEvent(ConsumerEvent ce) throws IOException {
final Schema schema = getSchema(ce.getEvent().getSchemaId());
- Object record = switch (consumer.getDeserializeType()) {
+ Object recordObj = switch (consumer.getDeserializeType()) {
case AVRO -> deserializeAvro(ce, schema);
case GENERIC_RECORD -> deserializeGenericRecord(ce, schema);
case SPECIFIC_RECORD -> deserializeSpecificRecord(ce, schema);
@@ -375,7 +375,7 @@ public class PubSubApiClient extends ServiceSupport {
case JSON -> deserializeJson(ce, schema);
};
String replayId =
PubSubApiClient.base64EncodeByteString(ce.getReplayId());
- consumer.processEvent(record, replayId);
+ consumer.processEvent(recordObj, replayId);
}
private Object deserializeAvro(ConsumerEvent ce, Schema schema) throws
IOException {
@@ -388,9 +388,9 @@ public class PubSubApiClient extends ServiceSupport {
}
private Object deserializeJson(ConsumerEvent ce, Schema schema) throws
IOException {
- final GenericRecord record = deserializeGenericRecord(ce, schema);
+ final GenericRecord genericRecord = deserializeGenericRecord(ce,
schema);
JsonAvroConverter converter = new JsonAvroConverter();
- final byte[] bytes = converter.convertToJson(record);
+ final byte[] bytes = converter.convertToJson(genericRecord);
return new String(bytes);
}
diff --git
a/core/camel-base-engine/src/main/java/org/apache/camel/impl/engine/RouteService.java
b/core/camel-base-engine/src/main/java/org/apache/camel/impl/engine/RouteService.java
index fd22706079d..98e5a3e6650 100644
---
a/core/camel-base-engine/src/main/java/org/apache/camel/impl/engine/RouteService.java
+++
b/core/camel-base-engine/src/main/java/org/apache/camel/impl/engine/RouteService.java
@@ -366,9 +366,9 @@ public class RouteService extends ChildServiceSupport {
for (Service service : services) {
StartupStep step = null;
// skip internal services / route pipeline (starting point for
route)
- boolean record
+ boolean shouldRecord
= !(service instanceof InternalProcessor ||
"RoutePipeline".equals(service.getClass().getSimpleName()));
- if (record) {
+ if (shouldRecord) {
step = beginStep(service, "Init");
}
ServiceHelper.initService(service);
@@ -384,9 +384,9 @@ public class RouteService extends ChildServiceSupport {
for (Service service : services) {
StartupStep step = null;
// skip internal services / route pipeline (starting point for
route)
- boolean record
+ boolean shouldRecord
= !(service instanceof InternalProcessor ||
"RoutePipeline".equals(service.getClass().getSimpleName()));
- if (record) {
+ if (shouldRecord) {
step = beginStep(service, "Start");
}
for (LifecycleStrategy strategy :
camelContext.getLifecycleStrategies()) {
diff --git
a/dsl/camel-kamelet-main/src/main/java/org/apache/camel/main/download/DownloadThreadPool.java
b/dsl/camel-kamelet-main/src/main/java/org/apache/camel/main/download/DownloadThreadPool.java
index e3bc60eba44..cf7dc6bf405 100644
---
a/dsl/camel-kamelet-main/src/main/java/org/apache/camel/main/download/DownloadThreadPool.java
+++
b/dsl/camel-kamelet-main/src/main/java/org/apache/camel/main/download/DownloadThreadPool.java
@@ -86,11 +86,11 @@ class DownloadThreadPool extends ServiceSupport implements
CamelContextAware {
}
MavenGav a = MavenGav.parseGav(gav);
- DownloadRecord record = downloader.getDownloadState(a.getGroupId(),
a.getArtifactId(), a.getVersion());
- if (record != null) {
+ DownloadRecord downloadRecord =
downloader.getDownloadState(a.getGroupId(), a.getArtifactId(), a.getVersion());
+ if (downloadRecord != null) {
long taken = watch.taken();
- String url = record.repoUrl();
- String id = record.repoId();
+ String url = downloadRecord.repoUrl();
+ String id = downloadRecord.repoId();
String msg = "Downloaded: " + gav + " (took: "
+ TimeUtils.printDuration(taken, true) + ") from: " +
id + "@" + url;
log.info(msg);