This is an automated email from the ASF dual-hosted git repository.

orpiske pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/camel.git


The following commit(s) were added to refs/heads/main by this push:
     new 642e185ad5b CAMEL-20161: fix usages of restricted identifiers (#12213)
642e185ad5b is described below

commit 642e185ad5ba8474d2c54a81c18622836b1a73bd
Author: Otavio Rodolfo Piske <[email protected]>
AuthorDate: Mon Nov 27 15:18:40 2023 +0100

    CAMEL-20161: fix usages of restricted identifiers (#12213)
---
 .../camel/dataformat/asn1/ASN1DataFormat.java      | 10 ++---
 .../aws2/ddbstream/Ddb2StreamConsumer.java         |  4 +-
 .../aws2/firehose/KinesisFirehose2Producer.java    |  6 +--
 .../timestream/write/Timestream2WriteProducer.java |  4 +-
 .../component/azure/cosmosdb/CosmosDbConsumer.java |  8 ++--
 .../camel/dataformat/bindy/BindyCsvFactory.java    | 40 ++++++++++----------
 .../dataformat/bindy/BindyFixedLengthFactory.java  | 44 +++++++++++-----------
 .../braintree/internal/BraintreeLogHandler.java    | 32 ++++++++--------
 .../camel/dataformat/csv/CsvRecordConverter.java   |  4 +-
 .../camel/dataformat/csv/CsvRecordConverters.java  | 16 ++++----
 .../camel/component/debezium/DebeziumEndpoint.java | 10 ++---
 .../component/dns/cloud/DnsServiceDiscovery.java   | 12 +++---
 .../camel/component/dns/types/DnsConverter.java    |  4 +-
 .../component/flatpack/FlatpackConverter.java      | 24 ++++++------
 .../component/influxdb2/InfluxDb2Producer.java     |  6 +--
 .../camel/component/influxdb2/data/Record.java     | 12 +++---
 .../camel/component/influxdb2/data/Records.java    | 16 ++++----
 .../kafka/consumer/AbstractCommitManager.java      |  4 +-
 .../kafka/consumer/AsyncCommitManager.java         |  4 +-
 .../component/kafka/consumer/CommitManager.java    |  2 +-
 .../consumer/support/KafkaRecordProcessor.java     |  8 ++--
 .../support/KafkaRecordProcessorFacade.java        | 20 +++++-----
 .../support/KafkaProducerMetadataCallBack.java     |  8 ++--
 .../kafka/SingleNodeKafkaResumeStrategy.java       | 14 +++----
 .../cloud/KubernetesDnsSrvServiceDiscovery.java    |  4 +-
 .../apache/camel/openapi/RestOpenApiSupport.java   |  8 ++--
 .../rest/openapi/RestOpenApiEndpoint.java          |  8 ++--
 .../component/salesforce/PubSubApiConsumer.java    |  4 +-
 .../internal/client/PubSubApiClient.java           | 12 +++---
 .../org/apache/camel/impl/engine/RouteService.java |  8 ++--
 .../camel/main/download/DownloadThreadPool.java    |  8 ++--
 31 files changed, 182 insertions(+), 182 deletions(-)

diff --git 
a/components/camel-asn1/src/main/java/org/apache/camel/dataformat/asn1/ASN1DataFormat.java
 
b/components/camel-asn1/src/main/java/org/apache/camel/dataformat/asn1/ASN1DataFormat.java
index 88d8d1abfdf..3d8630f2944 100644
--- 
a/components/camel-asn1/src/main/java/org/apache/camel/dataformat/asn1/ASN1DataFormat.java
+++ 
b/components/camel-asn1/src/main/java/org/apache/camel/dataformat/asn1/ASN1DataFormat.java
@@ -62,12 +62,12 @@ public class ASN1DataFormat extends ServiceSupport 
implements DataFormat, DataFo
                 encodeGenericTypeObject(exchange, stream);
                 return;
             }
-            Object record = exchange.getIn().getBody();
-            if (record instanceof ASN1Primitive) {
-                ASN1Primitive asn1Primitive = 
ObjectHelper.cast(ASN1Primitive.class, record);
+            Object body = exchange.getIn().getBody();
+            if (body instanceof ASN1Primitive) {
+                ASN1Primitive asn1Primitive = 
ObjectHelper.cast(ASN1Primitive.class, body);
                 berOut = new ByteArrayInputStream(asn1Primitive.getEncoded());
-            } else if (record instanceof byte[]) {
-                berOut = new 
ByteArrayInputStream(ObjectHelper.cast(byte[].class, record));
+            } else if (body instanceof byte[]) {
+                berOut = new 
ByteArrayInputStream(ObjectHelper.cast(byte[].class, body));
             }
         } else {
             byte[] byteInput = 
exchange.getContext().getTypeConverter().mandatoryConvertTo(byte[].class, 
exchange, graph);
diff --git 
a/components/camel-aws/camel-aws2-ddb/src/main/java/org/apache/camel/component/aws2/ddbstream/Ddb2StreamConsumer.java
 
b/components/camel-aws/camel-aws2-ddb/src/main/java/org/apache/camel/component/aws2/ddbstream/Ddb2StreamConsumer.java
index a51741329c7..62491771907 100644
--- 
a/components/camel-aws/camel-aws2-ddb/src/main/java/org/apache/camel/component/aws2/ddbstream/Ddb2StreamConsumer.java
+++ 
b/components/camel-aws/camel-aws2-ddb/src/main/java/org/apache/camel/component/aws2/ddbstream/Ddb2StreamConsumer.java
@@ -79,8 +79,8 @@ public class Ddb2StreamConsumer extends 
ScheduledBatchPollingConsumer {
             }
             List<Record> records = result.records();
             Queue<Exchange> exchanges = new ArrayDeque<>();
-            for (Record record : records) {
-                exchanges.add(createExchange(record));
+            for (Record polledRecord : records) {
+                exchanges.add(createExchange(polledRecord));
             }
             processedExchangeCount += processBatch(CastUtils.cast(exchanges));
 
diff --git 
a/components/camel-aws/camel-aws2-kinesis/src/main/java/org/apache/camel/component/aws2/firehose/KinesisFirehose2Producer.java
 
b/components/camel-aws/camel-aws2-kinesis/src/main/java/org/apache/camel/component/aws2/firehose/KinesisFirehose2Producer.java
index 0e8f4635dee..04a55dd7bee 100644
--- 
a/components/camel-aws/camel-aws2-kinesis/src/main/java/org/apache/camel/component/aws2/firehose/KinesisFirehose2Producer.java
+++ 
b/components/camel-aws/camel-aws2-kinesis/src/main/java/org/apache/camel/component/aws2/firehose/KinesisFirehose2Producer.java
@@ -178,12 +178,12 @@ public class KinesisFirehose2Producer extends 
DefaultProducer {
 
     private PutRecordRequest createRequest(Exchange exchange) {
         ByteBuffer body = exchange.getIn().getBody(ByteBuffer.class);
-        Record.Builder record = Record.builder();
-        record.data(SdkBytes.fromByteBuffer(body));
+        Record.Builder builder = Record.builder();
+        builder.data(SdkBytes.fromByteBuffer(body));
 
         PutRecordRequest.Builder putRecordRequest = PutRecordRequest.builder();
         
putRecordRequest.deliveryStreamName(getEndpoint().getConfiguration().getStreamName());
-        putRecordRequest.record(record.build());
+        putRecordRequest.record(builder.build());
         return putRecordRequest.build();
     }
 
diff --git 
a/components/camel-aws/camel-aws2-timestream/src/main/java/org/apache/camel/component/aws2/timestream/write/Timestream2WriteProducer.java
 
b/components/camel-aws/camel-aws2-timestream/src/main/java/org/apache/camel/component/aws2/timestream/write/Timestream2WriteProducer.java
index 18c346ab71a..4f0bb640917 100644
--- 
a/components/camel-aws/camel-aws2-timestream/src/main/java/org/apache/camel/component/aws2/timestream/write/Timestream2WriteProducer.java
+++ 
b/components/camel-aws/camel-aws2-timestream/src/main/java/org/apache/camel/component/aws2/timestream/write/Timestream2WriteProducer.java
@@ -714,8 +714,8 @@ public class Timestream2WriteProducer extends 
DefaultProducer {
                 builder.tableName(tableName);
             }
             if 
(ObjectHelper.isNotEmpty(exchange.getIn().getHeader(Timestream2Constants.RECORD)))
 {
-                Record record = 
exchange.getIn().getHeader(Timestream2Constants.RECORD, Record.class);
-                builder.commonAttributes(record);
+                Record recordObject = 
exchange.getIn().getHeader(Timestream2Constants.RECORD, Record.class);
+                builder.commonAttributes(recordObject);
             }
             if 
(ObjectHelper.isNotEmpty(exchange.getIn().getHeader(Timestream2Constants.RECORD_LIST)))
 {
                 List<Record> records = 
exchange.getIn().getHeader(Timestream2Constants.RECORD_LIST, List.class);
diff --git 
a/components/camel-azure/camel-azure-cosmosdb/src/main/java/org/apache/camel/component/azure/cosmosdb/CosmosDbConsumer.java
 
b/components/camel-azure/camel-azure-cosmosdb/src/main/java/org/apache/camel/component/azure/cosmosdb/CosmosDbConsumer.java
index 91e67820f67..054e7ede630 100644
--- 
a/components/camel-azure/camel-azure-cosmosdb/src/main/java/org/apache/camel/component/azure/cosmosdb/CosmosDbConsumer.java
+++ 
b/components/camel-azure/camel-azure-cosmosdb/src/main/java/org/apache/camel/component/azure/cosmosdb/CosmosDbConsumer.java
@@ -82,8 +82,8 @@ public class CosmosDbConsumer extends DefaultConsumer {
         return (CosmosDbEndpoint) super.getEndpoint();
     }
 
-    private void onEventListener(final List<Map<String, ?>> record) {
-        final Exchange exchange = createAzureCosmosDbExchange(record);
+    private void onEventListener(final List<Map<String, ?>> recordList) {
+        final Exchange exchange = createAzureCosmosDbExchange(recordList);
 
         // add exchange callback
         exchange.getExchangeExtension().addOnCompletion(onCompletion);
@@ -91,11 +91,11 @@ public class CosmosDbConsumer extends DefaultConsumer {
         getAsyncProcessor().process(exchange, EmptyAsyncCallback.get());
     }
 
-    private Exchange createAzureCosmosDbExchange(final List<Map<String, ?>> 
record) {
+    private Exchange createAzureCosmosDbExchange(final List<Map<String, ?>> 
recordList) {
         final Exchange exchange = createExchange(true);
         final Message message = exchange.getIn();
 
-        message.setBody(record);
+        message.setBody(recordList);
 
         return exchange;
     }
diff --git 
a/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyCsvFactory.java
 
b/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyCsvFactory.java
index a24adf46cda..5d12684e2f7 100644
--- 
a/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyCsvFactory.java
+++ 
b/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyCsvFactory.java
@@ -627,71 +627,71 @@ public class BindyCsvFactory extends BindyAbstractFactory 
implements BindyFactor
             for (Class<?> cl : models) {
 
                 // Get annotation @CsvRecord from the class
-                CsvRecord record = cl.getAnnotation(CsvRecord.class);
+                CsvRecord csvRecord = cl.getAnnotation(CsvRecord.class);
 
                 // Get annotation @Section from the class
                 Section section = cl.getAnnotation(Section.class);
 
-                if (record != null) {
-                    LOG.debug("Csv record: {}", record);
+                if (csvRecord != null) {
+                    LOG.debug("Csv record: {}", csvRecord);
 
                     // Get skipFirstLine parameter
-                    skipFirstLine = record.skipFirstLine();
+                    skipFirstLine = csvRecord.skipFirstLine();
                     LOG.debug("Skip First Line parameter of the CSV: {}", 
skipFirstLine);
 
                     // Get skipFirstLine parameter
-                    skipField = record.skipField();
+                    skipField = csvRecord.skipField();
                     LOG.debug("Skip Field parameter of the CSV: {}", 
skipField);
 
                     // Get generateHeaderColumnNames parameter
-                    generateHeaderColumnNames = record.generateHeaderColumns();
+                    generateHeaderColumnNames = 
csvRecord.generateHeaderColumns();
                     LOG.debug("Generate header column names parameter of the 
CSV: {}", generateHeaderColumnNames);
 
                     // Get Separator parameter
-                    
org.apache.camel.util.ObjectHelper.notNull(record.separator(),
+                    
org.apache.camel.util.ObjectHelper.notNull(csvRecord.separator(),
                             "No separator has been defined in the @Record 
annotation");
-                    separator = record.separator();
+                    separator = csvRecord.separator();
                     LOG.debug("Separator defined for the CSV: {}", separator);
 
                     // Get carriage return parameter
-                    crlf = record.crlf();
+                    crlf = csvRecord.crlf();
                     LOG.debug("Carriage return defined for the CSV: {}", crlf);
 
                     // Get isOrdered parameter
-                    messageOrdered = record.isOrdered();
+                    messageOrdered = csvRecord.isOrdered();
                     LOG.debug("Must CSV record be ordered: {}", 
messageOrdered);
 
-                    if 
(org.apache.camel.util.ObjectHelper.isNotEmpty(record.quote())) {
-                        quote = record.quote();
+                    if 
(org.apache.camel.util.ObjectHelper.isNotEmpty(csvRecord.quote())) {
+                        quote = csvRecord.quote();
                         LOG.debug("Quoting columns with: {}", quote);
                     }
 
-                    quoting = record.quoting();
+                    quoting = csvRecord.quoting();
                     LOG.debug("CSV will be quoted: {}", quoting);
 
-                    autospanLine = record.autospanLine();
+                    autospanLine = csvRecord.autospanLine();
                     LOG.debug("Autospan line in last record: {}", 
autospanLine);
 
                     // Get allowEmptyStream parameter
-                    allowEmptyStream = record.allowEmptyStream();
+                    allowEmptyStream = csvRecord.allowEmptyStream();
                     LOG.debug("Allow empty stream parameter of the CSV: {}", 
allowEmptyStream);
 
                     // Get quotingEscaped parameter
-                    quotingEscaped = record.quotingEscaped();
+                    quotingEscaped = csvRecord.quotingEscaped();
                     LOG.debug("Escape quote character flag of the CSV: {}", 
quotingEscaped);
 
                     // Get quotingOnlyWhenNeeded parameter
-                    quotingOnlyWhenNeeded = record.quotingOnlyWhenNeeded();
+                    quotingOnlyWhenNeeded = csvRecord.quotingOnlyWhenNeeded();
                     LOG.debug("Quoting only when needed: {}", 
quotingOnlyWhenNeeded);
 
                     // Get endWithLineBreak parameter
-                    endWithLineBreak = record.endWithLineBreak();
+                    endWithLineBreak = csvRecord.endWithLineBreak();
                     LOG.debug("End with line break: {}", endWithLineBreak);
 
-                    removeQuotes = record.removeQuotes();
+                    removeQuotes = csvRecord.removeQuotes();
                     LOG.debug("Remove quotes: {}", removeQuotes);
 
-                    trimLine = record.trimLine();
+                    trimLine = csvRecord.trimLine();
                     LOG.debug("Trim line: {}", trimLine);
                 }
 
diff --git 
a/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyFixedLengthFactory.java
 
b/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyFixedLengthFactory.java
index 88ce8298311..20f0048b1eb 100644
--- 
a/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyFixedLengthFactory.java
+++ 
b/components/camel-bindy/src/main/java/org/apache/camel/dataformat/bindy/BindyFixedLengthFactory.java
@@ -172,7 +172,7 @@ public class BindyFixedLengthFactory extends 
BindyAbstractFactory implements Bin
         String delimiter;
         Field field;
 
-        final UnicodeHelper record = new UnicodeHelper(
+        final UnicodeHelper unicodeHelper = new UnicodeHelper(
                 recordStr, (this.countGrapheme) ? 
UnicodeHelper.Method.GRAPHEME : UnicodeHelper.Method.CODEPOINTS);
 
         // Iterate through the list of positions
@@ -211,19 +211,19 @@ public class BindyFixedLengthFactory extends 
BindyAbstractFactory implements Bin
             }
 
             if (length > 0) {
-                if (record.length() < offset) {
+                if (unicodeHelper.length() < offset) {
                     token = "";
                 } else {
                     int endIndex = offset + length - 1;
-                    if (endIndex > record.length()) {
-                        endIndex = record.length();
+                    if (endIndex > unicodeHelper.length()) {
+                        endIndex = unicodeHelper.length();
                     }
-                    token = record.substring(offset - 1, endIndex);
+                    token = unicodeHelper.substring(offset - 1, endIndex);
                 }
                 offset += length;
             } else if (!delimiter.isEmpty()) {
                 final UnicodeHelper tempToken = new UnicodeHelper(
-                        record.substring(offset - 1, record.length()),
+                        unicodeHelper.substring(offset - 1, 
unicodeHelper.length()),
                         (this.countGrapheme) ? UnicodeHelper.Method.GRAPHEME : 
UnicodeHelper.Method.CODEPOINTS);
                 token = tempToken.substring(0, tempToken.indexOf(delimiter));
                 // include the delimiter in the offset calculation
@@ -324,7 +324,7 @@ public class BindyFixedLengthFactory extends 
BindyAbstractFactory implements Bin
         }
 
         // check for unmapped non-whitespace data at the end of the line
-        if (offset <= record.length() && !(record.substring(offset - 1, 
record.length())).isBlank()
+        if (offset <= unicodeHelper.length() && 
!(unicodeHelper.substring(offset - 1, unicodeHelper.length())).isBlank()
                 && !isIgnoreTrailingChars()) {
             throw new IllegalArgumentException(
                     "Unexpected / unmapped characters found at the end of the 
fixed-length record at line : " + line);
@@ -566,36 +566,36 @@ public class BindyFixedLengthFactory extends 
BindyAbstractFactory implements Bin
         for (Class<?> cl : models) {
 
             // Get annotation @FixedLengthRecord from the class
-            FixedLengthRecord record = 
cl.getAnnotation(FixedLengthRecord.class);
+            FixedLengthRecord fixedLengthRecord = 
cl.getAnnotation(FixedLengthRecord.class);
 
-            if (record != null) {
-                LOG.debug("Fixed length record: {}", record);
+            if (fixedLengthRecord != null) {
+                LOG.debug("Fixed length record: {}", fixedLengthRecord);
 
                 // Get carriage return parameter
-                crlf = record.crlf();
+                crlf = fixedLengthRecord.crlf();
                 LOG.debug("Carriage return defined for the CSV: {}", crlf);
 
-                eol = record.eol();
+                eol = fixedLengthRecord.eol();
                 LOG.debug("EOL(end-of-line) defined for the CSV: {}", eol);
 
                 // Get header parameter
-                header = record.header();
+                header = fixedLengthRecord.header();
                 LOG.debug("Header: {}", header);
                 hasHeader = header != void.class;
                 LOG.debug("Has Header: {}", hasHeader);
 
                 // Get skipHeader parameter
-                skipHeader = record.skipHeader();
+                skipHeader = fixedLengthRecord.skipHeader();
                 LOG.debug("Skip Header: {}", skipHeader);
 
                 // Get footer parameter
-                footer = record.footer();
+                footer = fixedLengthRecord.footer();
                 LOG.debug("Footer: {}", footer);
-                hasFooter = record.footer() != void.class;
+                hasFooter = fixedLengthRecord.footer() != void.class;
                 LOG.debug("Has Footer: {}", hasFooter);
 
                 // Get skipFooter parameter
-                skipFooter = record.skipFooter();
+                skipFooter = fixedLengthRecord.skipFooter();
                 LOG.debug("Skip Footer: {}", skipFooter);
 
                 // Get isHeader parameter
@@ -607,21 +607,21 @@ public class BindyFixedLengthFactory extends 
BindyAbstractFactory implements Bin
                 LOG.debug("Is Footer: {}", isFooter);
 
                 // Get padding character
-                paddingChar = record.paddingChar();
+                paddingChar = fixedLengthRecord.paddingChar();
                 LOG.debug("Padding char: {}", paddingChar);
 
                 // Get length of the record
-                recordLength = record.length();
+                recordLength = fixedLengthRecord.length();
                 LOG.debug("Length of the record: {}", recordLength);
 
                 // Get flag for ignore trailing characters
-                ignoreTrailingChars = record.ignoreTrailingChars();
+                ignoreTrailingChars = fixedLengthRecord.ignoreTrailingChars();
                 LOG.debug("Ignore trailing chars: {}", ignoreTrailingChars);
 
-                ignoreMissingChars = record.ignoreMissingChars();
+                ignoreMissingChars = fixedLengthRecord.ignoreMissingChars();
                 LOG.debug("Enable ignore missing chars: {}", 
ignoreMissingChars);
 
-                countGrapheme = record.countGrapheme();
+                countGrapheme = fixedLengthRecord.countGrapheme();
                 LOG.debug("Enable grapheme counting instead of codepoints: 
{}", countGrapheme);
             }
         }
diff --git 
a/components/camel-braintree/src/main/java/org/apache/camel/component/braintree/internal/BraintreeLogHandler.java
 
b/components/camel-braintree/src/main/java/org/apache/camel/component/braintree/internal/BraintreeLogHandler.java
index b0b3fdec554..5fcedcd2fbd 100644
--- 
a/components/camel-braintree/src/main/java/org/apache/camel/component/braintree/internal/BraintreeLogHandler.java
+++ 
b/components/camel-braintree/src/main/java/org/apache/camel/component/braintree/internal/BraintreeLogHandler.java
@@ -50,46 +50,46 @@ public final class BraintreeLogHandler extends Handler {
     }
 
     @Override
-    public void publish(LogRecord record) {
-        if (record != null) {
-            final Logger logger = getLogger(record);
-            final String message = formatMessage(record);
+    public void publish(LogRecord logRecord) {
+        if (logRecord != null) {
+            final Logger logger = getLogger(logRecord);
+            final String message = formatMessage(logRecord);
 
-            int level = record.getLevel().intValue();
+            int level = logRecord.getLevel().intValue();
             if (level <= TRACE_LEVEL_THRESHOLD) {
-                logger.trace(message, record.getThrown());
+                logger.trace(message, logRecord.getThrown());
             } else if (level <= DEBUG_LEVEL_THRESHOLD) {
-                logger.debug(message, record.getThrown());
+                logger.debug(message, logRecord.getThrown());
             } else if (level <= INFO_LEVEL_THRESHOLD) {
-                logger.info(message, record.getThrown());
+                logger.info(message, logRecord.getThrown());
             } else if (level <= WARN_LEVEL_THRESHOLD) {
-                logger.warn(message, record.getThrown());
+                logger.warn(message, logRecord.getThrown());
             } else {
-                logger.error(message, record.getThrown());
+                logger.error(message, logRecord.getThrown());
             }
         }
     }
 
-    private Logger getLogger(LogRecord record) {
-        String name = record.getLoggerName();
+    private Logger getLogger(LogRecord logRecord) {
+        String name = logRecord.getLoggerName();
         if (name == null) {
             name = DEFAULT_LOGGER_NAME;
         }
         return LoggerFactory.getLogger(name);
     }
 
-    private String formatMessage(LogRecord record) {
-        String message = record.getMessage();
+    private String formatMessage(LogRecord logRecord) {
+        String message = logRecord.getMessage();
         if (message != null) {
 
-            ResourceBundle bundle = record.getResourceBundle();
+            ResourceBundle bundle = logRecord.getResourceBundle();
             if (bundle != null) {
                 try {
                     message = bundle.getString(message);
                 } catch (MissingResourceException e) {
                 }
             }
-            Object[] params = record.getParameters();
+            Object[] params = logRecord.getParameters();
             // avoid formatting when there are no or 0 parameters. see also
             // http://jira.qos.ch/browse/SLF4J-203
             if (params != null && params.length > 0) {
diff --git 
a/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverter.java
 
b/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverter.java
index acaf80d9d90..6f03d48da20 100644
--- 
a/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverter.java
+++ 
b/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverter.java
@@ -31,8 +31,8 @@ public interface CsvRecordConverter<T> {
     /**
      * Converts the CSV record into another type.
      *
-     * @param  record CSV record to convert
+     * @param  csvRecord CSV record to convert
      * @return        converted CSV record
      */
-    T convertRecord(CSVRecord record);
+    T convertRecord(CSVRecord csvRecord);
 }
diff --git 
a/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverters.java
 
b/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverters.java
index cb7cc41b127..4c9ec4d5b27 100644
--- 
a/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverters.java
+++ 
b/components/camel-csv/src/main/java/org/apache/camel/dataformat/csv/CsvRecordConverters.java
@@ -45,10 +45,10 @@ final class CsvRecordConverters {
         private static final ListCsvRecordConverter SINGLETON = new 
ListCsvRecordConverter();
 
         @Override
-        public List<String> convertRecord(CSVRecord record) {
-            List<String> answer = new ArrayList<>(record.size());
-            for (int i = 0; i < record.size(); i++) {
-                answer.add(record.get(i));
+        public List<String> convertRecord(CSVRecord csvRecord) {
+            List<String> answer = new ArrayList<>(csvRecord.size());
+            for (int i = 0; i < csvRecord.size(); i++) {
+                answer.add(csvRecord.get(i));
             }
             return answer;
         }
@@ -67,8 +67,8 @@ final class CsvRecordConverters {
         private static final MapCsvRecordConverter SINGLETON = new 
MapCsvRecordConverter();
 
         @Override
-        public Map<String, String> convertRecord(CSVRecord record) {
-            return record.toMap();
+        public Map<String, String> convertRecord(CSVRecord csvRecord) {
+            return csvRecord.toMap();
         }
     }
 
@@ -86,8 +86,8 @@ final class CsvRecordConverters {
 
         @Override
         @SuppressWarnings("unchecked")
-        public Map<String, String> convertRecord(CSVRecord record) {
-            return record.toMap();
+        public Map<String, String> convertRecord(CSVRecord csvRecord) {
+            return csvRecord.toMap();
         }
     }
 }
diff --git 
a/components/camel-debezium/camel-debezium-common/camel-debezium-common-component/src/main/java/org/apache/camel/component/debezium/DebeziumEndpoint.java
 
b/components/camel-debezium/camel-debezium-common/camel-debezium-common-component/src/main/java/org/apache/camel/component/debezium/DebeziumEndpoint.java
index 311944f0d4a..9ca582f2ab2 100644
--- 
a/components/camel-debezium/camel-debezium-common/camel-debezium-common-component/src/main/java/org/apache/camel/component/debezium/DebeziumEndpoint.java
+++ 
b/components/camel-debezium/camel-debezium-common/camel-debezium-common-component/src/main/java/org/apache/camel/component/debezium/DebeziumEndpoint.java
@@ -64,7 +64,7 @@ public abstract class DebeziumEndpoint<C extends 
EmbeddedDebeziumConfiguration>
                 "DebeziumConsumer");
     }
 
-    public Exchange createDbzExchange(DebeziumConsumer consumer, final 
SourceRecord record) {
+    public Exchange createDbzExchange(DebeziumConsumer consumer, final 
SourceRecord sourceRecord) {
         final Exchange exchange;
         if (consumer != null) {
             exchange = consumer.createExchange(false);
@@ -74,8 +74,8 @@ public abstract class DebeziumEndpoint<C extends 
EmbeddedDebeziumConfiguration>
 
         final Message message = exchange.getIn();
 
-        final Schema valueSchema = record.valueSchema();
-        final Object value = record.value();
+        final Schema valueSchema = sourceRecord.valueSchema();
+        final Object value = sourceRecord.value();
 
         // extract values from SourceRecord
         final Map<String, Object> sourceMetadata = 
extractSourceMetadataValueFromValueStruct(valueSchema, value);
@@ -85,8 +85,8 @@ public abstract class DebeziumEndpoint<C extends 
EmbeddedDebeziumConfiguration>
         final Object timestamp = extractValueFromValueStruct(valueSchema, 
value, Envelope.FieldName.TIMESTAMP);
         final Object ddl = extractValueFromValueStruct(valueSchema, value, 
HistoryRecord.Fields.DDL_STATEMENTS);
         // set message headers
-        message.setHeader(DebeziumConstants.HEADER_IDENTIFIER, record.topic());
-        message.setHeader(DebeziumConstants.HEADER_KEY, record.key());
+        message.setHeader(DebeziumConstants.HEADER_IDENTIFIER, 
sourceRecord.topic());
+        message.setHeader(DebeziumConstants.HEADER_KEY, sourceRecord.key());
         message.setHeader(DebeziumConstants.HEADER_SOURCE_METADATA, 
sourceMetadata);
         message.setHeader(DebeziumConstants.HEADER_OPERATION, operation);
         message.setHeader(DebeziumConstants.HEADER_BEFORE, before);
diff --git 
a/components/camel-dns/src/main/java/org/apache/camel/component/dns/cloud/DnsServiceDiscovery.java
 
b/components/camel-dns/src/main/java/org/apache/camel/component/dns/cloud/DnsServiceDiscovery.java
index 735a8e4b972..022bfe4e2a3 100644
--- 
a/components/camel-dns/src/main/java/org/apache/camel/component/dns/cloud/DnsServiceDiscovery.java
+++ 
b/components/camel-dns/src/main/java/org/apache/camel/component/dns/cloud/DnsServiceDiscovery.java
@@ -59,7 +59,7 @@ public final class DnsServiceDiscovery extends 
DefaultServiceDiscovery {
                     .filter(SRVRecord.class::isInstance)
                     .map(SRVRecord.class::cast)
                     .sorted(COMPARATOR)
-                    .map(record -> asService(name, record))
+                    .map(srvRecord -> asService(name, srvRecord))
                     .collect(Collectors.toList());
         } else {
             services = Collections.emptyList();
@@ -85,15 +85,15 @@ public final class DnsServiceDiscovery extends 
DefaultServiceDiscovery {
         return byPriority.thenComparing(byWeight);
     }
 
-    private static ServiceDefinition asService(String serviceName, SRVRecord 
record) {
+    private static ServiceDefinition asService(String serviceName, SRVRecord 
srvRecord) {
         Map<String, String> meta = new HashMap<>();
-        ObjectHelper.ifNotEmpty(record.getPriority(), val -> 
meta.put("priority", Integer.toString(val)));
-        ObjectHelper.ifNotEmpty(record.getWeight(), val -> meta.put("weight", 
Integer.toString(val)));
+        ObjectHelper.ifNotEmpty(srvRecord.getPriority(), val -> 
meta.put("priority", Integer.toString(val)));
+        ObjectHelper.ifNotEmpty(srvRecord.getWeight(), val -> 
meta.put("weight", Integer.toString(val)));
 
         return new DefaultServiceDefinition(
                 serviceName,
-                record.getTarget().toString(true),
-                record.getPort(),
+                srvRecord.getTarget().toString(true),
+                srvRecord.getPort(),
                 meta);
     }
 }
diff --git 
a/components/camel-dns/src/main/java/org/apache/camel/component/dns/types/DnsConverter.java
 
b/components/camel-dns/src/main/java/org/apache/camel/component/dns/types/DnsConverter.java
index 4876a68cac9..4cc0f9f0883 100644
--- 
a/components/camel-dns/src/main/java/org/apache/camel/component/dns/types/DnsConverter.java
+++ 
b/components/camel-dns/src/main/java/org/apache/camel/component/dns/types/DnsConverter.java
@@ -36,8 +36,8 @@ public final class DnsConverter {
     }
 
     @Converter
-    public static String toString(Record record) {
-        return record.toString();
+    public static String toString(Record recordObject) {
+        return recordObject.toString();
     }
 
     @Converter
diff --git 
a/components/camel-flatpack/src/main/java/org/apache/camel/component/flatpack/FlatpackConverter.java
 
b/components/camel-flatpack/src/main/java/org/apache/camel/component/flatpack/FlatpackConverter.java
index 12ff3c4b7d9..6942677ee2b 100644
--- 
a/components/camel-flatpack/src/main/java/org/apache/camel/component/flatpack/FlatpackConverter.java
+++ 
b/components/camel-flatpack/src/main/java/org/apache/camel/component/flatpack/FlatpackConverter.java
@@ -40,12 +40,12 @@ public final class FlatpackConverter {
     }
 
     @Converter
-    public static Map<String, Object> toMap(Record record) {
+    public static Map<String, Object> toMap(Record recordObj) {
         Map<String, Object> map = new HashMap<>();
-        if (record instanceof DataSet dataSet) {
+        if (recordObj instanceof DataSet dataSet) {
             putValues(map, dataSet);
         } else {
-            putValues(map, record);
+            putValues(map, recordObj);
         }
 
         return map;
@@ -110,23 +110,23 @@ public final class FlatpackConverter {
     /**
      * Puts the values of the record into the map
      */
-    private static void putValues(Map<String, Object> map, Record record) {
-        String[] columns = record.getColumns();
+    private static void putValues(Map<String, Object> map, Record recordObj) {
+        String[] columns = recordObj.getColumns();
 
         for (String column : columns) {
-            String value = record.getString(column);
+            String value = recordObj.getString(column);
             map.put(column, value);
         }
     }
 
     private static Element createDatasetRecord(DataSet dataSet, Document doc) {
-        Element record;
+        Element element;
         if (dataSet.isRecordID(FlatpackComponent.HEADER_ID)) {
-            record = doc.createElement("DatasetHeader");
+            element = doc.createElement("DatasetHeader");
         } else if (dataSet.isRecordID(FlatpackComponent.TRAILER_ID)) {
-            record = doc.createElement("DatasetTrailer");
+            element = doc.createElement("DatasetTrailer");
         } else {
-            record = doc.createElement("DatasetRecord");
+            element = doc.createElement("DatasetRecord");
         }
 
         String[] columns = getColumns(dataSet);
@@ -138,10 +138,10 @@ public final class FlatpackConverter {
             columnElement.setAttribute("name", column);
             columnElement.setTextContent(value);
 
-            record.appendChild(columnElement);
+            element.appendChild(columnElement);
         }
 
-        return record;
+        return element;
     }
 
     private static String[] getColumns(DataSet dataSet) {
diff --git 
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/InfluxDb2Producer.java
 
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/InfluxDb2Producer.java
index 7a8abfc75f0..417414a552d 100644
--- 
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/InfluxDb2Producer.java
+++ 
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/InfluxDb2Producer.java
@@ -120,12 +120,12 @@ public class InfluxDb2Producer extends DefaultProducer {
 
     private void insertRecord(Exchange exchange, String orgName, String 
bucketName, WritePrecision writePrecision)
             throws InvalidPayloadException {
-        Record record = exchange.getIn().getMandatoryBody(Record.class);
+        Record recordObj = exchange.getIn().getMandatoryBody(Record.class);
         try {
             if (LOG.isDebugEnabled()) {
-                LOG.debug("Writing record {}", record);
+                LOG.debug("Writing record {}", recordObj);
             }
-            writeApi.writeRecord(bucketName, orgName, writePrecision, 
record.getInfluxRecord());
+            writeApi.writeRecord(bucketName, orgName, writePrecision, 
recordObj.getInfluxRecord());
         } catch (Exception ex) {
             exchange.setException(new CamelInfluxDb2Exception(ex));
         }
diff --git 
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Record.java
 
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Record.java
index 89ef913655f..ab4589c9990 100644
--- 
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Record.java
+++ 
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Record.java
@@ -21,10 +21,10 @@ package org.apache.camel.component.influxdb2.data;
  * Influxdb write Record
  */
 public class Record {
-    private String record;
+    private String recordObj;
 
-    public Record(String record) {
-        this.record = record;
+    public Record(String recordObj) {
+        this.recordObj = recordObj;
     }
 
     public static Record fromString(String record) {
@@ -32,10 +32,10 @@ public class Record {
     }
 
     public String getInfluxRecord() {
-        return record;
+        return recordObj;
     }
 
-    public void setRecord(String record) {
-        this.record = record;
+    public void setRecord(String recordObj) {
+        this.recordObj = recordObj;
     }
 }
diff --git 
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Records.java
 
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Records.java
index 8f4ffb2380b..dc4f6ee6d4b 100644
--- 
a/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Records.java
+++ 
b/components/camel-influxdb2/src/main/java/org/apache/camel/component/influxdb2/data/Records.java
@@ -53,21 +53,21 @@ public class Records {
         return new Records(records);
     }
 
-    public static Records create(Record record) {
-        return new Records().addRecord(record);
+    public static Records create(Record recordObj) {
+        return new Records().addRecord(recordObj);
     }
 
-    public static Records create(String record) {
-        return create(Record.fromString(record));
+    public static Records create(String recordObj) {
+        return create(Record.fromString(recordObj));
     }
 
-    public Records addRecord(String record) {
-        this.records.add(Record.fromString(record));
+    public Records addRecord(String recordObj) {
+        this.records.add(Record.fromString(recordObj));
         return this;
     }
 
-    public Records addRecord(Record record) {
-        this.records.add(record);
+    public Records addRecord(Record recordObj) {
+        this.records.add(recordObj);
         return this;
     }
 
diff --git 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AbstractCommitManager.java
 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AbstractCommitManager.java
index d35ac3d65c3..1e7638a80fa 100644
--- 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AbstractCommitManager.java
+++ 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AbstractCommitManager.java
@@ -71,14 +71,14 @@ public abstract class AbstractCommitManager implements 
CommitManager {
 
     @Override
     public KafkaManualCommit getManualCommit(
-            Exchange exchange, TopicPartition partition, 
ConsumerRecord<Object, Object> record) {
+            Exchange exchange, TopicPartition partition, 
ConsumerRecord<Object, Object> consumerRecord) {
 
         KafkaManualCommitFactory manualCommitFactory = 
kafkaConsumer.getEndpoint().getKafkaManualCommitFactory();
         if (manualCommitFactory == null) {
             manualCommitFactory = new DefaultKafkaManualCommitFactory();
         }
 
-        return getManualCommit(exchange, partition, record, 
manualCommitFactory);
+        return getManualCommit(exchange, partition, consumerRecord, 
manualCommitFactory);
     }
 
     @Override
diff --git 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AsyncCommitManager.java
 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AsyncCommitManager.java
index 7e117ee4d61..b7f776ac56b 100644
--- 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AsyncCommitManager.java
+++ 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/AsyncCommitManager.java
@@ -75,14 +75,14 @@ public class AsyncCommitManager extends 
AbstractCommitManager {
 
     @Override
     public KafkaManualCommit getManualCommit(
-            Exchange exchange, TopicPartition partition, 
ConsumerRecord<Object, Object> record) {
+            Exchange exchange, TopicPartition partition, 
ConsumerRecord<Object, Object> consumerRecord) {
 
         KafkaManualCommitFactory manualCommitFactory = 
kafkaConsumer.getEndpoint().getKafkaManualCommitFactory();
         if (manualCommitFactory == null) {
             manualCommitFactory = new DefaultKafkaManualAsyncCommitFactory();
         }
 
-        return getManualCommit(exchange, partition, record, 
manualCommitFactory);
+        return getManualCommit(exchange, partition, consumerRecord, 
manualCommitFactory);
     }
 
     @Override
diff --git 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/CommitManager.java
 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/CommitManager.java
index 999fde26563..c852f6a5fd2 100644
--- 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/CommitManager.java
+++ 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/CommitManager.java
@@ -23,7 +23,7 @@ import org.apache.kafka.common.TopicPartition;
 
 public interface CommitManager {
 
-    KafkaManualCommit getManualCommit(Exchange exchange, TopicPartition 
partition, ConsumerRecord<Object, Object> record);
+    KafkaManualCommit getManualCommit(Exchange exchange, TopicPartition 
partition, ConsumerRecord<Object, Object> consumerRecord);
 
     /**
      * Commits everything that has been cached
diff --git 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessor.java
 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessor.java
index 86af7396d87..2954836f5f6 100644
--- 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessor.java
+++ 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessor.java
@@ -128,17 +128,17 @@ public class KafkaRecordProcessor {
 
     private boolean processException(
             Exchange exchange, TopicPartition topicPartition,
-            ConsumerRecord<Object, Object> record, ProcessingResult lastResult,
+            ConsumerRecord<Object, Object> consumerRecord, ProcessingResult 
lastResult,
             ExceptionHandler exceptionHandler) {
 
         // processing failed due to an unhandled exception, what should we do
         if (configuration.isBreakOnFirstError()) {
             if (lastResult.getPartition() != -1 &&
-                    lastResult.getPartition() != record.partition()) {
+                    lastResult.getPartition() != consumerRecord.partition()) {
                 LOG.error("About to process an exception with UNEXPECTED 
partition & offset. Got topic partition {}. " +
                           " The last result was on partition {} with offset {} 
but was expecting partition {} with offset {}",
                         topicPartition.partition(), lastResult.getPartition(), 
lastResult.getPartitionLastOffset(),
-                        record.partition(), record.offset());
+                        consumerRecord.partition(), consumerRecord.offset());
             }
 
             // we are failing and we should break out
@@ -147,7 +147,7 @@ public class KafkaRecordProcessor {
                 LOG.warn("Error during processing {} from topic: {} due to 
{}", exchange, topicPartition.topic(),
                         exc.getMessage());
                 LOG.warn("Will seek consumer to offset {} on partition {} and 
start polling again.",
-                        record.offset(), record.partition());
+                        consumerRecord.offset(), consumerRecord.partition());
             }
 
             // force commit, so we resume on next poll where we failed
diff --git 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessorFacade.java
 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessorFacade.java
index 44573daa60d..5c7060400d7 100644
--- 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessorFacade.java
+++ 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/consumer/support/KafkaRecordProcessorFacade.java
@@ -77,16 +77,16 @@ public class KafkaRecordProcessorFacade {
             logRecordsInPartition(partitionRecords, partition);
 
             while (!lastResult.isBreakOnErrorHit() && recordIterator.hasNext() 
&& !isStopping()) {
-                ConsumerRecord<Object, Object> record = recordIterator.next();
+                ConsumerRecord<Object, Object> consumerRecord = 
recordIterator.next();
 
-                LOG.debug("Processing record on partition {} with offset {}", 
record.partition(), record.offset());
+                LOG.debug("Processing record on partition {} with offset {}", 
consumerRecord.partition(), consumerRecord.offset());
 
                 lastResult = processRecord(partition, 
partitionIterator.hasNext(), recordIterator.hasNext(), lastResult,
-                        kafkaRecordProcessor, record);
+                        kafkaRecordProcessor, consumerRecord);
 
                 LOG.debug(
                         "Processed record on partition {} with offset {} and 
got ProcessingResult for partition {} and offset {}",
-                        record.partition(), record.offset(), 
lastResult.getPartition(), lastResult.getPartitionLastOffset());
+                        consumerRecord.partition(), consumerRecord.offset(), 
lastResult.getPartition(), lastResult.getPartitionLastOffset());
 
                 if (consumerListener != null) {
                     if (!consumerListener.afterProcess(lastResult)) {
@@ -125,15 +125,15 @@ public class KafkaRecordProcessorFacade {
             boolean recordHasNext,
             final ProcessingResult lastResult,
             KafkaRecordProcessor kafkaRecordProcessor,
-            ConsumerRecord<Object, Object> record) {
+            ConsumerRecord<Object, Object> consumerRecord) {
 
-        logRecord(record);
+        logRecord(consumerRecord);
 
         Exchange exchange = camelKafkaConsumer.createExchange(false);
 
         ProcessingResult currentResult
                 = kafkaRecordProcessor.processExchange(exchange, partition, 
partitionHasNext,
-                        recordHasNext, record, lastResult, 
camelKafkaConsumer.getExceptionHandler());
+                        recordHasNext, consumerRecord, lastResult, 
camelKafkaConsumer.getExceptionHandler());
 
         if (!currentResult.isBreakOnErrorHit()) {
             commitManager.recordOffset(partition, 
currentResult.getPartitionLastOffset());
@@ -145,10 +145,10 @@ public class KafkaRecordProcessorFacade {
         return currentResult;
     }
 
-    private void logRecord(ConsumerRecord<Object, Object> record) {
+    private void logRecord(ConsumerRecord<Object, Object> consumerRecord) {
         if (LOG.isTraceEnabled()) {
-            LOG.trace("Partition = {}, offset = {}, key = {}, value = {}", 
record.partition(),
-                    record.offset(), record.key(), record.value());
+            LOG.trace("Partition = {}, offset = {}, key = {}, value = {}", 
consumerRecord.partition(),
+                    consumerRecord.offset(), consumerRecord.key(), 
consumerRecord.value());
         }
     }
 
diff --git 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/producer/support/KafkaProducerMetadataCallBack.java
 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/producer/support/KafkaProducerMetadataCallBack.java
index 37db4907049..053307a96b7 100644
--- 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/producer/support/KafkaProducerMetadataCallBack.java
+++ 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/producer/support/KafkaProducerMetadataCallBack.java
@@ -24,18 +24,18 @@ import static 
org.apache.camel.component.kafka.producer.support.ProducerUtil.set
 
 public class KafkaProducerMetadataCallBack implements Callback {
     private final Object body;
-    private final boolean record;
+    private final boolean recordMetadata;
 
-    public KafkaProducerMetadataCallBack(Object body, boolean record) {
+    public KafkaProducerMetadataCallBack(Object body, boolean recordMetadata) {
         this.body = body;
-        this.record = record;
+        this.recordMetadata = recordMetadata;
     }
 
     @Override
     public void onCompletion(RecordMetadata recordMetadata, Exception e) {
         setException(body, e);
 
-        if (record) {
+        if (this.recordMetadata) {
             setRecordMetadata(body, recordMetadata);
         }
     }
diff --git 
a/components/camel-kafka/src/main/java/org/apache/camel/processor/resume/kafka/SingleNodeKafkaResumeStrategy.java
 
b/components/camel-kafka/src/main/java/org/apache/camel/processor/resume/kafka/SingleNodeKafkaResumeStrategy.java
index 2f9113f2423..94be16f6258 100644
--- 
a/components/camel-kafka/src/main/java/org/apache/camel/processor/resume/kafka/SingleNodeKafkaResumeStrategy.java
+++ 
b/components/camel-kafka/src/main/java/org/apache/camel/processor/resume/kafka/SingleNodeKafkaResumeStrategy.java
@@ -108,9 +108,9 @@ public class SingleNodeKafkaResumeStrategy implements 
KafkaResumeStrategy, Camel
      *
      */
     protected void produce(byte[] key, byte[] message, UpdateCallBack 
updateCallBack) {
-        ProducerRecord<byte[], byte[]> record = new 
ProducerRecord<>(resumeStrategyConfiguration.getTopic(), key, message);
+        ProducerRecord<byte[], byte[]> producerRecord = new 
ProducerRecord<>(resumeStrategyConfiguration.getTopic(), key, message);
 
-        producer.send(record, (recordMetadata, e) -> {
+        producer.send(producerRecord, (recordMetadata, e) -> {
             if (e != null) {
                 LOG.error("Failed to send message {}", e.getMessage(), e);
             }
@@ -235,15 +235,15 @@ public class SingleNodeKafkaResumeStrategy implements 
KafkaResumeStrategy, Camel
         do {
             ConsumerRecords<byte[], byte[]> records = consume(consumer);
 
-            for (ConsumerRecord<byte[], byte[]> record : records) {
-                byte[] value = record.value();
+            for (ConsumerRecord<byte[], byte[]> consumerRecord : records) {
+                byte[] value = consumerRecord.value();
 
                 if (LOG.isTraceEnabled()) {
-                    LOG.trace("Read from Kafka at {} ({}): {}", 
Instant.ofEpochMilli(record.timestamp()),
-                            record.timestampType(), value);
+                    LOG.trace("Read from Kafka at {} ({}): {}", 
Instant.ofEpochMilli(consumerRecord.timestamp()),
+                            consumerRecord.timestampType(), value);
                 }
 
-                if (!deserializable.deserialize(ByteBuffer.wrap(record.key()), 
ByteBuffer.wrap(record.value()))) {
+                if 
(!deserializable.deserialize(ByteBuffer.wrap(consumerRecord.key()), 
ByteBuffer.wrap(consumerRecord.value()))) {
                     LOG.warn("Deserializer indicates that this is the last 
record to deserialize");
                 }
             }
diff --git 
a/components/camel-kubernetes/src/main/java/org/apache/camel/component/kubernetes/cloud/KubernetesDnsSrvServiceDiscovery.java
 
b/components/camel-kubernetes/src/main/java/org/apache/camel/component/kubernetes/cloud/KubernetesDnsSrvServiceDiscovery.java
index c3727547588..a4c203e9442 100644
--- 
a/components/camel-kubernetes/src/main/java/org/apache/camel/component/kubernetes/cloud/KubernetesDnsSrvServiceDiscovery.java
+++ 
b/components/camel-kubernetes/src/main/java/org/apache/camel/component/kubernetes/cloud/KubernetesDnsSrvServiceDiscovery.java
@@ -100,8 +100,8 @@ public class KubernetesDnsSrvServiceDiscovery extends 
KubernetesServiceDiscovery
                 List<ServiceDefinition> servers = new LinkedList<>();
 
                 while (resolved.hasMore()) {
-                    String record = (String) resolved.next();
-                    String[] items = record.split(" ", -1);
+                    String recordObj = (String) resolved.next();
+                    String[] items = recordObj.split(" ", -1);
                     String host = items[3].trim();
                     String port = items[2].trim();
 
diff --git 
a/components/camel-openapi-java/src/main/java/org/apache/camel/openapi/RestOpenApiSupport.java
 
b/components/camel-openapi-java/src/main/java/org/apache/camel/openapi/RestOpenApiSupport.java
index 764b5a12e65..6ec28f1f065 100644
--- 
a/components/camel-openapi-java/src/main/java/org/apache/camel/openapi/RestOpenApiSupport.java
+++ 
b/components/camel-openapi-java/src/main/java/org/apache/camel/openapi/RestOpenApiSupport.java
@@ -182,10 +182,10 @@ public class RestOpenApiSupport {
         Pattern p = Pattern.compile("\\{(.*?)}");
         Matcher m = p.matcher(url);
         while (m.find()) {
-            String var = m.group(1);
-            if (server != null && server.getVariables() != null && 
server.getVariables().get(var) != null) {
-                String varValue = server.getVariables().get(var).getDefault();
-                url = url.replace("{" + var + "}", varValue);
+            String variable = m.group(1);
+            if (server != null && server.getVariables() != null && 
server.getVariables().get(variable) != null) {
+                String varValue = 
server.getVariables().get(variable).getDefault();
+                url = url.replace("{" + variable + "}", varValue);
             }
         }
         return url;
diff --git 
a/components/camel-rest-openapi/src/main/java/org/apache/camel/component/rest/openapi/RestOpenApiEndpoint.java
 
b/components/camel-rest-openapi/src/main/java/org/apache/camel/component/rest/openapi/RestOpenApiEndpoint.java
index 67bfdbf219d..b6b20d75891 100644
--- 
a/components/camel-rest-openapi/src/main/java/org/apache/camel/component/rest/openapi/RestOpenApiEndpoint.java
+++ 
b/components/camel-rest-openapi/src/main/java/org/apache/camel/component/rest/openapi/RestOpenApiEndpoint.java
@@ -421,10 +421,10 @@ public final class RestOpenApiEndpoint extends 
DefaultEndpoint {
         Matcher m = p.matcher(url);
         while (m.find()) {
 
-            String var = m.group(1);
-            if (server != null && server.getVariables() != null && 
server.getVariables().get(var) != null) {
-                String varValue = server.getVariables().get(var).getDefault();
-                url = url.replace("{" + var + "}", varValue);
+            String variable = m.group(1);
+            if (server != null && server.getVariables() != null && 
server.getVariables().get(variable) != null) {
+                String varValue = 
server.getVariables().get(variable).getDefault();
+                url = url.replace("{" + variable + "}", varValue);
             }
         }
         return url;
diff --git 
a/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/PubSubApiConsumer.java
 
b/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/PubSubApiConsumer.java
index 66b9b15e6bf..e71d3eec75c 100644
--- 
a/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/PubSubApiConsumer.java
+++ 
b/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/PubSubApiConsumer.java
@@ -66,10 +66,10 @@ public class PubSubApiConsumer extends DefaultConsumer {
         }
     }
 
-    public void processEvent(Object record, String replayId) throws 
IOException {
+    public void processEvent(Object recordObj, String replayId) throws 
IOException {
         final Exchange exchange = createExchange(true);
         final Message in = exchange.getIn();
-        in.setBody(record);
+        in.setBody(recordObj);
         in.setHeader(HEADER_SALESFORCE_PUBSUB_REPLAY_ID, replayId);
         AsyncCallback cb = defaultConsumerCallback(exchange, true);
         getAsyncProcessor().process(exchange, cb);
diff --git 
a/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/internal/client/PubSubApiClient.java
 
b/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/internal/client/PubSubApiClient.java
index cb05734e191..d6caba9b317 100644
--- 
a/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/internal/client/PubSubApiClient.java
+++ 
b/components/camel-salesforce/camel-salesforce-component/src/main/java/org/apache/camel/component/salesforce/internal/client/PubSubApiClient.java
@@ -240,8 +240,8 @@ public class PubSubApiClient extends ServiceSupport {
         }
         byte[] bytes;
         if (body instanceof IndexedRecord indexedRecord) {
-            if (body instanceof GenericRecord record) {
-                bytes = getBytes(body, new 
GenericDatumWriter<>(record.getSchema()));
+            if (body instanceof GenericRecord genericRecord) {
+                bytes = getBytes(body, new 
GenericDatumWriter<>(genericRecord.getSchema()));
             } else if (body instanceof SpecificRecord) {
                 bytes = getBytes(body, new SpecificDatumWriter<>());
             } else {
@@ -367,7 +367,7 @@ public class PubSubApiClient extends ServiceSupport {
 
         private void processEvent(ConsumerEvent ce) throws IOException {
             final Schema schema = getSchema(ce.getEvent().getSchemaId());
-            Object record = switch (consumer.getDeserializeType()) {
+            Object recordObj = switch (consumer.getDeserializeType()) {
                 case AVRO -> deserializeAvro(ce, schema);
                 case GENERIC_RECORD -> deserializeGenericRecord(ce, schema);
                 case SPECIFIC_RECORD -> deserializeSpecificRecord(ce, schema);
@@ -375,7 +375,7 @@ public class PubSubApiClient extends ServiceSupport {
                 case JSON -> deserializeJson(ce, schema);
             };
             String replayId = 
PubSubApiClient.base64EncodeByteString(ce.getReplayId());
-            consumer.processEvent(record, replayId);
+            consumer.processEvent(recordObj, replayId);
         }
 
         private Object deserializeAvro(ConsumerEvent ce, Schema schema) throws 
IOException {
@@ -388,9 +388,9 @@ public class PubSubApiClient extends ServiceSupport {
         }
 
         private Object deserializeJson(ConsumerEvent ce, Schema schema) throws 
IOException {
-            final GenericRecord record = deserializeGenericRecord(ce, schema);
+            final GenericRecord genericRecord = deserializeGenericRecord(ce, 
schema);
             JsonAvroConverter converter = new JsonAvroConverter();
-            final byte[] bytes = converter.convertToJson(record);
+            final byte[] bytes = converter.convertToJson(genericRecord);
             return new String(bytes);
         }
 
diff --git 
a/core/camel-base-engine/src/main/java/org/apache/camel/impl/engine/RouteService.java
 
b/core/camel-base-engine/src/main/java/org/apache/camel/impl/engine/RouteService.java
index fd22706079d..98e5a3e6650 100644
--- 
a/core/camel-base-engine/src/main/java/org/apache/camel/impl/engine/RouteService.java
+++ 
b/core/camel-base-engine/src/main/java/org/apache/camel/impl/engine/RouteService.java
@@ -366,9 +366,9 @@ public class RouteService extends ChildServiceSupport {
         for (Service service : services) {
             StartupStep step = null;
             // skip internal services / route pipeline (starting point for 
route)
-            boolean record
+            boolean shouldRecord
                     = !(service instanceof InternalProcessor || 
"RoutePipeline".equals(service.getClass().getSimpleName()));
-            if (record) {
+            if (shouldRecord) {
                 step = beginStep(service, "Init");
             }
             ServiceHelper.initService(service);
@@ -384,9 +384,9 @@ public class RouteService extends ChildServiceSupport {
         for (Service service : services) {
             StartupStep step = null;
             // skip internal services / route pipeline (starting point for 
route)
-            boolean record
+            boolean shouldRecord
                     = !(service instanceof InternalProcessor || 
"RoutePipeline".equals(service.getClass().getSimpleName()));
-            if (record) {
+            if (shouldRecord) {
                 step = beginStep(service, "Start");
             }
             for (LifecycleStrategy strategy : 
camelContext.getLifecycleStrategies()) {
diff --git 
a/dsl/camel-kamelet-main/src/main/java/org/apache/camel/main/download/DownloadThreadPool.java
 
b/dsl/camel-kamelet-main/src/main/java/org/apache/camel/main/download/DownloadThreadPool.java
index e3bc60eba44..cf7dc6bf405 100644
--- 
a/dsl/camel-kamelet-main/src/main/java/org/apache/camel/main/download/DownloadThreadPool.java
+++ 
b/dsl/camel-kamelet-main/src/main/java/org/apache/camel/main/download/DownloadThreadPool.java
@@ -86,11 +86,11 @@ class DownloadThreadPool extends ServiceSupport implements 
CamelContextAware {
         }
 
         MavenGav a = MavenGav.parseGav(gav);
-        DownloadRecord record = downloader.getDownloadState(a.getGroupId(), 
a.getArtifactId(), a.getVersion());
-        if (record != null) {
+        DownloadRecord downloadRecord = 
downloader.getDownloadState(a.getGroupId(), a.getArtifactId(), a.getVersion());
+        if (downloadRecord != null) {
             long taken = watch.taken();
-            String url = record.repoUrl();
-            String id = record.repoId();
+            String url = downloadRecord.repoUrl();
+            String id = downloadRecord.repoId();
             String msg = "Downloaded: " + gav + " (took: "
                          + TimeUtils.printDuration(taken, true) + ") from: " + 
id + "@" + url;
             log.info(msg);

Reply via email to