This is an automated email from the ASF dual-hosted git repository.
acosentino pushed a commit to branch camel-master
in repository https://gitbox.apache.org/repos/asf/camel-kafka-connector.git
The following commit(s) were added to refs/heads/camel-master by this push:
new 087dfb5 [create-pull-request] automated change
087dfb5 is described below
commit 087dfb5612d786d259d8f30f8e638ab1ebdd1686
Author: github-actions[bot]
<41898282+github-actions[bot]@users.noreply.github.com>
AuthorDate: Sat Feb 13 03:39:39 2021 +0000
[create-pull-request] automated change
---
.../resources/connectors/camel-aws2-sqs-sink.json | 14 ++++++++++++++
.../generated/resources/connectors/camel-kafka-sink.json | 2 +-
.../generated/resources/connectors/camel-kafka-source.json | 2 +-
.../resources/connectors/camel-milo-client-sink.json | 6 ++++++
.../resources/connectors/camel-milo-client-source.json | 6 ++++++
.../resources/connectors/camel-vertx-kafka-sink.json | 6 ------
.../resources/connectors/camel-vertx-kafka-source.json | 6 ------
.../src/generated/resources/camel-aws2-sqs-sink.json | 14 ++++++++++++++
.../src/main/docs/camel-aws2-sqs-kafka-sink-connector.adoc | 4 +++-
.../aws2sqs/CamelAws2sqsSinkConnectorConfig.java | 8 ++++++++
.../src/generated/resources/camel-kafka-sink.json | 2 +-
.../src/generated/resources/camel-kafka-source.json | 2 +-
.../src/main/docs/camel-kafka-kafka-sink-connector.adoc | 2 +-
.../src/main/docs/camel-kafka-kafka-source-connector.adoc | 2 +-
.../kafka/CamelKafkaSinkConnectorConfig.java | 2 +-
.../kafka/CamelKafkaSourceConnectorConfig.java | 2 +-
.../src/generated/resources/camel-milo-client-sink.json | 6 ++++++
.../src/generated/resources/camel-milo-client-source.json | 6 ++++++
.../main/docs/camel-milo-client-kafka-sink-connector.adoc | 3 ++-
.../docs/camel-milo-client-kafka-source-connector.adoc | 3 ++-
.../miloclient/CamelMiloclientSinkConnectorConfig.java | 4 ++++
.../miloclient/CamelMiloclientSourceConnectorConfig.java | 4 ++++
.../src/generated/resources/camel-vertx-kafka-sink.json | 6 ------
.../src/generated/resources/camel-vertx-kafka-source.json | 6 ------
.../main/docs/camel-vertx-kafka-kafka-sink-connector.adoc | 3 +--
.../docs/camel-vertx-kafka-kafka-source-connector.adoc | 3 +--
.../vertxkafka/CamelVertxkafkaSinkConnectorConfig.java | 4 ----
.../vertxkafka/CamelVertxkafkaSourceConnectorConfig.java | 4 ----
.../connectors/camel-aws2-sqs-kafka-sink-connector.adoc | 4 +++-
.../pages/connectors/camel-kafka-kafka-sink-connector.adoc | 2 +-
.../connectors/camel-kafka-kafka-source-connector.adoc | 2 +-
.../connectors/camel-milo-client-kafka-sink-connector.adoc | 3 ++-
.../camel-milo-client-kafka-source-connector.adoc | 3 ++-
.../connectors/camel-vertx-kafka-kafka-sink-connector.adoc | 3 +--
.../camel-vertx-kafka-kafka-source-connector.adoc | 3 +--
35 files changed, 96 insertions(+), 56 deletions(-)
diff --git
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-aws2-sqs-sink.json
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-aws2-sqs-sink.json
index 8d185cb..be8e767 100644
---
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-aws2-sqs-sink.json
+++
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-aws2-sqs-sink.json
@@ -85,6 +85,13 @@
"priority": "MEDIUM",
"required": "false"
},
+ "camel.sink.endpoint.batchSeparator": {
+ "name": "camel.sink.endpoint.batchSeparator",
+ "description": "Set the separator when passing a String
to send batch message operation",
+ "defaultValue": "\",\"",
+ "priority": "MEDIUM",
+ "required": "false"
+ },
"camel.sink.endpoint.delaySeconds": {
"name": "camel.sink.endpoint.delaySeconds",
"description": "Delay sending messages for a number of
seconds.",
@@ -269,6 +276,13 @@
"priority": "MEDIUM",
"required": "false"
},
+ "camel.component.aws2-sqs.batchSeparator": {
+ "name": "camel.component.aws2-sqs.batchSeparator",
+ "description": "Set the separator when passing a String
to send batch message operation",
+ "defaultValue": "\",\"",
+ "priority": "MEDIUM",
+ "required": "false"
+ },
"camel.component.aws2-sqs.delaySeconds": {
"name": "camel.component.aws2-sqs.delaySeconds",
"description": "Delay sending messages for a number of
seconds.",
diff --git
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-kafka-sink.json
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-kafka-sink.json
index b997786..dfae88e 100644
---
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-kafka-sink.json
+++
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-kafka-sink.json
@@ -743,7 +743,7 @@
},
"camel.component.kafka.kafkaClientFactory": {
"name": "camel.component.kafka.kafkaClientFactory",
- "description": "Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances with logic that
extends the vanilla Kafka clients.",
+ "description": "Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create instances with logic that extends the
vanilla Kafka clients.",
"priority": "MEDIUM",
"required": "false"
},
diff --git
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-kafka-source.json
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-kafka-source.json
index ebfe2c8..7124ba4 100644
---
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-kafka-source.json
+++
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-kafka-source.json
@@ -698,7 +698,7 @@
},
"camel.component.kafka.kafkaClientFactory": {
"name": "camel.component.kafka.kafkaClientFactory",
- "description": "Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances with logic that
extends the vanilla Kafka clients.",
+ "description": "Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create instances with logic that extends the
vanilla Kafka clients.",
"priority": "MEDIUM",
"required": "false"
},
diff --git
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-milo-client-sink.json
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-milo-client-sink.json
index a626861..3ded443 100644
---
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-milo-client-sink.json
+++
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-milo-client-sink.json
@@ -313,6 +313,12 @@
"priority": "MEDIUM",
"required": "false"
},
+ "camel.component.milo-client.miloClientConnectionManager": {
+ "name":
"camel.component.milo-client.miloClientConnectionManager",
+ "description": "Instance for managing client
connections",
+ "priority": "MEDIUM",
+ "required": "false"
+ },
"camel.component.milo-client.overrideHost": {
"name": "camel.component.milo-client.overrideHost",
"description": "Override the server reported endpoint
host with the host from the endpoint URI.",
diff --git
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-milo-client-source.json
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-milo-client-source.json
index 83746c4..1b45a55 100644
---
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-milo-client-source.json
+++
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-milo-client-source.json
@@ -330,6 +330,12 @@
"priority": "MEDIUM",
"required": "false"
},
+ "camel.component.milo-client.miloClientConnectionManager": {
+ "name":
"camel.component.milo-client.miloClientConnectionManager",
+ "description": "Instance for managing client
connections",
+ "priority": "MEDIUM",
+ "required": "false"
+ },
"camel.component.milo-client.overrideHost": {
"name": "camel.component.milo-client.overrideHost",
"description": "Override the server reported endpoint
host with the host from the endpoint URI.",
diff --git
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-vertx-kafka-sink.json
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-vertx-kafka-sink.json
index 459b4d5..b0021be 100644
---
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-vertx-kafka-sink.json
+++
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-vertx-kafka-sink.json
@@ -296,12 +296,6 @@
"priority": "MEDIUM",
"required": "false"
},
- "camel.sink.endpoint.vertxKafkaClientFactory": {
- "name": "camel.sink.endpoint.vertxKafkaClientFactory",
- "description": "Factory to use for creating
io.vertx.kafka.client.consumer.KafkaConsumer and
io.vertx.kafka.client.consumer.KafkaProducer instances. This allows to
configure a custom factory to create custom KafkaConsumer and KafkaProducer
instances with logic that extends the vanilla VertX Kafka clients.",
- "priority": "MEDIUM",
- "required": "false"
- },
"camel.sink.endpoint.saslClientCallbackHandlerClass": {
"name":
"camel.sink.endpoint.saslClientCallbackHandlerClass",
"description": "The fully qualified name of a SASL
client callback handler class that implements the AuthenticateCallbackHandler
interface.",
diff --git
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-vertx-kafka-source.json
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-vertx-kafka-source.json
index f049e88..6294c79 100644
---
a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-vertx-kafka-source.json
+++
b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-vertx-kafka-source.json
@@ -365,12 +365,6 @@
"InOptionalOut"
]
},
- "camel.source.endpoint.vertxKafkaClientFactory": {
- "name": "camel.source.endpoint.vertxKafkaClientFactory",
- "description": "Factory to use for creating
io.vertx.kafka.client.consumer.KafkaConsumer and
io.vertx.kafka.client.consumer.KafkaProducer instances. This allows to
configure a custom factory to create custom KafkaConsumer and KafkaProducer
instances with logic that extends the vanilla VertX Kafka clients.",
- "priority": "MEDIUM",
- "required": "false"
- },
"camel.source.endpoint.saslClientCallbackHandlerClass": {
"name":
"camel.source.endpoint.saslClientCallbackHandlerClass",
"description": "The fully qualified name of a SASL
client callback handler class that implements the AuthenticateCallbackHandler
interface.",
diff --git
a/connectors/camel-aws2-sqs-kafka-connector/src/generated/resources/camel-aws2-sqs-sink.json
b/connectors/camel-aws2-sqs-kafka-connector/src/generated/resources/camel-aws2-sqs-sink.json
index 8d185cb..be8e767 100644
---
a/connectors/camel-aws2-sqs-kafka-connector/src/generated/resources/camel-aws2-sqs-sink.json
+++
b/connectors/camel-aws2-sqs-kafka-connector/src/generated/resources/camel-aws2-sqs-sink.json
@@ -85,6 +85,13 @@
"priority": "MEDIUM",
"required": "false"
},
+ "camel.sink.endpoint.batchSeparator": {
+ "name": "camel.sink.endpoint.batchSeparator",
+ "description": "Set the separator when passing a String
to send batch message operation",
+ "defaultValue": "\",\"",
+ "priority": "MEDIUM",
+ "required": "false"
+ },
"camel.sink.endpoint.delaySeconds": {
"name": "camel.sink.endpoint.delaySeconds",
"description": "Delay sending messages for a number of
seconds.",
@@ -269,6 +276,13 @@
"priority": "MEDIUM",
"required": "false"
},
+ "camel.component.aws2-sqs.batchSeparator": {
+ "name": "camel.component.aws2-sqs.batchSeparator",
+ "description": "Set the separator when passing a String
to send batch message operation",
+ "defaultValue": "\",\"",
+ "priority": "MEDIUM",
+ "required": "false"
+ },
"camel.component.aws2-sqs.delaySeconds": {
"name": "camel.component.aws2-sqs.delaySeconds",
"description": "Delay sending messages for a number of
seconds.",
diff --git
a/connectors/camel-aws2-sqs-kafka-connector/src/main/docs/camel-aws2-sqs-kafka-sink-connector.adoc
b/connectors/camel-aws2-sqs-kafka-connector/src/main/docs/camel-aws2-sqs-kafka-sink-connector.adoc
index 1ed8652..d218f26 100644
---
a/connectors/camel-aws2-sqs-kafka-connector/src/main/docs/camel-aws2-sqs-kafka-sink-connector.adoc
+++
b/connectors/camel-aws2-sqs-kafka-connector/src/main/docs/camel-aws2-sqs-kafka-sink-connector.adoc
@@ -24,7 +24,7 @@
connector.class=org.apache.camel.kafkaconnector.aws2sqs.CamelAws2sqsSinkConnecto
----
-The camel-aws2-sqs sink connector supports 54 options, which are listed below.
+The camel-aws2-sqs sink connector supports 56 options, which are listed below.
@@ -42,6 +42,7 @@ The camel-aws2-sqs sink connector supports 54 options, which
are listed below.
| *camel.sink.endpoint.region* | The region in which SQS client needs to work.
When using this parameter, the configuration will expect the lowercase name of
the region (for example ap-east-1) You'll need to use the name
Region.EU_WEST_1.id() | null | false | MEDIUM
| *camel.sink.endpoint.trustAllCertificates* | If we want to trust all
certificates in case of overriding the endpoint | false | false | MEDIUM
| *camel.sink.endpoint.useDefaultCredentialsProvider* | Set whether the SQS
client should expect to load credentials on an AWS infra instance or to expect
static credentials to be passed in. | false | false | MEDIUM
+| *camel.sink.endpoint.batchSeparator* | Set the separator when passing a
String to send batch message operation | "," | false | MEDIUM
| *camel.sink.endpoint.delaySeconds* | Delay sending messages for a number of
seconds. | null | false | MEDIUM
| *camel.sink.endpoint.lazyStartProducer* | Whether the producer should be
started lazy (on the first message). By starting lazy you can use this to allow
CamelContext and routes to startup in situations where a producer may otherwise
fail during starting and cause the route to fail being started. By deferring
this startup to be lazy then the startup failure can be handled during routing
messages via Camel's routing error handlers. Beware that when the first message
is processed then cre [...]
| *camel.sink.endpoint.messageDeduplicationIdStrategy* | Only for FIFO queues.
Strategy for setting the messageDeduplicationId on the message. Can be one of
the following options: useExchangeId, useContentBasedDeduplication. For the
useContentBasedDeduplication option, no messageDeduplicationId will be set on
the message. One of: [useExchangeId] [useContentBasedDeduplication] |
"useExchangeId" | false | MEDIUM
@@ -68,6 +69,7 @@ The camel-aws2-sqs sink connector supports 54 options, which
are listed below.
| *camel.component.aws2-sqs.region* | The region in which SQS client needs to
work. When using this parameter, the configuration will expect the lowercase
name of the region (for example ap-east-1) You'll need to use the name
Region.EU_WEST_1.id() | null | false | MEDIUM
| *camel.component.aws2-sqs.trustAllCertificates* | If we want to trust all
certificates in case of overriding the endpoint | false | false | MEDIUM
| *camel.component.aws2-sqs.useDefaultCredentials Provider* | Set whether the
SQS client should expect to load credentials on an AWS infra instance or to
expect static credentials to be passed in. | false | false | MEDIUM
+| *camel.component.aws2-sqs.batchSeparator* | Set the separator when passing a
String to send batch message operation | "," | false | MEDIUM
| *camel.component.aws2-sqs.delaySeconds* | Delay sending messages for a
number of seconds. | null | false | MEDIUM
| *camel.component.aws2-sqs.lazyStartProducer* | Whether the producer should
be started lazy (on the first message). By starting lazy you can use this to
allow CamelContext and routes to startup in situations where a producer may
otherwise fail during starting and cause the route to fail being started. By
deferring this startup to be lazy then the startup failure can be handled
during routing messages via Camel's routing error handlers. Beware that when
the first message is processed the [...]
| *camel.component.aws2-sqs.messageDeduplicationId Strategy* | Only for FIFO
queues. Strategy for setting the messageDeduplicationId on the message. Can be
one of the following options: useExchangeId, useContentBasedDeduplication. For
the useContentBasedDeduplication option, no messageDeduplicationId will be set
on the message. One of: [useExchangeId] [useContentBasedDeduplication] |
"useExchangeId" | false | MEDIUM
diff --git
a/connectors/camel-aws2-sqs-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2sqs/CamelAws2sqsSinkConnectorConfig.java
b/connectors/camel-aws2-sqs-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2sqs/CamelAws2sqsSinkConnectorConfig.java
index adcead1..6214a36 100644
---
a/connectors/camel-aws2-sqs-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2sqs/CamelAws2sqsSinkConnectorConfig.java
+++
b/connectors/camel-aws2-sqs-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2sqs/CamelAws2sqsSinkConnectorConfig.java
@@ -57,6 +57,9 @@ public class CamelAws2sqsSinkConnectorConfig extends
CamelSinkConnectorConfig {
public static final String
CAMEL_SINK_AWS2SQS_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_CONF =
"camel.sink.endpoint.useDefaultCredentialsProvider";
public static final String
CAMEL_SINK_AWS2SQS_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_DOC = "Set whether
the SQS client should expect to load credentials on an AWS infra instance or to
expect static credentials to be passed in.";
public static final Boolean
CAMEL_SINK_AWS2SQS_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_DEFAULT = false;
+ public static final String
CAMEL_SINK_AWS2SQS_ENDPOINT_BATCH_SEPARATOR_CONF =
"camel.sink.endpoint.batchSeparator";
+ public static final String CAMEL_SINK_AWS2SQS_ENDPOINT_BATCH_SEPARATOR_DOC
= "Set the separator when passing a String to send batch message operation";
+ public static final String
CAMEL_SINK_AWS2SQS_ENDPOINT_BATCH_SEPARATOR_DEFAULT = ",";
public static final String CAMEL_SINK_AWS2SQS_ENDPOINT_DELAY_SECONDS_CONF
= "camel.sink.endpoint.delaySeconds";
public static final String CAMEL_SINK_AWS2SQS_ENDPOINT_DELAY_SECONDS_DOC =
"Delay sending messages for a number of seconds.";
public static final String
CAMEL_SINK_AWS2SQS_ENDPOINT_DELAY_SECONDS_DEFAULT = null;
@@ -135,6 +138,9 @@ public class CamelAws2sqsSinkConnectorConfig extends
CamelSinkConnectorConfig {
public static final String
CAMEL_SINK_AWS2SQS_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_CONF =
"camel.component.aws2-sqs.useDefaultCredentialsProvider";
public static final String
CAMEL_SINK_AWS2SQS_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_DOC = "Set
whether the SQS client should expect to load credentials on an AWS infra
instance or to expect static credentials to be passed in.";
public static final Boolean
CAMEL_SINK_AWS2SQS_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_DEFAULT = false;
+ public static final String
CAMEL_SINK_AWS2SQS_COMPONENT_BATCH_SEPARATOR_CONF =
"camel.component.aws2-sqs.batchSeparator";
+ public static final String
CAMEL_SINK_AWS2SQS_COMPONENT_BATCH_SEPARATOR_DOC = "Set the separator when
passing a String to send batch message operation";
+ public static final String
CAMEL_SINK_AWS2SQS_COMPONENT_BATCH_SEPARATOR_DEFAULT = ",";
public static final String CAMEL_SINK_AWS2SQS_COMPONENT_DELAY_SECONDS_CONF
= "camel.component.aws2-sqs.delaySeconds";
public static final String CAMEL_SINK_AWS2SQS_COMPONENT_DELAY_SECONDS_DOC
= "Delay sending messages for a number of seconds.";
public static final String
CAMEL_SINK_AWS2SQS_COMPONENT_DELAY_SECONDS_DEFAULT = null;
@@ -210,6 +216,7 @@ public class CamelAws2sqsSinkConnectorConfig extends
CamelSinkConnectorConfig {
conf.define(CAMEL_SINK_AWS2SQS_ENDPOINT_REGION_CONF,
ConfigDef.Type.STRING, CAMEL_SINK_AWS2SQS_ENDPOINT_REGION_DEFAULT,
ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2SQS_ENDPOINT_REGION_DOC);
conf.define(CAMEL_SINK_AWS2SQS_ENDPOINT_TRUST_ALL_CERTIFICATES_CONF,
ConfigDef.Type.BOOLEAN,
CAMEL_SINK_AWS2SQS_ENDPOINT_TRUST_ALL_CERTIFICATES_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_AWS2SQS_ENDPOINT_TRUST_ALL_CERTIFICATES_DOC);
conf.define(CAMEL_SINK_AWS2SQS_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_CONF,
ConfigDef.Type.BOOLEAN,
CAMEL_SINK_AWS2SQS_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_AWS2SQS_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_DOC);
+ conf.define(CAMEL_SINK_AWS2SQS_ENDPOINT_BATCH_SEPARATOR_CONF,
ConfigDef.Type.STRING, CAMEL_SINK_AWS2SQS_ENDPOINT_BATCH_SEPARATOR_DEFAULT,
ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2SQS_ENDPOINT_BATCH_SEPARATOR_DOC);
conf.define(CAMEL_SINK_AWS2SQS_ENDPOINT_DELAY_SECONDS_CONF,
ConfigDef.Type.STRING, CAMEL_SINK_AWS2SQS_ENDPOINT_DELAY_SECONDS_DEFAULT,
ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2SQS_ENDPOINT_DELAY_SECONDS_DOC);
conf.define(CAMEL_SINK_AWS2SQS_ENDPOINT_LAZY_START_PRODUCER_CONF,
ConfigDef.Type.BOOLEAN,
CAMEL_SINK_AWS2SQS_ENDPOINT_LAZY_START_PRODUCER_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_AWS2SQS_ENDPOINT_LAZY_START_PRODUCER_DOC);
conf.define(CAMEL_SINK_AWS2SQS_ENDPOINT_MESSAGE_DEDUPLICATION_ID_STRATEGY_CONF,
ConfigDef.Type.STRING,
CAMEL_SINK_AWS2SQS_ENDPOINT_MESSAGE_DEDUPLICATION_ID_STRATEGY_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_AWS2SQS_ENDPOINT_MESSAGE_DEDUPLICATION_ID_STRATEGY_DOC);
@@ -236,6 +243,7 @@ public class CamelAws2sqsSinkConnectorConfig extends
CamelSinkConnectorConfig {
conf.define(CAMEL_SINK_AWS2SQS_COMPONENT_REGION_CONF,
ConfigDef.Type.STRING, CAMEL_SINK_AWS2SQS_COMPONENT_REGION_DEFAULT,
ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2SQS_COMPONENT_REGION_DOC);
conf.define(CAMEL_SINK_AWS2SQS_COMPONENT_TRUST_ALL_CERTIFICATES_CONF,
ConfigDef.Type.BOOLEAN,
CAMEL_SINK_AWS2SQS_COMPONENT_TRUST_ALL_CERTIFICATES_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_AWS2SQS_COMPONENT_TRUST_ALL_CERTIFICATES_DOC);
conf.define(CAMEL_SINK_AWS2SQS_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_CONF,
ConfigDef.Type.BOOLEAN,
CAMEL_SINK_AWS2SQS_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_AWS2SQS_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_DOC);
+ conf.define(CAMEL_SINK_AWS2SQS_COMPONENT_BATCH_SEPARATOR_CONF,
ConfigDef.Type.STRING, CAMEL_SINK_AWS2SQS_COMPONENT_BATCH_SEPARATOR_DEFAULT,
ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2SQS_COMPONENT_BATCH_SEPARATOR_DOC);
conf.define(CAMEL_SINK_AWS2SQS_COMPONENT_DELAY_SECONDS_CONF,
ConfigDef.Type.STRING, CAMEL_SINK_AWS2SQS_COMPONENT_DELAY_SECONDS_DEFAULT,
ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2SQS_COMPONENT_DELAY_SECONDS_DOC);
conf.define(CAMEL_SINK_AWS2SQS_COMPONENT_LAZY_START_PRODUCER_CONF,
ConfigDef.Type.BOOLEAN,
CAMEL_SINK_AWS2SQS_COMPONENT_LAZY_START_PRODUCER_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_AWS2SQS_COMPONENT_LAZY_START_PRODUCER_DOC);
conf.define(CAMEL_SINK_AWS2SQS_COMPONENT_MESSAGE_DEDUPLICATION_ID_STRATEGY_CONF,
ConfigDef.Type.STRING,
CAMEL_SINK_AWS2SQS_COMPONENT_MESSAGE_DEDUPLICATION_ID_STRATEGY_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_AWS2SQS_COMPONENT_MESSAGE_DEDUPLICATION_ID_STRATEGY_DOC);
diff --git
a/connectors/camel-kafka-kafka-connector/src/generated/resources/camel-kafka-sink.json
b/connectors/camel-kafka-kafka-connector/src/generated/resources/camel-kafka-sink.json
index b997786..dfae88e 100644
---
a/connectors/camel-kafka-kafka-connector/src/generated/resources/camel-kafka-sink.json
+++
b/connectors/camel-kafka-kafka-connector/src/generated/resources/camel-kafka-sink.json
@@ -743,7 +743,7 @@
},
"camel.component.kafka.kafkaClientFactory": {
"name": "camel.component.kafka.kafkaClientFactory",
- "description": "Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances with logic that
extends the vanilla Kafka clients.",
+ "description": "Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create instances with logic that extends the
vanilla Kafka clients.",
"priority": "MEDIUM",
"required": "false"
},
diff --git
a/connectors/camel-kafka-kafka-connector/src/generated/resources/camel-kafka-source.json
b/connectors/camel-kafka-kafka-connector/src/generated/resources/camel-kafka-source.json
index ebfe2c8..7124ba4 100644
---
a/connectors/camel-kafka-kafka-connector/src/generated/resources/camel-kafka-source.json
+++
b/connectors/camel-kafka-kafka-connector/src/generated/resources/camel-kafka-source.json
@@ -698,7 +698,7 @@
},
"camel.component.kafka.kafkaClientFactory": {
"name": "camel.component.kafka.kafkaClientFactory",
- "description": "Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances with logic that
extends the vanilla Kafka clients.",
+ "description": "Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create instances with logic that extends the
vanilla Kafka clients.",
"priority": "MEDIUM",
"required": "false"
},
diff --git
a/connectors/camel-kafka-kafka-connector/src/main/docs/camel-kafka-kafka-sink-connector.adoc
b/connectors/camel-kafka-kafka-connector/src/main/docs/camel-kafka-kafka-sink-connector.adoc
index ea98fa7..3e2b65e 100644
---
a/connectors/camel-kafka-kafka-connector/src/main/docs/camel-kafka-kafka-sink-connector.adoc
+++
b/connectors/camel-kafka-kafka-connector/src/main/docs/camel-kafka-kafka-sink-connector.adoc
@@ -137,7 +137,7 @@ The camel-kafka sink connector supports 135 options, which
are listed below.
| *camel.component.kafka.workerPoolCoreSize* | Number of core threads for the
worker pool for continue routing Exchange after kafka server has acknowledge
the message that was sent to it from KafkaProducer using asynchronous
non-blocking processing. | "10" | false | MEDIUM
| *camel.component.kafka.workerPoolMaxSize* | Maximum number of threads for
the worker pool for continue routing Exchange after kafka server has
acknowledge the message that was sent to it from KafkaProducer using
asynchronous non-blocking processing. | "20" | false | MEDIUM
| *camel.component.kafka.autowiredEnabled* | Whether autowiring is enabled.
This is used for automatic autowiring options (the option must be marked as
autowired) by looking up in the registry to find if there is a single instance
of matching type, which then gets configured on the component. This can be used
for automatic configuring JDBC data sources, JMS connection factories, AWS
Clients, etc. | true | false | MEDIUM
-| *camel.component.kafka.kafkaClientFactory* | Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances with logic that
extends the vanilla Kafka clients. | null | false | MEDIUM
+| *camel.component.kafka.kafkaClientFactory* | Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create instances with logic that extends the
vanilla Kafka clients. | null | false | MEDIUM
| *camel.component.kafka.synchronous* | Sets whether synchronous processing
should be strictly used | false | false | MEDIUM
| *camel.component.kafka.schemaRegistryURL* | URL of the Confluent Platform
schema registry servers to use. The format is host1:port1,host2:port2. This is
known as schema.registry.url in the Confluent Platform documentation. This
option is only available in the Confluent Platform (not standard Apache Kafka)
| null | false | MEDIUM
| *camel.component.kafka.interceptorClasses* | Sets interceptors for producer
or consumers. Producer interceptors have to be classes implementing
org.apache.kafka.clients.producer.ProducerInterceptor Consumer interceptors
have to be classes implementing
org.apache.kafka.clients.consumer.ConsumerInterceptor Note that if you use
Producer interceptor on a consumer it will throw a class cast exception in
runtime | null | false | MEDIUM
diff --git
a/connectors/camel-kafka-kafka-connector/src/main/docs/camel-kafka-kafka-source-connector.adoc
b/connectors/camel-kafka-kafka-connector/src/main/docs/camel-kafka-kafka-source-connector.adoc
index b41e14b..03f8c86 100644
---
a/connectors/camel-kafka-kafka-connector/src/main/docs/camel-kafka-kafka-source-connector.adoc
+++
b/connectors/camel-kafka-kafka-connector/src/main/docs/camel-kafka-kafka-source-connector.adoc
@@ -129,7 +129,7 @@ The camel-kafka source connector supports 122 options,
which are listed below.
| *camel.component.kafka.valueDeserializer* | Deserializer class for value
that implements the Deserializer interface. |
"org.apache.kafka.common.serialization.StringDeserializer" | false | MEDIUM
| *camel.component.kafka.kafkaManualCommitFactory* | Factory to use for
creating KafkaManualCommit instances. This allows to plugin a custom factory to
create custom KafkaManualCommit instances in case special logic is needed when
doing manual commits that deviates from the default implementation that comes
out of the box. | null | false | MEDIUM
| *camel.component.kafka.autowiredEnabled* | Whether autowiring is enabled.
This is used for automatic autowiring options (the option must be marked as
autowired) by looking up in the registry to find if there is a single instance
of matching type, which then gets configured on the component. This can be used
for automatic configuring JDBC data sources, JMS connection factories, AWS
Clients, etc. | true | false | MEDIUM
-| *camel.component.kafka.kafkaClientFactory* | Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances with logic that
extends the vanilla Kafka clients. | null | false | MEDIUM
+| *camel.component.kafka.kafkaClientFactory* | Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create instances with logic that extends the
vanilla Kafka clients. | null | false | MEDIUM
| *camel.component.kafka.synchronous* | Sets whether synchronous processing
should be strictly used | false | false | MEDIUM
| *camel.component.kafka.schemaRegistryURL* | URL of the Confluent Platform
schema registry servers to use. The format is host1:port1,host2:port2. This is
known as schema.registry.url in the Confluent Platform documentation. This
option is only available in the Confluent Platform (not standard Apache Kafka)
| null | false | MEDIUM
| *camel.component.kafka.interceptorClasses* | Sets interceptors for producer
or consumers. Producer interceptors have to be classes implementing
org.apache.kafka.clients.producer.ProducerInterceptor Consumer interceptors
have to be classes implementing
org.apache.kafka.clients.consumer.ConsumerInterceptor Note that if you use
Producer interceptor on a consumer it will throw a class cast exception in
runtime | null | false | MEDIUM
diff --git
a/connectors/camel-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/kafka/CamelKafkaSinkConnectorConfig.java
b/connectors/camel-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/kafka/CamelKafkaSinkConnectorConfig.java
index 327cd04..6ae1f14 100644
---
a/connectors/camel-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/kafka/CamelKafkaSinkConnectorConfig.java
+++
b/connectors/camel-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/kafka/CamelKafkaSinkConnectorConfig.java
@@ -343,7 +343,7 @@ public class CamelKafkaSinkConnectorConfig extends
CamelSinkConnectorConfig {
public static final String
CAMEL_SINK_KAFKA_COMPONENT_AUTOWIRED_ENABLED_DOC = "Whether autowiring is
enabled. This is used for automatic autowiring options (the option must be
marked as autowired) by looking up in the registry to find if there is a single
instance of matching type, which then gets configured on the component. This
can be used for automatic configuring JDBC data sources, JMS connection
factories, AWS Clients, etc.";
public static final Boolean
CAMEL_SINK_KAFKA_COMPONENT_AUTOWIRED_ENABLED_DEFAULT = true;
public static final String
CAMEL_SINK_KAFKA_COMPONENT_KAFKA_CLIENT_FACTORY_CONF =
"camel.component.kafka.kafkaClientFactory";
- public static final String
CAMEL_SINK_KAFKA_COMPONENT_KAFKA_CLIENT_FACTORY_DOC = "Factory to use for
creating org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances with logic that
extends the vanilla Kafka clients.";
+ public static final String
CAMEL_SINK_KAFKA_COMPONENT_KAFKA_CLIENT_FACTORY_DOC = "Factory to use for
creating org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create instances with logic that extends the
vanilla Kafka clients.";
public static final String
CAMEL_SINK_KAFKA_COMPONENT_KAFKA_CLIENT_FACTORY_DEFAULT = null;
public static final String CAMEL_SINK_KAFKA_COMPONENT_SYNCHRONOUS_CONF =
"camel.component.kafka.synchronous";
public static final String CAMEL_SINK_KAFKA_COMPONENT_SYNCHRONOUS_DOC =
"Sets whether synchronous processing should be strictly used";
diff --git
a/connectors/camel-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/kafka/CamelKafkaSourceConnectorConfig.java
b/connectors/camel-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/kafka/CamelKafkaSourceConnectorConfig.java
index 50627a8..8f81152 100644
---
a/connectors/camel-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/kafka/CamelKafkaSourceConnectorConfig.java
+++
b/connectors/camel-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/kafka/CamelKafkaSourceConnectorConfig.java
@@ -321,7 +321,7 @@ public class CamelKafkaSourceConnectorConfig
public static final String
CAMEL_SOURCE_KAFKA_COMPONENT_AUTOWIRED_ENABLED_DOC = "Whether autowiring is
enabled. This is used for automatic autowiring options (the option must be
marked as autowired) by looking up in the registry to find if there is a single
instance of matching type, which then gets configured on the component. This
can be used for automatic configuring JDBC data sources, JMS connection
factories, AWS Clients, etc.";
public static final Boolean
CAMEL_SOURCE_KAFKA_COMPONENT_AUTOWIRED_ENABLED_DEFAULT = true;
public static final String
CAMEL_SOURCE_KAFKA_COMPONENT_KAFKA_CLIENT_FACTORY_CONF =
"camel.component.kafka.kafkaClientFactory";
- public static final String
CAMEL_SOURCE_KAFKA_COMPONENT_KAFKA_CLIENT_FACTORY_DOC = "Factory to use for
creating org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances with logic that
extends the vanilla Kafka clients.";
+ public static final String
CAMEL_SOURCE_KAFKA_COMPONENT_KAFKA_CLIENT_FACTORY_DOC = "Factory to use for
creating org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create instances with logic that extends the
vanilla Kafka clients.";
public static final String
CAMEL_SOURCE_KAFKA_COMPONENT_KAFKA_CLIENT_FACTORY_DEFAULT = null;
public static final String CAMEL_SOURCE_KAFKA_COMPONENT_SYNCHRONOUS_CONF =
"camel.component.kafka.synchronous";
public static final String CAMEL_SOURCE_KAFKA_COMPONENT_SYNCHRONOUS_DOC =
"Sets whether synchronous processing should be strictly used";
diff --git
a/connectors/camel-milo-client-kafka-connector/src/generated/resources/camel-milo-client-sink.json
b/connectors/camel-milo-client-kafka-connector/src/generated/resources/camel-milo-client-sink.json
index a626861..3ded443 100644
---
a/connectors/camel-milo-client-kafka-connector/src/generated/resources/camel-milo-client-sink.json
+++
b/connectors/camel-milo-client-kafka-connector/src/generated/resources/camel-milo-client-sink.json
@@ -313,6 +313,12 @@
"priority": "MEDIUM",
"required": "false"
},
+ "camel.component.milo-client.miloClientConnectionManager": {
+ "name":
"camel.component.milo-client.miloClientConnectionManager",
+ "description": "Instance for managing client
connections",
+ "priority": "MEDIUM",
+ "required": "false"
+ },
"camel.component.milo-client.overrideHost": {
"name": "camel.component.milo-client.overrideHost",
"description": "Override the server reported endpoint
host with the host from the endpoint URI.",
diff --git
a/connectors/camel-milo-client-kafka-connector/src/generated/resources/camel-milo-client-source.json
b/connectors/camel-milo-client-kafka-connector/src/generated/resources/camel-milo-client-source.json
index 83746c4..1b45a55 100644
---
a/connectors/camel-milo-client-kafka-connector/src/generated/resources/camel-milo-client-source.json
+++
b/connectors/camel-milo-client-kafka-connector/src/generated/resources/camel-milo-client-source.json
@@ -330,6 +330,12 @@
"priority": "MEDIUM",
"required": "false"
},
+ "camel.component.milo-client.miloClientConnectionManager": {
+ "name":
"camel.component.milo-client.miloClientConnectionManager",
+ "description": "Instance for managing client
connections",
+ "priority": "MEDIUM",
+ "required": "false"
+ },
"camel.component.milo-client.overrideHost": {
"name": "camel.component.milo-client.overrideHost",
"description": "Override the server reported endpoint
host with the host from the endpoint URI.",
diff --git
a/connectors/camel-milo-client-kafka-connector/src/main/docs/camel-milo-client-kafka-sink-connector.adoc
b/connectors/camel-milo-client-kafka-connector/src/main/docs/camel-milo-client-kafka-sink-connector.adoc
index 53546d0..82650ab 100644
---
a/connectors/camel-milo-client-kafka-connector/src/main/docs/camel-milo-client-kafka-sink-connector.adoc
+++
b/connectors/camel-milo-client-kafka-connector/src/main/docs/camel-milo-client-kafka-sink-connector.adoc
@@ -24,7 +24,7 @@
connector.class=org.apache.camel.kafkaconnector.miloclient.CamelMiloclientSinkCo
----
-The camel-milo-client sink connector supports 53 options, which are listed
below.
+The camel-milo-client sink connector supports 54 options, which are listed
below.
@@ -78,6 +78,7 @@ The camel-milo-client sink connector supports 53 options,
which are listed below
| *camel.component.milo-client.keyStoreUrl* | The URL where the key should be
loaded from | null | false | MEDIUM
| *camel.component.milo-client.maxPendingPublish Requests* | The maximum
number of pending publish requests | null | false | MEDIUM
| *camel.component.milo-client.maxResponseMessageSize* | The maximum number of
bytes a response message may have | null | false | MEDIUM
+| *camel.component.milo-client.miloClientConnection Manager* | Instance for
managing client connections | null | false | MEDIUM
| *camel.component.milo-client.overrideHost* | Override the server reported
endpoint host with the host from the endpoint URI. | false | false | MEDIUM
| *camel.component.milo-client.productUri* | The product URI |
"http://camel.apache.org/EclipseMilo" | false | MEDIUM
| *camel.component.milo-client.requestedPublishing Interval* | The requested
publishing interval in milliseconds | "1_000.0" | false | MEDIUM
diff --git
a/connectors/camel-milo-client-kafka-connector/src/main/docs/camel-milo-client-kafka-source-connector.adoc
b/connectors/camel-milo-client-kafka-connector/src/main/docs/camel-milo-client-kafka-source-connector.adoc
index 1f1ece7..755b388 100644
---
a/connectors/camel-milo-client-kafka-connector/src/main/docs/camel-milo-client-kafka-source-connector.adoc
+++
b/connectors/camel-milo-client-kafka-connector/src/main/docs/camel-milo-client-kafka-source-connector.adoc
@@ -24,7 +24,7 @@
connector.class=org.apache.camel.kafkaconnector.miloclient.CamelMiloclientSource
----
-The camel-milo-client source connector supports 55 options, which are listed
below.
+The camel-milo-client source connector supports 56 options, which are listed
below.
@@ -80,6 +80,7 @@ The camel-milo-client source connector supports 55 options,
which are listed bel
| *camel.component.milo-client.keyStoreUrl* | The URL where the key should be
loaded from | null | false | MEDIUM
| *camel.component.milo-client.maxPendingPublish Requests* | The maximum
number of pending publish requests | null | false | MEDIUM
| *camel.component.milo-client.maxResponseMessageSize* | The maximum number of
bytes a response message may have | null | false | MEDIUM
+| *camel.component.milo-client.miloClientConnection Manager* | Instance for
managing client connections | null | false | MEDIUM
| *camel.component.milo-client.overrideHost* | Override the server reported
endpoint host with the host from the endpoint URI. | false | false | MEDIUM
| *camel.component.milo-client.productUri* | The product URI |
"http://camel.apache.org/EclipseMilo" | false | MEDIUM
| *camel.component.milo-client.requestedPublishing Interval* | The requested
publishing interval in milliseconds | "1_000.0" | false | MEDIUM
diff --git
a/connectors/camel-milo-client-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/miloclient/CamelMiloclientSinkConnectorConfig.java
b/connectors/camel-milo-client-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/miloclient/CamelMiloclientSinkConnectorConfig.java
index 9ec86e8..868e10e 100644
---
a/connectors/camel-milo-client-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/miloclient/CamelMiloclientSinkConnectorConfig.java
+++
b/connectors/camel-milo-client-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/miloclient/CamelMiloclientSinkConnectorConfig.java
@@ -167,6 +167,9 @@ public class CamelMiloclientSinkConnectorConfig
public static final String
CAMEL_SINK_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_CONF =
"camel.component.milo-client.maxResponseMessageSize";
public static final String
CAMEL_SINK_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_DOC = "The maximum
number of bytes a response message may have";
public static final String
CAMEL_SINK_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_DEFAULT = null;
+ public static final String
CAMEL_SINK_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_CONF =
"camel.component.milo-client.miloClientConnectionManager";
+ public static final String
CAMEL_SINK_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_DOC = "Instance
for managing client connections";
+ public static final String
CAMEL_SINK_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_DEFAULT = null;
public static final String
CAMEL_SINK_MILOCLIENT_COMPONENT_OVERRIDE_HOST_CONF =
"camel.component.milo-client.overrideHost";
public static final String
CAMEL_SINK_MILOCLIENT_COMPONENT_OVERRIDE_HOST_DOC = "Override the server
reported endpoint host with the host from the endpoint URI.";
public static final Boolean
CAMEL_SINK_MILOCLIENT_COMPONENT_OVERRIDE_HOST_DEFAULT = false;
@@ -245,6 +248,7 @@ public class CamelMiloclientSinkConnectorConfig
conf.define(CAMEL_SINK_MILOCLIENT_COMPONENT_KEY_STORE_URL_CONF,
ConfigDef.Type.STRING, CAMEL_SINK_MILOCLIENT_COMPONENT_KEY_STORE_URL_DEFAULT,
ConfigDef.Importance.MEDIUM, CAMEL_SINK_MILOCLIENT_COMPONENT_KEY_STORE_URL_DOC);
conf.define(CAMEL_SINK_MILOCLIENT_COMPONENT_MAX_PENDING_PUBLISH_REQUESTS_CONF,
ConfigDef.Type.STRING,
CAMEL_SINK_MILOCLIENT_COMPONENT_MAX_PENDING_PUBLISH_REQUESTS_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_MILOCLIENT_COMPONENT_MAX_PENDING_PUBLISH_REQUESTS_DOC);
conf.define(CAMEL_SINK_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_CONF,
ConfigDef.Type.STRING,
CAMEL_SINK_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_DOC);
+
conf.define(CAMEL_SINK_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_CONF,
ConfigDef.Type.STRING,
CAMEL_SINK_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_DOC);
conf.define(CAMEL_SINK_MILOCLIENT_COMPONENT_OVERRIDE_HOST_CONF,
ConfigDef.Type.BOOLEAN, CAMEL_SINK_MILOCLIENT_COMPONENT_OVERRIDE_HOST_DEFAULT,
ConfigDef.Importance.MEDIUM, CAMEL_SINK_MILOCLIENT_COMPONENT_OVERRIDE_HOST_DOC);
conf.define(CAMEL_SINK_MILOCLIENT_COMPONENT_PRODUCT_URI_CONF,
ConfigDef.Type.STRING, CAMEL_SINK_MILOCLIENT_COMPONENT_PRODUCT_URI_DEFAULT,
ConfigDef.Importance.MEDIUM, CAMEL_SINK_MILOCLIENT_COMPONENT_PRODUCT_URI_DOC);
conf.define(CAMEL_SINK_MILOCLIENT_COMPONENT_REQUESTED_PUBLISHING_INTERVAL_CONF,
ConfigDef.Type.STRING,
CAMEL_SINK_MILOCLIENT_COMPONENT_REQUESTED_PUBLISHING_INTERVAL_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_MILOCLIENT_COMPONENT_REQUESTED_PUBLISHING_INTERVAL_DOC);
diff --git
a/connectors/camel-milo-client-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/miloclient/CamelMiloclientSourceConnectorConfig.java
b/connectors/camel-milo-client-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/miloclient/CamelMiloclientSourceConnectorConfig.java
index b0785e5..f791706 100644
---
a/connectors/camel-milo-client-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/miloclient/CamelMiloclientSourceConnectorConfig.java
+++
b/connectors/camel-milo-client-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/miloclient/CamelMiloclientSourceConnectorConfig.java
@@ -173,6 +173,9 @@ public class CamelMiloclientSourceConnectorConfig
public static final String
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_CONF =
"camel.component.milo-client.maxResponseMessageSize";
public static final String
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_DOC = "The maximum
number of bytes a response message may have";
public static final String
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_DEFAULT = null;
+ public static final String
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_CONF =
"camel.component.milo-client.miloClientConnectionManager";
+ public static final String
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_DOC =
"Instance for managing client connections";
+ public static final String
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_DEFAULT = null;
public static final String
CAMEL_SOURCE_MILOCLIENT_COMPONENT_OVERRIDE_HOST_CONF =
"camel.component.milo-client.overrideHost";
public static final String
CAMEL_SOURCE_MILOCLIENT_COMPONENT_OVERRIDE_HOST_DOC = "Override the server
reported endpoint host with the host from the endpoint URI.";
public static final Boolean
CAMEL_SOURCE_MILOCLIENT_COMPONENT_OVERRIDE_HOST_DEFAULT = false;
@@ -253,6 +256,7 @@ public class CamelMiloclientSourceConnectorConfig
conf.define(CAMEL_SOURCE_MILOCLIENT_COMPONENT_KEY_STORE_URL_CONF,
ConfigDef.Type.STRING, CAMEL_SOURCE_MILOCLIENT_COMPONENT_KEY_STORE_URL_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_MILOCLIENT_COMPONENT_KEY_STORE_URL_DOC);
conf.define(CAMEL_SOURCE_MILOCLIENT_COMPONENT_MAX_PENDING_PUBLISH_REQUESTS_CONF,
ConfigDef.Type.STRING,
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MAX_PENDING_PUBLISH_REQUESTS_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MAX_PENDING_PUBLISH_REQUESTS_DOC);
conf.define(CAMEL_SOURCE_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_CONF,
ConfigDef.Type.STRING,
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MAX_RESPONSE_MESSAGE_SIZE_DOC);
+
conf.define(CAMEL_SOURCE_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_CONF,
ConfigDef.Type.STRING,
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_MILOCLIENT_COMPONENT_MILO_CLIENT_CONNECTION_MANAGER_DOC);
conf.define(CAMEL_SOURCE_MILOCLIENT_COMPONENT_OVERRIDE_HOST_CONF,
ConfigDef.Type.BOOLEAN,
CAMEL_SOURCE_MILOCLIENT_COMPONENT_OVERRIDE_HOST_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_MILOCLIENT_COMPONENT_OVERRIDE_HOST_DOC);
conf.define(CAMEL_SOURCE_MILOCLIENT_COMPONENT_PRODUCT_URI_CONF,
ConfigDef.Type.STRING, CAMEL_SOURCE_MILOCLIENT_COMPONENT_PRODUCT_URI_DEFAULT,
ConfigDef.Importance.MEDIUM, CAMEL_SOURCE_MILOCLIENT_COMPONENT_PRODUCT_URI_DOC);
conf.define(CAMEL_SOURCE_MILOCLIENT_COMPONENT_REQUESTED_PUBLISHING_INTERVAL_CONF,
ConfigDef.Type.STRING,
CAMEL_SOURCE_MILOCLIENT_COMPONENT_REQUESTED_PUBLISHING_INTERVAL_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_MILOCLIENT_COMPONENT_REQUESTED_PUBLISHING_INTERVAL_DOC);
diff --git
a/connectors/camel-vertx-kafka-kafka-connector/src/generated/resources/camel-vertx-kafka-sink.json
b/connectors/camel-vertx-kafka-kafka-connector/src/generated/resources/camel-vertx-kafka-sink.json
index 459b4d5..b0021be 100644
---
a/connectors/camel-vertx-kafka-kafka-connector/src/generated/resources/camel-vertx-kafka-sink.json
+++
b/connectors/camel-vertx-kafka-kafka-connector/src/generated/resources/camel-vertx-kafka-sink.json
@@ -296,12 +296,6 @@
"priority": "MEDIUM",
"required": "false"
},
- "camel.sink.endpoint.vertxKafkaClientFactory": {
- "name": "camel.sink.endpoint.vertxKafkaClientFactory",
- "description": "Factory to use for creating
io.vertx.kafka.client.consumer.KafkaConsumer and
io.vertx.kafka.client.consumer.KafkaProducer instances. This allows to
configure a custom factory to create custom KafkaConsumer and KafkaProducer
instances with logic that extends the vanilla VertX Kafka clients.",
- "priority": "MEDIUM",
- "required": "false"
- },
"camel.sink.endpoint.saslClientCallbackHandlerClass": {
"name":
"camel.sink.endpoint.saslClientCallbackHandlerClass",
"description": "The fully qualified name of a SASL
client callback handler class that implements the AuthenticateCallbackHandler
interface.",
diff --git
a/connectors/camel-vertx-kafka-kafka-connector/src/generated/resources/camel-vertx-kafka-source.json
b/connectors/camel-vertx-kafka-kafka-connector/src/generated/resources/camel-vertx-kafka-source.json
index f049e88..6294c79 100644
---
a/connectors/camel-vertx-kafka-kafka-connector/src/generated/resources/camel-vertx-kafka-source.json
+++
b/connectors/camel-vertx-kafka-kafka-connector/src/generated/resources/camel-vertx-kafka-source.json
@@ -365,12 +365,6 @@
"InOptionalOut"
]
},
- "camel.source.endpoint.vertxKafkaClientFactory": {
- "name": "camel.source.endpoint.vertxKafkaClientFactory",
- "description": "Factory to use for creating
io.vertx.kafka.client.consumer.KafkaConsumer and
io.vertx.kafka.client.consumer.KafkaProducer instances. This allows to
configure a custom factory to create custom KafkaConsumer and KafkaProducer
instances with logic that extends the vanilla VertX Kafka clients.",
- "priority": "MEDIUM",
- "required": "false"
- },
"camel.source.endpoint.saslClientCallbackHandlerClass": {
"name":
"camel.source.endpoint.saslClientCallbackHandlerClass",
"description": "The fully qualified name of a SASL
client callback handler class that implements the AuthenticateCallbackHandler
interface.",
diff --git
a/connectors/camel-vertx-kafka-kafka-connector/src/main/docs/camel-vertx-kafka-kafka-sink-connector.adoc
b/connectors/camel-vertx-kafka-kafka-connector/src/main/docs/camel-vertx-kafka-kafka-sink-connector.adoc
index 3dd4808..d6e756b 100644
---
a/connectors/camel-vertx-kafka-kafka-connector/src/main/docs/camel-vertx-kafka-kafka-sink-connector.adoc
+++
b/connectors/camel-vertx-kafka-kafka-connector/src/main/docs/camel-vertx-kafka-kafka-sink-connector.adoc
@@ -24,7 +24,7 @@
connector.class=org.apache.camel.kafkaconnector.vertxkafka.CamelVertxkafkaSinkCo
----
-The camel-vertx-kafka sink connector supports 155 options, which are listed
below.
+The camel-vertx-kafka sink connector supports 154 options, which are listed
below.
@@ -71,7 +71,6 @@ The camel-vertx-kafka sink connector supports 155 options,
which are listed belo
| *camel.sink.endpoint.transactionalId* | The TransactionalId to use for
transactional delivery. This enables reliability semantics which span multiple
producer sessions since it allows the client to guarantee that transactions
using the same TransactionalId have been completed prior to starting any new
transactions. If no TransactionalId is provided, then the producer is limited
to idempotent delivery. If a TransactionalId is configured, enable.idempotence
is implied. By default the Tra [...]
| *camel.sink.endpoint.transactionTimeoutMs* | The maximum amount of time in
ms that the transaction coordinator will wait for a transaction status update
from the producer before proactively aborting the ongoing transaction.If this
value is larger than the transaction.max.timeout.ms setting in the broker, the
request will fail with a InvalidTxnTimeoutException error. | 60000 | false |
MEDIUM
| *camel.sink.endpoint.valueSerializer* | Serializer class for value that
implements the org.apache.kafka.common.serialization.Serializer interface. |
"org.apache.kafka.common.serialization.StringSerializer" | false | MEDIUM
-| *camel.sink.endpoint.vertxKafkaClientFactory* | Factory to use for creating
io.vertx.kafka.client.consumer.KafkaConsumer and
io.vertx.kafka.client.consumer.KafkaProducer instances. This allows to
configure a custom factory to create custom KafkaConsumer and KafkaProducer
instances with logic that extends the vanilla VertX Kafka clients. | null |
false | MEDIUM
| *camel.sink.endpoint.saslClientCallbackHandlerClass* | The fully qualified
name of a SASL client callback handler class that implements the
AuthenticateCallbackHandler interface. | null | false | MEDIUM
| *camel.sink.endpoint.saslJaasConfig* | JAAS login context parameters for
SASL connections in the format used by JAAS configuration files. JAAS
configuration file format is described here. The format for the value is:
'loginModuleClass controlFlag (optionName=optionValue);'. For brokers, the
config must be prefixed with listener prefix and SASL mechanism name in
lower-case. For example,
listener.name.sasl_ssl.scram-sha-256.sasl.jaas.config=com.example.ScramLoginModule
required; | null | [...]
| *camel.sink.endpoint.saslKerberosKinitCmd* | Kerberos kinit command path. |
"/usr/bin/kinit" | false | MEDIUM
diff --git
a/connectors/camel-vertx-kafka-kafka-connector/src/main/docs/camel-vertx-kafka-kafka-source-connector.adoc
b/connectors/camel-vertx-kafka-kafka-connector/src/main/docs/camel-vertx-kafka-kafka-source-connector.adoc
index abd4742..b12a000 100644
---
a/connectors/camel-vertx-kafka-kafka-connector/src/main/docs/camel-vertx-kafka-kafka-source-connector.adoc
+++
b/connectors/camel-vertx-kafka-kafka-connector/src/main/docs/camel-vertx-kafka-kafka-source-connector.adoc
@@ -24,7 +24,7 @@
connector.class=org.apache.camel.kafkaconnector.vertxkafka.CamelVertxkafkaSource
----
-The camel-vertx-kafka source connector supports 171 options, which are listed
below.
+The camel-vertx-kafka source connector supports 170 options, which are listed
below.
@@ -80,7 +80,6 @@ The camel-vertx-kafka source connector supports 171 options,
which are listed be
| *camel.source.endpoint.valueDeserializer* | Deserializer class for value
that implements the org.apache.kafka.common.serialization.Deserializer
interface. | "org.apache.kafka.common.serialization.StringDeserializer" | false
| MEDIUM
| *camel.source.endpoint.exceptionHandler* | To let the consumer use a custom
ExceptionHandler. Notice if the option bridgeErrorHandler is enabled then this
option is not in use. By default the consumer will deal with exceptions, that
will be logged at WARN or ERROR level and ignored. | null | false | MEDIUM
| *camel.source.endpoint.exchangePattern* | Sets the exchange pattern when the
consumer creates an exchange. One of: [InOnly] [InOut] [InOptionalOut] | null |
false | MEDIUM
-| *camel.source.endpoint.vertxKafkaClientFactory* | Factory to use for
creating io.vertx.kafka.client.consumer.KafkaConsumer and
io.vertx.kafka.client.consumer.KafkaProducer instances. This allows to
configure a custom factory to create custom KafkaConsumer and KafkaProducer
instances with logic that extends the vanilla VertX Kafka clients. | null |
false | MEDIUM
| *camel.source.endpoint.saslClientCallbackHandler Class* | The fully
qualified name of a SASL client callback handler class that implements the
AuthenticateCallbackHandler interface. | null | false | MEDIUM
| *camel.source.endpoint.saslJaasConfig* | JAAS login context parameters for
SASL connections in the format used by JAAS configuration files. JAAS
configuration file format is described here. The format for the value is:
'loginModuleClass controlFlag (optionName=optionValue);'. For brokers, the
config must be prefixed with listener prefix and SASL mechanism name in
lower-case. For example,
listener.name.sasl_ssl.scram-sha-256.sasl.jaas.config=com.example.ScramLoginModule
required; | null [...]
| *camel.source.endpoint.saslKerberosKinitCmd* | Kerberos kinit command path.
| "/usr/bin/kinit" | false | MEDIUM
diff --git
a/connectors/camel-vertx-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/vertxkafka/CamelVertxkafkaSinkConnectorConfig.java
b/connectors/camel-vertx-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/vertxkafka/CamelVertxkafkaSinkConnectorConfig.java
index 5dbe5d5..65bdfbc 100644
---
a/connectors/camel-vertx-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/vertxkafka/CamelVertxkafkaSinkConnectorConfig.java
+++
b/connectors/camel-vertx-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/vertxkafka/CamelVertxkafkaSinkConnectorConfig.java
@@ -146,9 +146,6 @@ public class CamelVertxkafkaSinkConnectorConfig
public static final String
CAMEL_SINK_VERTXKAFKA_ENDPOINT_VALUE_SERIALIZER_CONF =
"camel.sink.endpoint.valueSerializer";
public static final String
CAMEL_SINK_VERTXKAFKA_ENDPOINT_VALUE_SERIALIZER_DOC = "Serializer class for
value that implements the org.apache.kafka.common.serialization.Serializer
interface.";
public static final String
CAMEL_SINK_VERTXKAFKA_ENDPOINT_VALUE_SERIALIZER_DEFAULT =
"org.apache.kafka.common.serialization.StringSerializer";
- public static final String
CAMEL_SINK_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_CONF =
"camel.sink.endpoint.vertxKafkaClientFactory";
- public static final String
CAMEL_SINK_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_DOC = "Factory to use
for creating io.vertx.kafka.client.consumer.KafkaConsumer and
io.vertx.kafka.client.consumer.KafkaProducer instances. This allows to
configure a custom factory to create custom KafkaConsumer and KafkaProducer
instances with logic that extends the vanilla VertX Kafka clients.";
- public static final String
CAMEL_SINK_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_DEFAULT = null;
public static final String
CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_CONF =
"camel.sink.endpoint.saslClientCallbackHandlerClass";
public static final String
CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_DOC = "The
fully qualified name of a SASL client callback handler class that implements
the AuthenticateCallbackHandler interface.";
public static final String
CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_DEFAULT =
null;
@@ -544,7 +541,6 @@ public class CamelVertxkafkaSinkConnectorConfig
conf.define(CAMEL_SINK_VERTXKAFKA_ENDPOINT_TRANSACTIONAL_ID_CONF,
ConfigDef.Type.STRING, CAMEL_SINK_VERTXKAFKA_ENDPOINT_TRANSACTIONAL_ID_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_VERTXKAFKA_ENDPOINT_TRANSACTIONAL_ID_DOC);
conf.define(CAMEL_SINK_VERTXKAFKA_ENDPOINT_TRANSACTION_TIMEOUT_MS_CONF,
ConfigDef.Type.INT,
CAMEL_SINK_VERTXKAFKA_ENDPOINT_TRANSACTION_TIMEOUT_MS_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_VERTXKAFKA_ENDPOINT_TRANSACTION_TIMEOUT_MS_DOC);
conf.define(CAMEL_SINK_VERTXKAFKA_ENDPOINT_VALUE_SERIALIZER_CONF,
ConfigDef.Type.STRING, CAMEL_SINK_VERTXKAFKA_ENDPOINT_VALUE_SERIALIZER_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_VERTXKAFKA_ENDPOINT_VALUE_SERIALIZER_DOC);
-
conf.define(CAMEL_SINK_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_CONF,
ConfigDef.Type.STRING,
CAMEL_SINK_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_DOC);
conf.define(CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_CONF,
ConfigDef.Type.STRING,
CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_DOC);
conf.define(CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_JAAS_CONFIG_CONF,
ConfigDef.Type.STRING, CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_JAAS_CONFIG_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_JAAS_CONFIG_DOC);
conf.define(CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_KERBEROS_KINIT_CMD_CONF,
ConfigDef.Type.STRING,
CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_KERBEROS_KINIT_CMD_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SINK_VERTXKAFKA_ENDPOINT_SASL_KERBEROS_KINIT_CMD_DOC);
diff --git
a/connectors/camel-vertx-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/vertxkafka/CamelVertxkafkaSourceConnectorConfig.java
b/connectors/camel-vertx-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/vertxkafka/CamelVertxkafkaSourceConnectorConfig.java
index f6a55eb..c91d7c6 100644
---
a/connectors/camel-vertx-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/vertxkafka/CamelVertxkafkaSourceConnectorConfig.java
+++
b/connectors/camel-vertx-kafka-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/vertxkafka/CamelVertxkafkaSourceConnectorConfig.java
@@ -173,9 +173,6 @@ public class CamelVertxkafkaSourceConnectorConfig
public static final String
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_EXCHANGE_PATTERN_CONF =
"camel.source.endpoint.exchangePattern";
public static final String
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_EXCHANGE_PATTERN_DOC = "Sets the exchange
pattern when the consumer creates an exchange. One of: [InOnly] [InOut]
[InOptionalOut]";
public static final String
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_EXCHANGE_PATTERN_DEFAULT = null;
- public static final String
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_CONF =
"camel.source.endpoint.vertxKafkaClientFactory";
- public static final String
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_DOC = "Factory to
use for creating io.vertx.kafka.client.consumer.KafkaConsumer and
io.vertx.kafka.client.consumer.KafkaProducer instances. This allows to
configure a custom factory to create custom KafkaConsumer and KafkaProducer
instances with logic that extends the vanilla VertX Kafka clients.";
- public static final String
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_DEFAULT = null;
public static final String
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_CONF =
"camel.source.endpoint.saslClientCallbackHandlerClass";
public static final String
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_DOC = "The
fully qualified name of a SASL client callback handler class that implements
the AuthenticateCallbackHandler interface.";
public static final String
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_DEFAULT =
null;
@@ -601,7 +598,6 @@ public class CamelVertxkafkaSourceConnectorConfig
conf.define(CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_VALUE_DESERIALIZER_CONF,
ConfigDef.Type.STRING,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_VALUE_DESERIALIZER_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_VALUE_DESERIALIZER_DOC);
conf.define(CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_EXCEPTION_HANDLER_CONF,
ConfigDef.Type.STRING,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_EXCEPTION_HANDLER_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_EXCEPTION_HANDLER_DOC);
conf.define(CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_EXCHANGE_PATTERN_CONF,
ConfigDef.Type.STRING,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_EXCHANGE_PATTERN_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_EXCHANGE_PATTERN_DOC);
-
conf.define(CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_CONF,
ConfigDef.Type.STRING,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_VERTX_KAFKA_CLIENT_FACTORY_DOC);
conf.define(CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_CONF,
ConfigDef.Type.STRING,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_CLIENT_CALLBACK_HANDLER_CLASS_DOC);
conf.define(CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_JAAS_CONFIG_CONF,
ConfigDef.Type.STRING,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_JAAS_CONFIG_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_JAAS_CONFIG_DOC);
conf.define(CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_KERBEROS_KINIT_CMD_CONF,
ConfigDef.Type.STRING,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_KERBEROS_KINIT_CMD_DEFAULT,
ConfigDef.Importance.MEDIUM,
CAMEL_SOURCE_VERTXKAFKA_ENDPOINT_SASL_KERBEROS_KINIT_CMD_DOC);
diff --git
a/docs/modules/ROOT/pages/connectors/camel-aws2-sqs-kafka-sink-connector.adoc
b/docs/modules/ROOT/pages/connectors/camel-aws2-sqs-kafka-sink-connector.adoc
index 1ed8652..d218f26 100644
---
a/docs/modules/ROOT/pages/connectors/camel-aws2-sqs-kafka-sink-connector.adoc
+++
b/docs/modules/ROOT/pages/connectors/camel-aws2-sqs-kafka-sink-connector.adoc
@@ -24,7 +24,7 @@
connector.class=org.apache.camel.kafkaconnector.aws2sqs.CamelAws2sqsSinkConnecto
----
-The camel-aws2-sqs sink connector supports 54 options, which are listed below.
+The camel-aws2-sqs sink connector supports 56 options, which are listed below.
@@ -42,6 +42,7 @@ The camel-aws2-sqs sink connector supports 54 options, which
are listed below.
| *camel.sink.endpoint.region* | The region in which SQS client needs to work.
When using this parameter, the configuration will expect the lowercase name of
the region (for example ap-east-1) You'll need to use the name
Region.EU_WEST_1.id() | null | false | MEDIUM
| *camel.sink.endpoint.trustAllCertificates* | If we want to trust all
certificates in case of overriding the endpoint | false | false | MEDIUM
| *camel.sink.endpoint.useDefaultCredentialsProvider* | Set whether the SQS
client should expect to load credentials on an AWS infra instance or to expect
static credentials to be passed in. | false | false | MEDIUM
+| *camel.sink.endpoint.batchSeparator* | Set the separator when passing a
String to send batch message operation | "," | false | MEDIUM
| *camel.sink.endpoint.delaySeconds* | Delay sending messages for a number of
seconds. | null | false | MEDIUM
| *camel.sink.endpoint.lazyStartProducer* | Whether the producer should be
started lazy (on the first message). By starting lazy you can use this to allow
CamelContext and routes to startup in situations where a producer may otherwise
fail during starting and cause the route to fail being started. By deferring
this startup to be lazy then the startup failure can be handled during routing
messages via Camel's routing error handlers. Beware that when the first message
is processed then cre [...]
| *camel.sink.endpoint.messageDeduplicationIdStrategy* | Only for FIFO queues.
Strategy for setting the messageDeduplicationId on the message. Can be one of
the following options: useExchangeId, useContentBasedDeduplication. For the
useContentBasedDeduplication option, no messageDeduplicationId will be set on
the message. One of: [useExchangeId] [useContentBasedDeduplication] |
"useExchangeId" | false | MEDIUM
@@ -68,6 +69,7 @@ The camel-aws2-sqs sink connector supports 54 options, which
are listed below.
| *camel.component.aws2-sqs.region* | The region in which SQS client needs to
work. When using this parameter, the configuration will expect the lowercase
name of the region (for example ap-east-1) You'll need to use the name
Region.EU_WEST_1.id() | null | false | MEDIUM
| *camel.component.aws2-sqs.trustAllCertificates* | If we want to trust all
certificates in case of overriding the endpoint | false | false | MEDIUM
| *camel.component.aws2-sqs.useDefaultCredentials Provider* | Set whether the
SQS client should expect to load credentials on an AWS infra instance or to
expect static credentials to be passed in. | false | false | MEDIUM
+| *camel.component.aws2-sqs.batchSeparator* | Set the separator when passing a
String to send batch message operation | "," | false | MEDIUM
| *camel.component.aws2-sqs.delaySeconds* | Delay sending messages for a
number of seconds. | null | false | MEDIUM
| *camel.component.aws2-sqs.lazyStartProducer* | Whether the producer should
be started lazy (on the first message). By starting lazy you can use this to
allow CamelContext and routes to startup in situations where a producer may
otherwise fail during starting and cause the route to fail being started. By
deferring this startup to be lazy then the startup failure can be handled
during routing messages via Camel's routing error handlers. Beware that when
the first message is processed the [...]
| *camel.component.aws2-sqs.messageDeduplicationId Strategy* | Only for FIFO
queues. Strategy for setting the messageDeduplicationId on the message. Can be
one of the following options: useExchangeId, useContentBasedDeduplication. For
the useContentBasedDeduplication option, no messageDeduplicationId will be set
on the message. One of: [useExchangeId] [useContentBasedDeduplication] |
"useExchangeId" | false | MEDIUM
diff --git
a/docs/modules/ROOT/pages/connectors/camel-kafka-kafka-sink-connector.adoc
b/docs/modules/ROOT/pages/connectors/camel-kafka-kafka-sink-connector.adoc
index ea98fa7..3e2b65e 100644
--- a/docs/modules/ROOT/pages/connectors/camel-kafka-kafka-sink-connector.adoc
+++ b/docs/modules/ROOT/pages/connectors/camel-kafka-kafka-sink-connector.adoc
@@ -137,7 +137,7 @@ The camel-kafka sink connector supports 135 options, which
are listed below.
| *camel.component.kafka.workerPoolCoreSize* | Number of core threads for the
worker pool for continue routing Exchange after kafka server has acknowledge
the message that was sent to it from KafkaProducer using asynchronous
non-blocking processing. | "10" | false | MEDIUM
| *camel.component.kafka.workerPoolMaxSize* | Maximum number of threads for
the worker pool for continue routing Exchange after kafka server has
acknowledge the message that was sent to it from KafkaProducer using
asynchronous non-blocking processing. | "20" | false | MEDIUM
| *camel.component.kafka.autowiredEnabled* | Whether autowiring is enabled.
This is used for automatic autowiring options (the option must be marked as
autowired) by looking up in the registry to find if there is a single instance
of matching type, which then gets configured on the component. This can be used
for automatic configuring JDBC data sources, JMS connection factories, AWS
Clients, etc. | true | false | MEDIUM
-| *camel.component.kafka.kafkaClientFactory* | Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances with logic that
extends the vanilla Kafka clients. | null | false | MEDIUM
+| *camel.component.kafka.kafkaClientFactory* | Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create instances with logic that extends the
vanilla Kafka clients. | null | false | MEDIUM
| *camel.component.kafka.synchronous* | Sets whether synchronous processing
should be strictly used | false | false | MEDIUM
| *camel.component.kafka.schemaRegistryURL* | URL of the Confluent Platform
schema registry servers to use. The format is host1:port1,host2:port2. This is
known as schema.registry.url in the Confluent Platform documentation. This
option is only available in the Confluent Platform (not standard Apache Kafka)
| null | false | MEDIUM
| *camel.component.kafka.interceptorClasses* | Sets interceptors for producer
or consumers. Producer interceptors have to be classes implementing
org.apache.kafka.clients.producer.ProducerInterceptor Consumer interceptors
have to be classes implementing
org.apache.kafka.clients.consumer.ConsumerInterceptor Note that if you use
Producer interceptor on a consumer it will throw a class cast exception in
runtime | null | false | MEDIUM
diff --git
a/docs/modules/ROOT/pages/connectors/camel-kafka-kafka-source-connector.adoc
b/docs/modules/ROOT/pages/connectors/camel-kafka-kafka-source-connector.adoc
index b41e14b..03f8c86 100644
--- a/docs/modules/ROOT/pages/connectors/camel-kafka-kafka-source-connector.adoc
+++ b/docs/modules/ROOT/pages/connectors/camel-kafka-kafka-source-connector.adoc
@@ -129,7 +129,7 @@ The camel-kafka source connector supports 122 options,
which are listed below.
| *camel.component.kafka.valueDeserializer* | Deserializer class for value
that implements the Deserializer interface. |
"org.apache.kafka.common.serialization.StringDeserializer" | false | MEDIUM
| *camel.component.kafka.kafkaManualCommitFactory* | Factory to use for
creating KafkaManualCommit instances. This allows to plugin a custom factory to
create custom KafkaManualCommit instances in case special logic is needed when
doing manual commits that deviates from the default implementation that comes
out of the box. | null | false | MEDIUM
| *camel.component.kafka.autowiredEnabled* | Whether autowiring is enabled.
This is used for automatic autowiring options (the option must be marked as
autowired) by looking up in the registry to find if there is a single instance
of matching type, which then gets configured on the component. This can be used
for automatic configuring JDBC data sources, JMS connection factories, AWS
Clients, etc. | true | false | MEDIUM
-| *camel.component.kafka.kafkaClientFactory* | Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances with logic that
extends the vanilla Kafka clients. | null | false | MEDIUM
+| *camel.component.kafka.kafkaClientFactory* | Factory to use for creating
org.apache.kafka.clients.consumer.KafkaConsumer and
org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
configure a custom factory to create instances with logic that extends the
vanilla Kafka clients. | null | false | MEDIUM
| *camel.component.kafka.synchronous* | Sets whether synchronous processing
should be strictly used | false | false | MEDIUM
| *camel.component.kafka.schemaRegistryURL* | URL of the Confluent Platform
schema registry servers to use. The format is host1:port1,host2:port2. This is
known as schema.registry.url in the Confluent Platform documentation. This
option is only available in the Confluent Platform (not standard Apache Kafka)
| null | false | MEDIUM
| *camel.component.kafka.interceptorClasses* | Sets interceptors for producer
or consumers. Producer interceptors have to be classes implementing
org.apache.kafka.clients.producer.ProducerInterceptor Consumer interceptors
have to be classes implementing
org.apache.kafka.clients.consumer.ConsumerInterceptor Note that if you use
Producer interceptor on a consumer it will throw a class cast exception in
runtime | null | false | MEDIUM
diff --git
a/docs/modules/ROOT/pages/connectors/camel-milo-client-kafka-sink-connector.adoc
b/docs/modules/ROOT/pages/connectors/camel-milo-client-kafka-sink-connector.adoc
index 53546d0..82650ab 100644
---
a/docs/modules/ROOT/pages/connectors/camel-milo-client-kafka-sink-connector.adoc
+++
b/docs/modules/ROOT/pages/connectors/camel-milo-client-kafka-sink-connector.adoc
@@ -24,7 +24,7 @@
connector.class=org.apache.camel.kafkaconnector.miloclient.CamelMiloclientSinkCo
----
-The camel-milo-client sink connector supports 53 options, which are listed
below.
+The camel-milo-client sink connector supports 54 options, which are listed
below.
@@ -78,6 +78,7 @@ The camel-milo-client sink connector supports 53 options,
which are listed below
| *camel.component.milo-client.keyStoreUrl* | The URL where the key should be
loaded from | null | false | MEDIUM
| *camel.component.milo-client.maxPendingPublish Requests* | The maximum
number of pending publish requests | null | false | MEDIUM
| *camel.component.milo-client.maxResponseMessageSize* | The maximum number of
bytes a response message may have | null | false | MEDIUM
+| *camel.component.milo-client.miloClientConnection Manager* | Instance for
managing client connections | null | false | MEDIUM
| *camel.component.milo-client.overrideHost* | Override the server reported
endpoint host with the host from the endpoint URI. | false | false | MEDIUM
| *camel.component.milo-client.productUri* | The product URI |
"http://camel.apache.org/EclipseMilo" | false | MEDIUM
| *camel.component.milo-client.requestedPublishing Interval* | The requested
publishing interval in milliseconds | "1_000.0" | false | MEDIUM
diff --git
a/docs/modules/ROOT/pages/connectors/camel-milo-client-kafka-source-connector.adoc
b/docs/modules/ROOT/pages/connectors/camel-milo-client-kafka-source-connector.adoc
index 1f1ece7..755b388 100644
---
a/docs/modules/ROOT/pages/connectors/camel-milo-client-kafka-source-connector.adoc
+++
b/docs/modules/ROOT/pages/connectors/camel-milo-client-kafka-source-connector.adoc
@@ -24,7 +24,7 @@
connector.class=org.apache.camel.kafkaconnector.miloclient.CamelMiloclientSource
----
-The camel-milo-client source connector supports 55 options, which are listed
below.
+The camel-milo-client source connector supports 56 options, which are listed
below.
@@ -80,6 +80,7 @@ The camel-milo-client source connector supports 55 options,
which are listed bel
| *camel.component.milo-client.keyStoreUrl* | The URL where the key should be
loaded from | null | false | MEDIUM
| *camel.component.milo-client.maxPendingPublish Requests* | The maximum
number of pending publish requests | null | false | MEDIUM
| *camel.component.milo-client.maxResponseMessageSize* | The maximum number of
bytes a response message may have | null | false | MEDIUM
+| *camel.component.milo-client.miloClientConnection Manager* | Instance for
managing client connections | null | false | MEDIUM
| *camel.component.milo-client.overrideHost* | Override the server reported
endpoint host with the host from the endpoint URI. | false | false | MEDIUM
| *camel.component.milo-client.productUri* | The product URI |
"http://camel.apache.org/EclipseMilo" | false | MEDIUM
| *camel.component.milo-client.requestedPublishing Interval* | The requested
publishing interval in milliseconds | "1_000.0" | false | MEDIUM
diff --git
a/docs/modules/ROOT/pages/connectors/camel-vertx-kafka-kafka-sink-connector.adoc
b/docs/modules/ROOT/pages/connectors/camel-vertx-kafka-kafka-sink-connector.adoc
index 3dd4808..d6e756b 100644
---
a/docs/modules/ROOT/pages/connectors/camel-vertx-kafka-kafka-sink-connector.adoc
+++
b/docs/modules/ROOT/pages/connectors/camel-vertx-kafka-kafka-sink-connector.adoc
@@ -24,7 +24,7 @@
connector.class=org.apache.camel.kafkaconnector.vertxkafka.CamelVertxkafkaSinkCo
----
-The camel-vertx-kafka sink connector supports 155 options, which are listed
below.
+The camel-vertx-kafka sink connector supports 154 options, which are listed
below.
@@ -71,7 +71,6 @@ The camel-vertx-kafka sink connector supports 155 options,
which are listed belo
| *camel.sink.endpoint.transactionalId* | The TransactionalId to use for
transactional delivery. This enables reliability semantics which span multiple
producer sessions since it allows the client to guarantee that transactions
using the same TransactionalId have been completed prior to starting any new
transactions. If no TransactionalId is provided, then the producer is limited
to idempotent delivery. If a TransactionalId is configured, enable.idempotence
is implied. By default the Tra [...]
| *camel.sink.endpoint.transactionTimeoutMs* | The maximum amount of time in
ms that the transaction coordinator will wait for a transaction status update
from the producer before proactively aborting the ongoing transaction.If this
value is larger than the transaction.max.timeout.ms setting in the broker, the
request will fail with a InvalidTxnTimeoutException error. | 60000 | false |
MEDIUM
| *camel.sink.endpoint.valueSerializer* | Serializer class for value that
implements the org.apache.kafka.common.serialization.Serializer interface. |
"org.apache.kafka.common.serialization.StringSerializer" | false | MEDIUM
-| *camel.sink.endpoint.vertxKafkaClientFactory* | Factory to use for creating
io.vertx.kafka.client.consumer.KafkaConsumer and
io.vertx.kafka.client.consumer.KafkaProducer instances. This allows to
configure a custom factory to create custom KafkaConsumer and KafkaProducer
instances with logic that extends the vanilla VertX Kafka clients. | null |
false | MEDIUM
| *camel.sink.endpoint.saslClientCallbackHandlerClass* | The fully qualified
name of a SASL client callback handler class that implements the
AuthenticateCallbackHandler interface. | null | false | MEDIUM
| *camel.sink.endpoint.saslJaasConfig* | JAAS login context parameters for
SASL connections in the format used by JAAS configuration files. JAAS
configuration file format is described here. The format for the value is:
'loginModuleClass controlFlag (optionName=optionValue);'. For brokers, the
config must be prefixed with listener prefix and SASL mechanism name in
lower-case. For example,
listener.name.sasl_ssl.scram-sha-256.sasl.jaas.config=com.example.ScramLoginModule
required; | null | [...]
| *camel.sink.endpoint.saslKerberosKinitCmd* | Kerberos kinit command path. |
"/usr/bin/kinit" | false | MEDIUM
diff --git
a/docs/modules/ROOT/pages/connectors/camel-vertx-kafka-kafka-source-connector.adoc
b/docs/modules/ROOT/pages/connectors/camel-vertx-kafka-kafka-source-connector.adoc
index abd4742..b12a000 100644
---
a/docs/modules/ROOT/pages/connectors/camel-vertx-kafka-kafka-source-connector.adoc
+++
b/docs/modules/ROOT/pages/connectors/camel-vertx-kafka-kafka-source-connector.adoc
@@ -24,7 +24,7 @@
connector.class=org.apache.camel.kafkaconnector.vertxkafka.CamelVertxkafkaSource
----
-The camel-vertx-kafka source connector supports 171 options, which are listed
below.
+The camel-vertx-kafka source connector supports 170 options, which are listed
below.
@@ -80,7 +80,6 @@ The camel-vertx-kafka source connector supports 171 options,
which are listed be
| *camel.source.endpoint.valueDeserializer* | Deserializer class for value
that implements the org.apache.kafka.common.serialization.Deserializer
interface. | "org.apache.kafka.common.serialization.StringDeserializer" | false
| MEDIUM
| *camel.source.endpoint.exceptionHandler* | To let the consumer use a custom
ExceptionHandler. Notice if the option bridgeErrorHandler is enabled then this
option is not in use. By default the consumer will deal with exceptions, that
will be logged at WARN or ERROR level and ignored. | null | false | MEDIUM
| *camel.source.endpoint.exchangePattern* | Sets the exchange pattern when the
consumer creates an exchange. One of: [InOnly] [InOut] [InOptionalOut] | null |
false | MEDIUM
-| *camel.source.endpoint.vertxKafkaClientFactory* | Factory to use for
creating io.vertx.kafka.client.consumer.KafkaConsumer and
io.vertx.kafka.client.consumer.KafkaProducer instances. This allows to
configure a custom factory to create custom KafkaConsumer and KafkaProducer
instances with logic that extends the vanilla VertX Kafka clients. | null |
false | MEDIUM
| *camel.source.endpoint.saslClientCallbackHandler Class* | The fully
qualified name of a SASL client callback handler class that implements the
AuthenticateCallbackHandler interface. | null | false | MEDIUM
| *camel.source.endpoint.saslJaasConfig* | JAAS login context parameters for
SASL connections in the format used by JAAS configuration files. JAAS
configuration file format is described here. The format for the value is:
'loginModuleClass controlFlag (optionName=optionValue);'. For brokers, the
config must be prefixed with listener prefix and SASL mechanism name in
lower-case. For example,
listener.name.sasl_ssl.scram-sha-256.sasl.jaas.config=com.example.ScramLoginModule
required; | null [...]
| *camel.source.endpoint.saslKerberosKinitCmd* | Kerberos kinit command path.
| "/usr/bin/kinit" | false | MEDIUM