Hi Konstantine,

I appreciate you taking the time to respond
So I have set CONNECT_LOG4J_ROOT_LEVEL=INFO and that’s the output I got
below
Now I understand I need to set CONNECT_LOG4J_LOGGERS also. Can I please
have an example of how to set that value to suppress some debug statements?

For Example, I tried
CONNECT_LOG4_LOGGERS="org.reflections=INFO,org.apache.kafka=INFO” and yet
I’m still seeing all the DEBUG statements… like
22:34:06.444 [CLASSPATH traversal thread.] DEBUG
org.reflections.Reflections - could not scan file
unit/kafka/producer/ProducerTest.scala in url
file:/usr/bin/../share/java/kafka/kafka_2.11-0.10.1.0-cp2-test-sources.jar
with scanner SubTypesScanner

And it seems the bootstrap did take the variables into account as they get
written successfully:
root@6da04b77c18e:/# cat /etc/kafka/connect-log4j.properties

log4j.rootLogger=INFO, stdout

log4j.appender.stdout=org.apache.log4j.ConsoleAppender
log4j.appender.stdout.layout=org.apache.log4j.PatternLayout
log4j.appender.stdout.layout.ConversionPattern=[%d] %p %m (%c)%n

log4j.logger.org.reflections=INFO, stdout
log4j.logger.org.apache.kafka=INFO, stdout

I’m new to LOG4J properties so thanks for your help.

Regards,
Stephane


On 18 January 2017 at 8:06:16 am, Konstantine Karantasis (
konstant...@confluent.io) wrote:

Hi Stephane,

if you are using the docker images from confluent, a way to set the levels
to specific loggers is described here:

http://docs.confluent.io/3.1.1/cp-docker-images/docs/operations/logging.html#log4j-log-levels

For Connect, you would need to set the environment variable
CONNECT_LOG4J_LOGGERS in a similar way that KAFKA_LOG4J_LOGGERS is set in
the "docker run" command described above.

Regarding the redirection to stdout, if you are using Docker this is not
configurable with the current templates because this allows you to view the
logs for each container directly through docker via the command "docker
logs <ContainerName>", which is the preferred way.

Hope this helps,
Konstantine


On Mon, Jan 16, 2017 at 9:51 PM, Stephane Maarek <
steph...@simplemachines.com.au> wrote:

> The kind of output is the following:
>
> 05:15:34.878 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name connections-closed:
> 05:15:34.879 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name connections-created:
> 05:15:34.880 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name bytes-sent-received:
> 05:15:34.881 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name bytes-sent:
> 05:15:34.882 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name bytes-received:
> 05:15:34.882 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name select-time:
> 05:15:34.884 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name io-time:
> 05:15:34.905 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name heartbeat-latency
> 05:15:34.906 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name join-latency
> 05:15:34.907 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name sync-latency
> 05:15:34.970 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> connections-closed:
> 05:15:34.971 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> connections-created:
> 05:15:34.971 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> bytes-sent-received:
> 05:15:34.972 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> bytes-sent:
> 05:15:34.975 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> bytes-received:
> 05:15:34.977 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> select-time:
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-39
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
_connect_offsets-6
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-35
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
_connect_offsets-2
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-31
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-26
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-22
> 05:15:35.991 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-18
> 05:46:58.401 [CLASSPATH traversal thread.] DEBUG
> org.reflections.Reflections - could not scan file
> groovy/ui/icons/page_copy.png in url
> file:/usr/share/java/kafka-connect-hdfs/groovy-all-2.1.6.jar with scanner
> TypeAnnotationsScanner
> 05:46:58.401 [CLASSPATH traversal thread.] DEBUG
> org.reflections.Reflections - could not scan file
> groovy/ui/icons/page_copy.png in url
> file:/usr/share/java/kafka-connect-hdfs/groovy-all-2.1.6.jar with scanner
> SubTypesScanner
>
>
> *How do I stop all these loggers?*
>
> That’s what my connect-log4j.properties looks like:
>
>
> log4j.rootLogger=INFO, stdout
>
> log4j.appender.stdout=org.apache.log4j.ConsoleAppender
> log4j.appender.stdout.layout=org.apache.log4j.PatternLayout
> log4j.appender.stdout.layout.ConversionPattern=[%d] %p %m (%c)%n
>
> log4j.logger.org.apache.kafka.clients.consumer=INFO, stdout
>
> I’m surprise because I set INFO at the rootLogger and it doesn’t seem to
be
> taken into account
>
>
>
>
> On 16 January 2017 at 7:01:50 pm, Stephane Maarek (
> steph...@simplemachines.com.au) wrote:
>
>
> Hi,
>
> I created my own connector and I’m launching it in cluster mode, but
every
> DEBUG statement is still going to the console.
> How can I control the log level of Kafka Connect and its associated
> connectors? I’m using the confluent docker image btw
>
> Thanks
> Stephane
>

Reply via email to