Here is the code which consuming messages

>>>>>>>>
while(true && startShutdown == false) {
    Context context = new Context();
    JSONObject notifJSON = new JSONObject();
    String notificationMsg = "";
    NotificationEvent notifEvent = null;
    initializeContext();
    try {
        consumerConnect();
        ConsumerRecords<String, String> records = consumer.poll(100);
        if(records.count() == 0) {
            //logger.trace("No records in topic: "+this.topic);
            continue;
        }
        for(ConsumerRecord<String, String> record : records) {
            try {
                long totalStart = System.currentTimeMillis();
                notificationMsg = record.value();
                JSONParser jsonParser = new JSONParser();
                logger.trace("Kafka-Msg: >>"+notificationMsg);
                if(notificationMsg.equals("")) {
                    continue;
                }
                Profiler.start(workerId, "json-parse");
                notifJSON           =
(JSONObject)jsonParser.parse(notificationMsg);
                Profiler.end(workerId, "json-parse");
                notifEvent    = new NotificationEvent(notifJSON);
                if( notifEvent.getTransactionID().equals("") == true ) {
                    notifEvent.generateTransactionID();
                }
                context.setEventObject(notifEvent);
                updateContext(context);

                //======== Fetch template ==========//
                Profiler.start(workerId, "tpl-fetch");
                long start = System.currentTimeMillis();
                Template template   =
notifTplMngr.fetchTemplate(notifEvent);

logger.trace("fetch-tpl:"+(System.currentTimeMillis()-start));
                Profiler.end(workerId, "tpl-fetch");

                //======== Personalise template ==========//
                Profiler.start(workerId, "personalisation");
                start = System.currentTimeMillis();
                String message      =
NotificationTemplatePersonaliser.personaliseAuto(template, notifEvent);

notifEvent.setMaskedMessage(NotificationTemplatePersonaliser.getMaskedContent(template,
notifEvent));

logger.trace("personalise:"+(System.currentTimeMillis()-start));
                Profiler.end(workerId, "personalisation");

                context.setEventObject(notifEvent);
                updateContext(context);

                //======== Send notification==========//
                Profiler.start(workerId, "notif-dispatch");
                postOffice.sendNotification(message, notifEvent);
                Profiler.end(workerId, "notif-dispatch");

                retryCount = 0;
                logger.debug("Time to complete notification dispatch
:"+(System.currentTimeMillis()-totalStart));
                if(startShutdown == true) {
                    break;
                }
            } catch (Exception ex) {
                if(ex instanceof RetriableException) {
                    kafkaLogger.error(ex);
                    logger.warn("",ex);
                    addToFailedQueue(notifJSON, ex.getMessage(),
CODE_RETRIABLE_FAILURE);
                } else if(ex instanceof InvalidEventException) {

                    JsonLog jsonLog = new JsonLog();
                    jsonLog.setDescription("Invalid event message. Reason:
"+ex.getMessage());
                    jsonLog.setOriginalPayload(notificationMsg);
                    jsonLog.setEventType("ERROR");
                    jsonLog.setCode("InvalidEventException");
                    jsonLog.setComponent(kafkaLogger.getSourceClass(ex));
                    jsonLog.setSubComponent(notifEvent.getChannelName());
                    kafkaLogger.log(jsonLog);
                    //kafkaLogger.error(ex);
                    addToFailedQueue(notifJSON, ex.getMessage(),
CODE_PERMANENT_FAILURE);
                    logger.warn("Invalid event message. Reason:
"+ex.getMessage());

                } else if(ex instanceof EventFailedException) {
                    addToFailedQueue(notifJSON, ex.getMessage(),
CODE_PERMANENT_FAILURE);
                    kafkaLogger.error(ex);
                    logger.warn("Notification event failed. Reason:
"+ex.getMessage());

                } else if(ex instanceof
org.json.simple.parser.ParseException) {
                    kafkaLogger.error("Exception while parsing notification
JSON message.");
                    logger.warn("Exception while parsing notification JSON
message.");
                } else {
                    kafkaLogger.error(ex);
                    addToFailedQueue(notifJSON, ex.getMessage(),
CODE_PERMANENT_FAILURE);
                    logger.warn("",ex);
                }
            } finally {
                eventsProcessed++;
            }
        }
    } catch (Exception ex) {
        kafkaLogger.error(ex);
        addToFailedQueue(notifJSON, ex.getMessage(),
CODE_PERMANENT_FAILURE);
        logger.warn("",ex);
    }
}
<<<<<<<<<<

And here are server properties.

broker.id=0
port=9092
delete.topic.enable=true
message.max.bytes=1500000
listeners=SSL://x.x.x.x:9092
advertised.listeners=SSL://x.x.x.x:9092
num.network.threads=3
num.io.threads=8
socket.send.buffer.bytes=102400
socket.receive.buffer.bytes=102400
socket.request.max.bytes=104857600
log.dirs=/lotus/kafka-logs
num.partitions=3
auto.topic.creation.enable=false
num.recovery.threads.per.data.dir=1
log.retention.hours=168
log.segment.bytes=1073741824
log.retention.check.interval.ms=300000
ssl.keystore.location=/opt/kafka/certificates/kafka.keystore.jks
ssl.keystore.password=xxxx
ssl.key.password=xxxx
ssl.truststore.location=/opt/kafka/certificates/kafka.truststore.jks
ssl.truststore.password=xxxx
security.inter.broker.protocol=SSL
zookeeper.connect=x.x.x.x:2181,x.x.x.x:2181,x.x.x.x:2181
zookeeper.connection.timeout.ms=6000

On Tue, May 29, 2018 at 5:59 PM M. Manna <manme...@gmail.com> wrote:

> Thanks..
>
> Where is your consumer code that is consuming messages?
>
> On 29 May 2018 at 13:18, Shantanu Deshmukh <shantanu...@gmail.com> wrote:
>
> > No problem, here are consumer properties
> > ---------
> > auto.commit.interval.ms = 3000
> > auto.offset.reset = latest
> > bootstrap.servers = [x.x.x.x:9092, x.x.x.x:9092, x.x.x.x:9092]
> > check.crcs = true
> > client.id =
> > connections.max.idle.ms = 540000
> > enable.auto.commit = true
> > exclude.internal.topics = true
> > fetch.max.bytes = 52428800
> > fetch.max.wait.ms = 500
> > fetch.min.bytes = 1
> > group.id = otp-notifications-consumer
> > heartbeat.interval.ms = 3000
> > interceptor.classes = null
> > key.deserializer = class
> > org.apache.kafka.common.serialization.StringDeserializer
> > max.partition.fetch.bytes = 1048576
> > max.poll.interval.ms = 300000
> > max.poll.records = 5
> > metadata.max.age.ms = 300000
> > metric.reporters = []
> > metrics.num.samples = 2
> > metrics.sample.window.ms = 30000
> > partition.assignment.strategy = [class
> > org.apache.kafka.clients.consumer.RangeAssignor]
> > receive.buffer.bytes = 65536
> > reconnect.backoff.ms = 50
> > request.timeout.ms = 305000
> > retry.backoff.ms = 100
> > sasl.kerberos.kinit.cmd = /usr/bin/kinit
> > sasl.kerberos.min.time.before.relogin = 60000
> > sasl.kerberos.service.name = null
> > sasl.kerberos.ticket.renew.jitter = 0.05
> > sasl.kerberos.ticket.renew.window.factor = 0.8
> > sasl.mechanism = GSSAPI
> > security.protocol = SSL
> > send.buffer.bytes = 131072
> > session.timeout.ms = 300000
> > ssl.cipher.suites = null
> > ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
> > ssl.endpoint.identification.algorithm = null
> > ssl.key.password = null
> > ssl.keymanager.algorithm = SunX509
> > ssl.keystore.location = null
> > ssl.keystore.password = null
> > ssl.keystore.type = JKS
> > ssl.protocol = TLS
> > ssl.provider = null
> > ssl.secure.random.implementation = null
> > ssl.trustmanager.algorithm = PKIX
> > ssl.truststore.location = ****
> > ssl.truststore.password = [hidden]
> > ssl.truststore.type = JKS
> > value.deserializer = class
> > org.apache.kafka.common.serialization.StringDeserializer
> > ------------
> >
> > On Tue, May 29, 2018 at 5:36 PM M. Manna <manme...@gmail.com> wrote:
> >
> > > Hi,
> > >
> > > It's not possible to answer questions based on text. You need to share
> > your
> > > consumer.properties, and server.properties file, and also, what exactly
> > you
> > > have changed from default configuration.
> > >
> > >
> > >
> > > On 29 May 2018 at 12:51, Shantanu Deshmukh <shantanu...@gmail.com>
> > wrote:
> > >
> > > > Hello,
> > > >
> > > > We have 3 broker Kafka 0.10.0.1 cluster. We have 5 topics, each with
> 10
> > > > partitions. I have an application which consumes from all these
> topics
> > by
> > > > creating multiple consumer processes. All of these consumers are
> under
> > a
> > > > same consumer group. I am noticing that every time we restart this
> > > > application. It takes almost 5 minutes for consumers to start
> > consuming.
> > > > What might be going wrong?
> > > >
> > >
> >
>

Reply via email to