Have you tried increase the poll time higher, e.g. 4000 and see if that helps matters?
On 29 May 2018 at 13:44, Shantanu Deshmukh <shantanu...@gmail.com> wrote: > Here is the code which consuming messages > > >>>>>>>> > while(true && startShutdown == false) { > Context context = new Context(); > JSONObject notifJSON = new JSONObject(); > String notificationMsg = ""; > NotificationEvent notifEvent = null; > initializeContext(); > try { > consumerConnect(); > ConsumerRecords<String, String> records = consumer.poll(100); > if(records.count() == 0) { > //logger.trace("No records in topic: "+this.topic); > continue; > } > for(ConsumerRecord<String, String> record : records) { > try { > long totalStart = System.currentTimeMillis(); > notificationMsg = record.value(); > JSONParser jsonParser = new JSONParser(); > logger.trace("Kafka-Msg: >>"+notificationMsg); > if(notificationMsg.equals("")) { > continue; > } > Profiler.start(workerId, "json-parse"); > notifJSON = > (JSONObject)jsonParser.parse(notificationMsg); > Profiler.end(workerId, "json-parse"); > notifEvent = new NotificationEvent(notifJSON); > if( notifEvent.getTransactionID().equals("") == true ) { > notifEvent.generateTransactionID(); > } > context.setEventObject(notifEvent); > updateContext(context); > > //======== Fetch template ==========// > Profiler.start(workerId, "tpl-fetch"); > long start = System.currentTimeMillis(); > Template template = > notifTplMngr.fetchTemplate(notifEvent); > > logger.trace("fetch-tpl:"+(System.currentTimeMillis()-start)); > Profiler.end(workerId, "tpl-fetch"); > > //======== Personalise template ==========// > Profiler.start(workerId, "personalisation"); > start = System.currentTimeMillis(); > String message = > NotificationTemplatePersonaliser.personaliseAuto(template, notifEvent); > > notifEvent.setMaskedMessage(NotificationTemplatePersonalis > er.getMaskedContent(template, > notifEvent)); > > logger.trace("personalise:"+(System.currentTimeMillis()-start)); > Profiler.end(workerId, "personalisation"); > > context.setEventObject(notifEvent); > updateContext(context); > > //======== Send notification==========// > Profiler.start(workerId, "notif-dispatch"); > postOffice.sendNotification(message, notifEvent); > Profiler.end(workerId, "notif-dispatch"); > > retryCount = 0; > logger.debug("Time to complete notification dispatch > :"+(System.currentTimeMillis()-totalStart)); > if(startShutdown == true) { > break; > } > } catch (Exception ex) { > if(ex instanceof RetriableException) { > kafkaLogger.error(ex); > logger.warn("",ex); > addToFailedQueue(notifJSON, ex.getMessage(), > CODE_RETRIABLE_FAILURE); > } else if(ex instanceof InvalidEventException) { > > JsonLog jsonLog = new JsonLog(); > jsonLog.setDescription("Invalid event message. Reason: > "+ex.getMessage()); > jsonLog.setOriginalPayload(notificationMsg); > jsonLog.setEventType("ERROR"); > jsonLog.setCode("InvalidEventException"); > jsonLog.setComponent(kafkaLogger.getSourceClass(ex)); > jsonLog.setSubComponent(notifEvent.getChannelName()); > kafkaLogger.log(jsonLog); > //kafkaLogger.error(ex); > addToFailedQueue(notifJSON, ex.getMessage(), > CODE_PERMANENT_FAILURE); > logger.warn("Invalid event message. Reason: > "+ex.getMessage()); > > } else if(ex instanceof EventFailedException) { > addToFailedQueue(notifJSON, ex.getMessage(), > CODE_PERMANENT_FAILURE); > kafkaLogger.error(ex); > logger.warn("Notification event failed. Reason: > "+ex.getMessage()); > > } else if(ex instanceof > org.json.simple.parser.ParseException) { > kafkaLogger.error("Exception while parsing notification > JSON message."); > logger.warn("Exception while parsing notification JSON > message."); > } else { > kafkaLogger.error(ex); > addToFailedQueue(notifJSON, ex.getMessage(), > CODE_PERMANENT_FAILURE); > logger.warn("",ex); > } > } finally { > eventsProcessed++; > } > } > } catch (Exception ex) { > kafkaLogger.error(ex); > addToFailedQueue(notifJSON, ex.getMessage(), > CODE_PERMANENT_FAILURE); > logger.warn("",ex); > } > } > <<<<<<<<<< > > And here are server properties. > > broker.id=0 > port=9092 > delete.topic.enable=true > message.max.bytes=1500000 > listeners=SSL://x.x.x.x:9092 > advertised.listeners=SSL://x.x.x.x:9092 > num.network.threads=3 > num.io.threads=8 > socket.send.buffer.bytes=102400 > socket.receive.buffer.bytes=102400 > socket.request.max.bytes=104857600 > log.dirs=/lotus/kafka-logs > num.partitions=3 > auto.topic.creation.enable=false > num.recovery.threads.per.data.dir=1 > log.retention.hours=168 > log.segment.bytes=1073741824 > log.retention.check.interval.ms=300000 > ssl.keystore.location=/opt/kafka/certificates/kafka.keystore.jks > ssl.keystore.password=xxxx > ssl.key.password=xxxx > ssl.truststore.location=/opt/kafka/certificates/kafka.truststore.jks > ssl.truststore.password=xxxx > security.inter.broker.protocol=SSL > zookeeper.connect=x.x.x.x:2181,x.x.x.x:2181,x.x.x.x:2181 > zookeeper.connection.timeout.ms=6000 > > On Tue, May 29, 2018 at 5:59 PM M. Manna <manme...@gmail.com> wrote: > > > Thanks.. > > > > Where is your consumer code that is consuming messages? > > > > On 29 May 2018 at 13:18, Shantanu Deshmukh <shantanu...@gmail.com> > wrote: > > > > > No problem, here are consumer properties > > > --------- > > > auto.commit.interval.ms = 3000 > > > auto.offset.reset = latest > > > bootstrap.servers = [x.x.x.x:9092, x.x.x.x:9092, x.x.x.x:9092] > > > check.crcs = true > > > client.id = > > > connections.max.idle.ms = 540000 > > > enable.auto.commit = true > > > exclude.internal.topics = true > > > fetch.max.bytes = 52428800 > > > fetch.max.wait.ms = 500 > > > fetch.min.bytes = 1 > > > group.id = otp-notifications-consumer > > > heartbeat.interval.ms = 3000 > > > interceptor.classes = null > > > key.deserializer = class > > > org.apache.kafka.common.serialization.StringDeserializer > > > max.partition.fetch.bytes = 1048576 > > > max.poll.interval.ms = 300000 > > > max.poll.records = 5 > > > metadata.max.age.ms = 300000 > > > metric.reporters = [] > > > metrics.num.samples = 2 > > > metrics.sample.window.ms = 30000 > > > partition.assignment.strategy = [class > > > org.apache.kafka.clients.consumer.RangeAssignor] > > > receive.buffer.bytes = 65536 > > > reconnect.backoff.ms = 50 > > > request.timeout.ms = 305000 > > > retry.backoff.ms = 100 > > > sasl.kerberos.kinit.cmd = /usr/bin/kinit > > > sasl.kerberos.min.time.before.relogin = 60000 > > > sasl.kerberos.service.name = null > > > sasl.kerberos.ticket.renew.jitter = 0.05 > > > sasl.kerberos.ticket.renew.window.factor = 0.8 > > > sasl.mechanism = GSSAPI > > > security.protocol = SSL > > > send.buffer.bytes = 131072 > > > session.timeout.ms = 300000 > > > ssl.cipher.suites = null > > > ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] > > > ssl.endpoint.identification.algorithm = null > > > ssl.key.password = null > > > ssl.keymanager.algorithm = SunX509 > > > ssl.keystore.location = null > > > ssl.keystore.password = null > > > ssl.keystore.type = JKS > > > ssl.protocol = TLS > > > ssl.provider = null > > > ssl.secure.random.implementation = null > > > ssl.trustmanager.algorithm = PKIX > > > ssl.truststore.location = **** > > > ssl.truststore.password = [hidden] > > > ssl.truststore.type = JKS > > > value.deserializer = class > > > org.apache.kafka.common.serialization.StringDeserializer > > > ------------ > > > > > > On Tue, May 29, 2018 at 5:36 PM M. Manna <manme...@gmail.com> wrote: > > > > > > > Hi, > > > > > > > > It's not possible to answer questions based on text. You need to > share > > > your > > > > consumer.properties, and server.properties file, and also, what > exactly > > > you > > > > have changed from default configuration. > > > > > > > > > > > > > > > > On 29 May 2018 at 12:51, Shantanu Deshmukh <shantanu...@gmail.com> > > > wrote: > > > > > > > > > Hello, > > > > > > > > > > We have 3 broker Kafka 0.10.0.1 cluster. We have 5 topics, each > with > > 10 > > > > > partitions. I have an application which consumes from all these > > topics > > > by > > > > > creating multiple consumer processes. All of these consumers are > > under > > > a > > > > > same consumer group. I am noticing that every time we restart this > > > > > application. It takes almost 5 minutes for consumers to start > > > consuming. > > > > > What might be going wrong? > > > > > > > > > > > > > > >