Hi, my Topology description..

- Topologies:
   Sub-topology: 0
    Source: SeSource (topics: [buzzard.decoded.amon.stats.ap,
buzzard.app.monitoring.internal.clientsestats])
      --> SeProcess
    Processor: SeProcess (stores: [BucketData, CacheData])
      --> AeSink
      <-- SeSource
    Sink: AeSink (topic: buzzard.app.monitoring.internal.stats)
      <-- SeProcess

  Sub-topology: 1
    Source: StationSource (topics: [buzzard.decoded.amon.stats.station])
      --> StationProcess
    Processor: StationProcess (stores: [StationKeyValue, StationCacheData])
      --> StationSink
      <-- StationSource
    Sink: StationSink (topic: buzzard.app.monitoring.internal.clientsestats)
      <-- StationProcess

Regards
Ashik

On Thu, Jul 18, 2019 at 1:31 AM Sophie Blee-Goldman <sop...@confluent.io>
wrote:

> Hm. These directories shouldn't be created if using only an in memory
> store. Can you print your topology?
>
> On Wed, Jul 17, 2019 at 11:02 AM Muhammed Ashik <ashi...@gmail.com> wrote:
>
> > Hi I just did `du -mh` on `\tmp\kafka-streams` below are the folders
> listed
> > with some .lock files inside.
> > not sure if these are coming from rocksdb.. and looks like the sizes of
> > these files are less.
> >
> > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/0_5
> > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/0_3
> > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/0_2
> > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/1_4
> > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/0_1
> > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/0_4
> > 28K ./buzzard.MonitoringSeWlanStatsAggregator
> > 32K .
> >
> >
> >
> > On Wed, Jul 17, 2019 at 11:18 PM Sophie Blee-Goldman <
> sop...@confluent.io>
> > wrote:
> >
> > > Hm. Just to be absolutely sure, could you try throwing an exception or
> > > something in your RocksDBConfigSetter?
> > >
> > > On Wed, Jul 17, 2019 at 10:43 AM Muhammed Ashik <ashi...@gmail.com>
> > wrote:
> > >
> > > > I can confirm the /tmp/kafka-streams doesn't have any data related to
> > > > rocksdb.
> > > >
> > > > On Wed, Jul 17, 2019 at 11:11 PM Sophie Blee-Goldman <
> > > sop...@confluent.io>
> > > > wrote:
> > > >
> > > > > You can describe your topology to see if there are any state stores
> > in
> > > it
> > > > > that you aren't aware of. Alternatively you could check out the
> state
> > > > > directory (/tmp/kafka-streams by default) and see if there is any
> > data
> > > in
> > > > > there
> > > > >
> > > > > On Wed, Jul 17, 2019 at 10:36 AM Muhammed Ashik <ashi...@gmail.com
> >
> > > > wrote:
> > > > >
> > > > > > Thanks How can I verify If there is some data really going on
> > rocksdb
> > > > > > I tried printing the statistics with no success.
> > > > > >
> > > > > > class CustomRocksDBConfig extends RocksDBConfigSetter {
> > > > > >   override def setConfig(storeName: String, options: Options,
> > > configs:
> > > > > > util.Map[String, AnyRef]): Unit = {
> > > > > >
> > > > > >     val stats = new Statistics
> > > > > >     stats.setStatsLevel(StatsLevel.ALL)
> > > > > >     options.setStatistics(stats)
> > > > > >       .setStatsDumpPeriodSec(600)
> > > > > >     options
> > > > > >       .setInfoLogLevel(InfoLogLevel.INFO_LEVEL)
> > > > > >     options.setDbLogDir("/tmp/dump")
> > > > > >
> > > > > >   }
> > > > > > }
> > > > > >
> > > > > >
> > > > > > and included in the stream config ..
> > > > > >
> > > > > > settings.put(StreamsConfig.ROCKSDB_CONFIG_SETTER_CLASS_CONFIG,
> > > > > > classOf[CustomRocksDBConfig])
> > > > > >
> > > > > >
> > > > > > Regards
> > > > > > Ashik
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > > On Wed, Jul 17, 2019 at 10:52 PM Sophie Blee-Goldman <
> > > > > sop...@confluent.io>
> > > > > > wrote:
> > > > > >
> > > > > > > Sorry, didn't see the "off-heap" part of the email. Are you
> using
> > > any
> > > > > > > stateful DSL operators? The default stores are persistent, so
> you
> > > may
> > > > > > have
> > > > > > > a RocksDB store in your topology without explicitly using one.
> > > > > > >
> > > > > > > On Wed, Jul 17, 2019 at 10:12 AM Sophie Blee-Goldman <
> > > > > > sop...@confluent.io>
> > > > > > > wrote:
> > > > > > >
> > > > > > > > If you are using inMemoryKeyValueStore, the records are
> stored
> > by
> > > > > > > > definition in memory. RocksDB is not used at all. This store
> > will
> > > > > > > continue
> > > > > > > > to grow proportionally to your keyspace. If you do not have
> > > > > sufficient
> > > > > > > > memory to hold your entire dataset in memory, consider adding
> > > > another
> > > > > > > > instance or switching to the RocksDB store
> > > > > > > >
> > > > > > > > On Wed, Jul 17, 2019 at 6:22 AM Muhammed Ashik <
> > > ashi...@gmail.com>
> > > > > > > wrote:
> > > > > > > >
> > > > > > > >> Kafka Streams version - 2.0.0
> > > > > > > >>
> > > > > > > >> Hi, in our streaming instance we are observing a steady
> growth
> > > in
> > > > > the
> > > > > > > >> off-heap memory (out of 2gb allocated memory 1.3 is reserved
> > for
> > > > > heap
> > > > > > > >> memory and the ~700mb memory is utilised over a time of
> ~6hrs
> > > and
> > > > > the
> > > > > > > >> process is OOM killed eventually).
> > > > > > > >>
> > > > > > > >> we are using only the inMemoryKeyValueStore and not doing
> any
> > > > > > > persistence.
> > > > > > > >> as suggested the iterators are closed at the places it is
> > > > used(using
> > > > > > it
> > > > > > > in
> > > > > > > >> only once).
> > > > > > > >>
> > > > > > > >> Some forums were relating such issues with rocksdb but we
> are
> > > not
> > > > > > > >> specifying rocksdb in the config explicitly though. I was
> not
> > > sure
> > > > > > > whether
> > > > > > > >> it is used as a default in memory store by kafka streams.
> > > > > > > >>
> > > > > > > >> Regards
> > > > > > > >> Ashik
> > > > > > > >>
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
>

Reply via email to