And all four stores (BucketData, CacheData, StationKeyValue,
StationCacheData) are definitely explicitly added as
Stores.inMemoryKeyValueStore("name")? Hm.

Does du -h show any contents within the
./buzzard.MonitoringSeWlanStatsAggregator/0_1 directories that match any of
these store names? Also if you rerun the app with a different state.dir, do
all the same directories get created?

On Thu, Jul 18, 2019 at 12:15 AM Muhammed Ashik <ashi...@gmail.com> wrote:

> Hi, my Topology description..
>
> - Topologies:
>    Sub-topology: 0
>     Source: SeSource (topics: [buzzard.decoded.amon.stats.ap,
> buzzard.app.monitoring.internal.clientsestats])
>       --> SeProcess
>     Processor: SeProcess (stores: [BucketData, CacheData])
>       --> AeSink
>       <-- SeSource
>     Sink: AeSink (topic: buzzard.app.monitoring.internal.stats)
>       <-- SeProcess
>
>   Sub-topology: 1
>     Source: StationSource (topics: [buzzard.decoded.amon.stats.station])
>       --> StationProcess
>     Processor: StationProcess (stores: [StationKeyValue, StationCacheData])
>       --> StationSink
>       <-- StationSource
>     Sink: StationSink (topic:
> buzzard.app.monitoring.internal.clientsestats)
>       <-- StationProcess
>
> Regards
> Ashik
>
> On Thu, Jul 18, 2019 at 1:31 AM Sophie Blee-Goldman <sop...@confluent.io>
> wrote:
>
> > Hm. These directories shouldn't be created if using only an in memory
> > store. Can you print your topology?
> >
> > On Wed, Jul 17, 2019 at 11:02 AM Muhammed Ashik <ashi...@gmail.com>
> wrote:
> >
> > > Hi I just did `du -mh` on `\tmp\kafka-streams` below are the folders
> > listed
> > > with some .lock files inside.
> > > not sure if these are coming from rocksdb.. and looks like the sizes of
> > > these files are less.
> > >
> > > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/0_5
> > > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/0_3
> > > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/0_2
> > > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/1_4
> > > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/0_1
> > > 4.0K ./buzzard.MonitoringSeWlanStatsAggregator/0_4
> > > 28K ./buzzard.MonitoringSeWlanStatsAggregator
> > > 32K .
> > >
> > >
> > >
> > > On Wed, Jul 17, 2019 at 11:18 PM Sophie Blee-Goldman <
> > sop...@confluent.io>
> > > wrote:
> > >
> > > > Hm. Just to be absolutely sure, could you try throwing an exception
> or
> > > > something in your RocksDBConfigSetter?
> > > >
> > > > On Wed, Jul 17, 2019 at 10:43 AM Muhammed Ashik <ashi...@gmail.com>
> > > wrote:
> > > >
> > > > > I can confirm the /tmp/kafka-streams doesn't have any data related
> to
> > > > > rocksdb.
> > > > >
> > > > > On Wed, Jul 17, 2019 at 11:11 PM Sophie Blee-Goldman <
> > > > sop...@confluent.io>
> > > > > wrote:
> > > > >
> > > > > > You can describe your topology to see if there are any state
> stores
> > > in
> > > > it
> > > > > > that you aren't aware of. Alternatively you could check out the
> > state
> > > > > > directory (/tmp/kafka-streams by default) and see if there is any
> > > data
> > > > in
> > > > > > there
> > > > > >
> > > > > > On Wed, Jul 17, 2019 at 10:36 AM Muhammed Ashik <
> ashi...@gmail.com
> > >
> > > > > wrote:
> > > > > >
> > > > > > > Thanks How can I verify If there is some data really going on
> > > rocksdb
> > > > > > > I tried printing the statistics with no success.
> > > > > > >
> > > > > > > class CustomRocksDBConfig extends RocksDBConfigSetter {
> > > > > > >   override def setConfig(storeName: String, options: Options,
> > > > configs:
> > > > > > > util.Map[String, AnyRef]): Unit = {
> > > > > > >
> > > > > > >     val stats = new Statistics
> > > > > > >     stats.setStatsLevel(StatsLevel.ALL)
> > > > > > >     options.setStatistics(stats)
> > > > > > >       .setStatsDumpPeriodSec(600)
> > > > > > >     options
> > > > > > >       .setInfoLogLevel(InfoLogLevel.INFO_LEVEL)
> > > > > > >     options.setDbLogDir("/tmp/dump")
> > > > > > >
> > > > > > >   }
> > > > > > > }
> > > > > > >
> > > > > > >
> > > > > > > and included in the stream config ..
> > > > > > >
> > > > > > > settings.put(StreamsConfig.ROCKSDB_CONFIG_SETTER_CLASS_CONFIG,
> > > > > > > classOf[CustomRocksDBConfig])
> > > > > > >
> > > > > > >
> > > > > > > Regards
> > > > > > > Ashik
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > On Wed, Jul 17, 2019 at 10:52 PM Sophie Blee-Goldman <
> > > > > > sop...@confluent.io>
> > > > > > > wrote:
> > > > > > >
> > > > > > > > Sorry, didn't see the "off-heap" part of the email. Are you
> > using
> > > > any
> > > > > > > > stateful DSL operators? The default stores are persistent, so
> > you
> > > > may
> > > > > > > have
> > > > > > > > a RocksDB store in your topology without explicitly using
> one.
> > > > > > > >
> > > > > > > > On Wed, Jul 17, 2019 at 10:12 AM Sophie Blee-Goldman <
> > > > > > > sop...@confluent.io>
> > > > > > > > wrote:
> > > > > > > >
> > > > > > > > > If you are using inMemoryKeyValueStore, the records are
> > stored
> > > by
> > > > > > > > > definition in memory. RocksDB is not used at all. This
> store
> > > will
> > > > > > > > continue
> > > > > > > > > to grow proportionally to your keyspace. If you do not have
> > > > > > sufficient
> > > > > > > > > memory to hold your entire dataset in memory, consider
> adding
> > > > > another
> > > > > > > > > instance or switching to the RocksDB store
> > > > > > > > >
> > > > > > > > > On Wed, Jul 17, 2019 at 6:22 AM Muhammed Ashik <
> > > > ashi...@gmail.com>
> > > > > > > > wrote:
> > > > > > > > >
> > > > > > > > >> Kafka Streams version - 2.0.0
> > > > > > > > >>
> > > > > > > > >> Hi, in our streaming instance we are observing a steady
> > growth
> > > > in
> > > > > > the
> > > > > > > > >> off-heap memory (out of 2gb allocated memory 1.3 is
> reserved
> > > for
> > > > > > heap
> > > > > > > > >> memory and the ~700mb memory is utilised over a time of
> > ~6hrs
> > > > and
> > > > > > the
> > > > > > > > >> process is OOM killed eventually).
> > > > > > > > >>
> > > > > > > > >> we are using only the inMemoryKeyValueStore and not doing
> > any
> > > > > > > > persistence.
> > > > > > > > >> as suggested the iterators are closed at the places it is
> > > > > used(using
> > > > > > > it
> > > > > > > > in
> > > > > > > > >> only once).
> > > > > > > > >>
> > > > > > > > >> Some forums were relating such issues with rocksdb but we
> > are
> > > > not
> > > > > > > > >> specifying rocksdb in the config explicitly though. I was
> > not
> > > > sure
> > > > > > > > whether
> > > > > > > > >> it is used as a default in memory store by kafka streams.
> > > > > > > > >>
> > > > > > > > >> Regards
> > > > > > > > >> Ashik
> > > > > > > > >>
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
>

Reply via email to