Ahh, I was going to add it to the run-class.sh script.

Yeah, it's already there by default:


# Metrics
metrics.reporters=snapshot,jmx
metrics.reporter.snapshot.class=org.apache.samza.metrics.reporter.MetricsSnapshotReporterFactory
metrics.reporter.snapshot.stream=kafka.metrics
metrics.reporter.jmx.class=org.apache.samza.metrics.reporter.JmxReporterFactory

So, where would I see those metrics?

On Mon, Mar 23, 2015 at 5:15 PM, Ash W Matheson <ash.mathe...@gmail.com>
wrote:

> read: I'm a C++ programmer looking at Java for the first time in > 10 years
>
> On Mon, Mar 23, 2015 at 5:13 PM, Ash W Matheson <ash.mathe...@gmail.com>
> wrote:
>
>> I'm assuming I have Jmx defined ... where would that get set?
>>
>> On Mon, Mar 23, 2015 at 5:08 PM, Chinmay Soman <chinmay.cere...@gmail.com
>> > wrote:
>>
>>> Hey Ash,
>>>
>>> Can you see your job metrics (if you have the Jmx metrics defined) to see
>>> if your job is actually doing anything ? My only guess at this point is
>>> the
>>> process method is not being called because somehow there's no incoming
>>> data. I could be totally wrong of course.
>>>
>>> On Mon, Mar 23, 2015 at 4:28 PM, Ash W Matheson <ash.mathe...@gmail.com>
>>> wrote:
>>>
>>> > Just to be clear, here's what's changed from the default hello-samza
>>> repo:
>>> >
>>> > wikipedia-parser.properties==========================
>>> > task.inputs=kafka.myTopic
>>> > systems.kafka.consumer.zookeeper.connect=
>>> > ec2-xxx-xxx-xxx-xxx.compute-1.amazonaws.com:2181/
>>> > systems.kafka.consumer.auto.offset.reset=smallest
>>> >
>>> > WikipediaParserStreamTask.java =====================
>>> >   public void process(IncomingMessageEnvelope envelope,
>>> MessageCollector
>>> > collector, TaskCoordinator coordinator) {
>>> >     Map<String, Object> jsonObject = (Map<String, Object>)
>>> > envelope.getMessage();
>>> >     WikipediaFeedEvent event = new WikipediaFeedEvent(jsonObject);
>>> >
>>> >     try {
>>> >       System.out.println(event.getRawEvent());
>>> >       // Map<String, Object> parsedJsonObject =
>>> parse(event.getRawEvent());
>>> >
>>> >       // parsedJsonObject.put("channel", event.getChannel());
>>> >       // parsedJsonObject.put("source", event.getSource());
>>> >       // parsedJsonObject.put("time", event.getTime());
>>> >
>>> >       // collector.send(new OutgoingMessageEnvelope(new
>>> > SystemStream("kafka", "wikipedia-edits"), parsedJsonObject));
>>> >
>>> > as well as the aforementioned changes to the log4j.xml file.
>>> >
>>> > The data pushed into the 'myTopic' topic is nothing more than a
>>> sentence.
>>> >
>>> >
>>> > On Mon, Mar 23, 2015 at 4:16 PM, Ash W Matheson <
>>> ash.mathe...@gmail.com>
>>> > wrote:
>>> >
>>> > > yep, modified log4j.xml to look like this:
>>> > >
>>> > >   <root>
>>> > >     <priority value="debug" />
>>> > >     <appender-ref ref="RollingAppender"/>
>>> > >     <appender-ref ref="jmx" />
>>> > >   </root>
>>> > >
>>> > > Not sure what you mean by #2.
>>> > >
>>> > > However, I'm running now, not seeing any exceptions, but still not
>>> seeing
>>> > > any output from System.out.println(...)
>>> > >
>>> > > On Mon, Mar 23, 2015 at 11:29 AM, Naveen Somasundaram <
>>> > > nsomasunda...@linkedin.com.invalid> wrote:
>>> > >
>>> > >> Hey Ash,
>>> > >>                1. Did you happen to modify your log4j.xml ?
>>> > >>                2. Can you print the class path that was printed
>>> when the
>>> > >> job started ? I am wondering if log4j was not loaded or not present
>>> in
>>> > the
>>> > >> path where it’s looking for. If you have been using hello samza, it
>>> > should
>>> > >> have pulled it from Maven.
>>> > >>
>>> > >> Thanks,
>>> > >> Naveen
>>> > >>
>>> > >> On Mar 22, 2015, at 10:35 AM, Ash W Matheson <
>>> ash.mathe...@gmail.com>
>>> > >> wrote:
>>> > >>
>>> > >> > Hey all,
>>> > >> >
>>> > >> > Evaluating Samza currently and am running into some odd issues.
>>> > >> >
>>> > >> > I'm currently working off the 'hello-samza' repo and trying to
>>> parse a
>>> > >> > simple kafka topic that I've produced through an extenal java app
>>> > >> (nothing
>>> > >> > other than a series of sentences) and it's failing pretty hard
>>> for me.
>>> > >> The
>>> > >> > base 'hello-samza' set of apps works fine, but as soon as I
>>> change the
>>> > >> > configuration to look at a different Kafka/zookeeper I get the
>>> > >> following in
>>> > >> > the userlogs:
>>> > >> >
>>> > >> > 2015-03-22 17:07:09 KafkaSystemAdmin [WARN] Unable to fetch last
>>> > offsets
>>> > >> > for streams [myTopic] due to kafka.common.KafkaException: fetching
>>> > topic
>>> > >> > metadata for topics [Set(myTopic)] from broker
>>> > >> > [ArrayBuffer(id:0,host:redacted,port:9092)] failed. Retrying.
>>> > >> >
>>> > >> >
>>> > >> > The modifications are pretty straightforward.  In the
>>> > >> > Wikipedia-parser.properties, I've changed the following:
>>> > >> > task.inputs=kafka.myTopic
>>> > >> > systems.kafka.consumer.zookeeper.connect=redacted:2181/
>>> > >> > systems.kafka.consumer.auto.offset.reset=smallest
>>> > >> > systems.kafka.producer.metadata.broker.list=redacted:9092
>>> > >> >
>>> > >> > and in the actual java file WikipediaParserStreamTask.java
>>> > >> >  public void process(IncomingMessageEnvelope envelope,
>>> > MessageCollector
>>> > >> > collector, TaskCoordinator coordinator) {
>>> > >> >    Map<String, Object> jsonObject = (Map<String, Object>)
>>> > >> > envelope.getMessage();
>>> > >> >    WikipediaFeedEvent event = new WikipediaFeedEvent(jsonObject);
>>> > >> >
>>> > >> >    try {
>>> > >> >        System.out.println(event.getRawEvent());
>>> > >> >
>>> > >> > And then following the compile/extract/run process outlined in the
>>> > >> > hello-samza website.
>>> > >> >
>>> > >> > Any thoughts?  I've looked online for any 'super simple' examples
>>> of
>>> > >> > ingesting kafka in samza with very little success.
>>> > >>
>>> > >>
>>> > >
>>> >
>>>
>>>
>>>
>>> --
>>> Thanks and regards
>>>
>>> Chinmay Soman
>>>
>>
>>
>

Reply via email to