Ahh, I was going to add it to the run-class.sh script. Yeah, it's already there by default:
# Metrics metrics.reporters=snapshot,jmx metrics.reporter.snapshot.class=org.apache.samza.metrics.reporter.MetricsSnapshotReporterFactory metrics.reporter.snapshot.stream=kafka.metrics metrics.reporter.jmx.class=org.apache.samza.metrics.reporter.JmxReporterFactory So, where would I see those metrics? On Mon, Mar 23, 2015 at 5:15 PM, Ash W Matheson <ash.mathe...@gmail.com> wrote: > read: I'm a C++ programmer looking at Java for the first time in > 10 years > > On Mon, Mar 23, 2015 at 5:13 PM, Ash W Matheson <ash.mathe...@gmail.com> > wrote: > >> I'm assuming I have Jmx defined ... where would that get set? >> >> On Mon, Mar 23, 2015 at 5:08 PM, Chinmay Soman <chinmay.cere...@gmail.com >> > wrote: >> >>> Hey Ash, >>> >>> Can you see your job metrics (if you have the Jmx metrics defined) to see >>> if your job is actually doing anything ? My only guess at this point is >>> the >>> process method is not being called because somehow there's no incoming >>> data. I could be totally wrong of course. >>> >>> On Mon, Mar 23, 2015 at 4:28 PM, Ash W Matheson <ash.mathe...@gmail.com> >>> wrote: >>> >>> > Just to be clear, here's what's changed from the default hello-samza >>> repo: >>> > >>> > wikipedia-parser.properties========================== >>> > task.inputs=kafka.myTopic >>> > systems.kafka.consumer.zookeeper.connect= >>> > ec2-xxx-xxx-xxx-xxx.compute-1.amazonaws.com:2181/ >>> > systems.kafka.consumer.auto.offset.reset=smallest >>> > >>> > WikipediaParserStreamTask.java ===================== >>> > public void process(IncomingMessageEnvelope envelope, >>> MessageCollector >>> > collector, TaskCoordinator coordinator) { >>> > Map<String, Object> jsonObject = (Map<String, Object>) >>> > envelope.getMessage(); >>> > WikipediaFeedEvent event = new WikipediaFeedEvent(jsonObject); >>> > >>> > try { >>> > System.out.println(event.getRawEvent()); >>> > // Map<String, Object> parsedJsonObject = >>> parse(event.getRawEvent()); >>> > >>> > // parsedJsonObject.put("channel", event.getChannel()); >>> > // parsedJsonObject.put("source", event.getSource()); >>> > // parsedJsonObject.put("time", event.getTime()); >>> > >>> > // collector.send(new OutgoingMessageEnvelope(new >>> > SystemStream("kafka", "wikipedia-edits"), parsedJsonObject)); >>> > >>> > as well as the aforementioned changes to the log4j.xml file. >>> > >>> > The data pushed into the 'myTopic' topic is nothing more than a >>> sentence. >>> > >>> > >>> > On Mon, Mar 23, 2015 at 4:16 PM, Ash W Matheson < >>> ash.mathe...@gmail.com> >>> > wrote: >>> > >>> > > yep, modified log4j.xml to look like this: >>> > > >>> > > <root> >>> > > <priority value="debug" /> >>> > > <appender-ref ref="RollingAppender"/> >>> > > <appender-ref ref="jmx" /> >>> > > </root> >>> > > >>> > > Not sure what you mean by #2. >>> > > >>> > > However, I'm running now, not seeing any exceptions, but still not >>> seeing >>> > > any output from System.out.println(...) >>> > > >>> > > On Mon, Mar 23, 2015 at 11:29 AM, Naveen Somasundaram < >>> > > nsomasunda...@linkedin.com.invalid> wrote: >>> > > >>> > >> Hey Ash, >>> > >> 1. Did you happen to modify your log4j.xml ? >>> > >> 2. Can you print the class path that was printed >>> when the >>> > >> job started ? I am wondering if log4j was not loaded or not present >>> in >>> > the >>> > >> path where it’s looking for. If you have been using hello samza, it >>> > should >>> > >> have pulled it from Maven. >>> > >> >>> > >> Thanks, >>> > >> Naveen >>> > >> >>> > >> On Mar 22, 2015, at 10:35 AM, Ash W Matheson < >>> ash.mathe...@gmail.com> >>> > >> wrote: >>> > >> >>> > >> > Hey all, >>> > >> > >>> > >> > Evaluating Samza currently and am running into some odd issues. >>> > >> > >>> > >> > I'm currently working off the 'hello-samza' repo and trying to >>> parse a >>> > >> > simple kafka topic that I've produced through an extenal java app >>> > >> (nothing >>> > >> > other than a series of sentences) and it's failing pretty hard >>> for me. >>> > >> The >>> > >> > base 'hello-samza' set of apps works fine, but as soon as I >>> change the >>> > >> > configuration to look at a different Kafka/zookeeper I get the >>> > >> following in >>> > >> > the userlogs: >>> > >> > >>> > >> > 2015-03-22 17:07:09 KafkaSystemAdmin [WARN] Unable to fetch last >>> > offsets >>> > >> > for streams [myTopic] due to kafka.common.KafkaException: fetching >>> > topic >>> > >> > metadata for topics [Set(myTopic)] from broker >>> > >> > [ArrayBuffer(id:0,host:redacted,port:9092)] failed. Retrying. >>> > >> > >>> > >> > >>> > >> > The modifications are pretty straightforward. In the >>> > >> > Wikipedia-parser.properties, I've changed the following: >>> > >> > task.inputs=kafka.myTopic >>> > >> > systems.kafka.consumer.zookeeper.connect=redacted:2181/ >>> > >> > systems.kafka.consumer.auto.offset.reset=smallest >>> > >> > systems.kafka.producer.metadata.broker.list=redacted:9092 >>> > >> > >>> > >> > and in the actual java file WikipediaParserStreamTask.java >>> > >> > public void process(IncomingMessageEnvelope envelope, >>> > MessageCollector >>> > >> > collector, TaskCoordinator coordinator) { >>> > >> > Map<String, Object> jsonObject = (Map<String, Object>) >>> > >> > envelope.getMessage(); >>> > >> > WikipediaFeedEvent event = new WikipediaFeedEvent(jsonObject); >>> > >> > >>> > >> > try { >>> > >> > System.out.println(event.getRawEvent()); >>> > >> > >>> > >> > And then following the compile/extract/run process outlined in the >>> > >> > hello-samza website. >>> > >> > >>> > >> > Any thoughts? I've looked online for any 'super simple' examples >>> of >>> > >> > ingesting kafka in samza with very little success. >>> > >> >>> > >> >>> > > >>> > >>> >>> >>> >>> -- >>> Thanks and regards >>> >>> Chinmay Soman >>> >> >> >