read: I'm a C++ programmer looking at Java for the first time in > 10 years

On Mon, Mar 23, 2015 at 5:13 PM, Ash W Matheson <ash.mathe...@gmail.com>
wrote:

> I'm assuming I have Jmx defined ... where would that get set?
>
> On Mon, Mar 23, 2015 at 5:08 PM, Chinmay Soman <chinmay.cere...@gmail.com>
> wrote:
>
>> Hey Ash,
>>
>> Can you see your job metrics (if you have the Jmx metrics defined) to see
>> if your job is actually doing anything ? My only guess at this point is
>> the
>> process method is not being called because somehow there's no incoming
>> data. I could be totally wrong of course.
>>
>> On Mon, Mar 23, 2015 at 4:28 PM, Ash W Matheson <ash.mathe...@gmail.com>
>> wrote:
>>
>> > Just to be clear, here's what's changed from the default hello-samza
>> repo:
>> >
>> > wikipedia-parser.properties==========================
>> > task.inputs=kafka.myTopic
>> > systems.kafka.consumer.zookeeper.connect=
>> > ec2-xxx-xxx-xxx-xxx.compute-1.amazonaws.com:2181/
>> > systems.kafka.consumer.auto.offset.reset=smallest
>> >
>> > WikipediaParserStreamTask.java =====================
>> >   public void process(IncomingMessageEnvelope envelope, MessageCollector
>> > collector, TaskCoordinator coordinator) {
>> >     Map<String, Object> jsonObject = (Map<String, Object>)
>> > envelope.getMessage();
>> >     WikipediaFeedEvent event = new WikipediaFeedEvent(jsonObject);
>> >
>> >     try {
>> >       System.out.println(event.getRawEvent());
>> >       // Map<String, Object> parsedJsonObject =
>> parse(event.getRawEvent());
>> >
>> >       // parsedJsonObject.put("channel", event.getChannel());
>> >       // parsedJsonObject.put("source", event.getSource());
>> >       // parsedJsonObject.put("time", event.getTime());
>> >
>> >       // collector.send(new OutgoingMessageEnvelope(new
>> > SystemStream("kafka", "wikipedia-edits"), parsedJsonObject));
>> >
>> > as well as the aforementioned changes to the log4j.xml file.
>> >
>> > The data pushed into the 'myTopic' topic is nothing more than a
>> sentence.
>> >
>> >
>> > On Mon, Mar 23, 2015 at 4:16 PM, Ash W Matheson <ash.mathe...@gmail.com
>> >
>> > wrote:
>> >
>> > > yep, modified log4j.xml to look like this:
>> > >
>> > >   <root>
>> > >     <priority value="debug" />
>> > >     <appender-ref ref="RollingAppender"/>
>> > >     <appender-ref ref="jmx" />
>> > >   </root>
>> > >
>> > > Not sure what you mean by #2.
>> > >
>> > > However, I'm running now, not seeing any exceptions, but still not
>> seeing
>> > > any output from System.out.println(...)
>> > >
>> > > On Mon, Mar 23, 2015 at 11:29 AM, Naveen Somasundaram <
>> > > nsomasunda...@linkedin.com.invalid> wrote:
>> > >
>> > >> Hey Ash,
>> > >>                1. Did you happen to modify your log4j.xml ?
>> > >>                2. Can you print the class path that was printed when
>> the
>> > >> job started ? I am wondering if log4j was not loaded or not present
>> in
>> > the
>> > >> path where it’s looking for. If you have been using hello samza, it
>> > should
>> > >> have pulled it from Maven.
>> > >>
>> > >> Thanks,
>> > >> Naveen
>> > >>
>> > >> On Mar 22, 2015, at 10:35 AM, Ash W Matheson <ash.mathe...@gmail.com
>> >
>> > >> wrote:
>> > >>
>> > >> > Hey all,
>> > >> >
>> > >> > Evaluating Samza currently and am running into some odd issues.
>> > >> >
>> > >> > I'm currently working off the 'hello-samza' repo and trying to
>> parse a
>> > >> > simple kafka topic that I've produced through an extenal java app
>> > >> (nothing
>> > >> > other than a series of sentences) and it's failing pretty hard for
>> me.
>> > >> The
>> > >> > base 'hello-samza' set of apps works fine, but as soon as I change
>> the
>> > >> > configuration to look at a different Kafka/zookeeper I get the
>> > >> following in
>> > >> > the userlogs:
>> > >> >
>> > >> > 2015-03-22 17:07:09 KafkaSystemAdmin [WARN] Unable to fetch last
>> > offsets
>> > >> > for streams [myTopic] due to kafka.common.KafkaException: fetching
>> > topic
>> > >> > metadata for topics [Set(myTopic)] from broker
>> > >> > [ArrayBuffer(id:0,host:redacted,port:9092)] failed. Retrying.
>> > >> >
>> > >> >
>> > >> > The modifications are pretty straightforward.  In the
>> > >> > Wikipedia-parser.properties, I've changed the following:
>> > >> > task.inputs=kafka.myTopic
>> > >> > systems.kafka.consumer.zookeeper.connect=redacted:2181/
>> > >> > systems.kafka.consumer.auto.offset.reset=smallest
>> > >> > systems.kafka.producer.metadata.broker.list=redacted:9092
>> > >> >
>> > >> > and in the actual java file WikipediaParserStreamTask.java
>> > >> >  public void process(IncomingMessageEnvelope envelope,
>> > MessageCollector
>> > >> > collector, TaskCoordinator coordinator) {
>> > >> >    Map<String, Object> jsonObject = (Map<String, Object>)
>> > >> > envelope.getMessage();
>> > >> >    WikipediaFeedEvent event = new WikipediaFeedEvent(jsonObject);
>> > >> >
>> > >> >    try {
>> > >> >        System.out.println(event.getRawEvent());
>> > >> >
>> > >> > And then following the compile/extract/run process outlined in the
>> > >> > hello-samza website.
>> > >> >
>> > >> > Any thoughts?  I've looked online for any 'super simple' examples
>> of
>> > >> > ingesting kafka in samza with very little success.
>> > >>
>> > >>
>> > >
>> >
>>
>>
>>
>> --
>> Thanks and regards
>>
>> Chinmay Soman
>>
>
>

Reply via email to