yep, modified log4j.xml to look like this:

  <root>
    <priority value="debug" />
    <appender-ref ref="RollingAppender"/>
    <appender-ref ref="jmx" />
  </root>

Not sure what you mean by #2.

However, I'm running now, not seeing any exceptions, but still not seeing
any output from System.out.println(...)

On Mon, Mar 23, 2015 at 11:29 AM, Naveen Somasundaram <
nsomasunda...@linkedin.com.invalid> wrote:

> Hey Ash,
>                1. Did you happen to modify your log4j.xml ?
>                2. Can you print the class path that was printed when the
> job started ? I am wondering if log4j was not loaded or not present in the
> path where it’s looking for. If you have been using hello samza, it should
> have pulled it from Maven.
>
> Thanks,
> Naveen
>
> On Mar 22, 2015, at 10:35 AM, Ash W Matheson <ash.mathe...@gmail.com>
> wrote:
>
> > Hey all,
> >
> > Evaluating Samza currently and am running into some odd issues.
> >
> > I'm currently working off the 'hello-samza' repo and trying to parse a
> > simple kafka topic that I've produced through an extenal java app
> (nothing
> > other than a series of sentences) and it's failing pretty hard for me.
> The
> > base 'hello-samza' set of apps works fine, but as soon as I change the
> > configuration to look at a different Kafka/zookeeper I get the following
> in
> > the userlogs:
> >
> > 2015-03-22 17:07:09 KafkaSystemAdmin [WARN] Unable to fetch last offsets
> > for streams [myTopic] due to kafka.common.KafkaException: fetching topic
> > metadata for topics [Set(myTopic)] from broker
> > [ArrayBuffer(id:0,host:redacted,port:9092)] failed. Retrying.
> >
> >
> > The modifications are pretty straightforward.  In the
> > Wikipedia-parser.properties, I've changed the following:
> > task.inputs=kafka.myTopic
> > systems.kafka.consumer.zookeeper.connect=redacted:2181/
> > systems.kafka.consumer.auto.offset.reset=smallest
> > systems.kafka.producer.metadata.broker.list=redacted:9092
> >
> > and in the actual java file WikipediaParserStreamTask.java
> >  public void process(IncomingMessageEnvelope envelope, MessageCollector
> > collector, TaskCoordinator coordinator) {
> >    Map<String, Object> jsonObject = (Map<String, Object>)
> > envelope.getMessage();
> >    WikipediaFeedEvent event = new WikipediaFeedEvent(jsonObject);
> >
> >    try {
> >        System.out.println(event.getRawEvent());
> >
> > And then following the compile/extract/run process outlined in the
> > hello-samza website.
> >
> > Any thoughts?  I've looked online for any 'super simple' examples of
> > ingesting kafka in samza with very little success.
>
>

Reply via email to