Yes, I have updated on all cluster nodes and restarted entire cluster. Do you see any problems with the steps that I followed?
Regards, Vijay Sent from my iPhone > On Mar 23, 2016, at 7:18 AM, Aljoscha Krettek <aljos...@apache.org> wrote: > > Hi, > did you update the log4j.properties file on all nodes where the TaskManagers > run and did you restart the whole cluster? > > Cheers, > Aljoscha >> On 23 Mar 2016, at 15:02, Vijay <vijikar...@yahoo.com.INVALID> wrote: >> >> Hi Aljoscha, >> >> I am using standalone flink cluster (3 node). I am running flink job by >> submitting/uploading jar through Flink UI. >> >> I have built flink from maven and modified the RollingSink code to add new >> debug statements. >> >> I have also packaged the streaming file system connector package (including >> RollingSink changes) to the job jar file. Modified changes include both >> Sytem.out as well as logger statements. >> >> Updated log4j property file to DEBUG >> >> Regards, >> Vijay >> >> Sent from my iPhone >> >>> On Mar 23, 2016, at 6:48 AM, Aljoscha Krettek <aljos...@apache.org> wrote: >>> >>> Hi, >>> what where the steps you took? By the way, are you running this on yarn or >>> in standalone mode? How are you starting the Flink job? Do you still don’t >>> see DEBUG entries in the log? >>> >>> Cheers, >>> Aljoscha >>>> On 23 Mar 2016, at 14:32, Vijay <vijikar...@yahoo.com> wrote: >>>> >>>> I have changed the properties file but it did not help. >>>> >>>> Regards, >>>> Vijay >>>> >>>> Sent from my iPhone >>>> >>>>> On Mar 23, 2016, at 5:39 AM, Aljoscha Krettek <aljos...@apache.org> wrote: >>>>> >>>>> Ok, then you should be able to change the log level to DEBUG in >>>>> conf/log4j.properties. >>>>> >>>>>> On 23 Mar 2016, at 12:41, Vijay <vijikar...@yahoo.com> wrote: >>>>>> >>>>>> I think only the ERROR category gets displayed in the log file >>>>>> >>>>>> Regards, >>>>>> Vijay >>>>>> >>>>>> Sent from my iPhone >>>>>> >>>>>>> On Mar 23, 2016, at 2:30 AM, Aljoscha Krettek <aljos...@apache.org> >>>>>>> wrote: >>>>>>> >>>>>>> Hi, >>>>>>> are you seeing the regular log output from the RollingSink in the >>>>>>> TaskManager logs? >>>>>>> >>>>>>> Cheers, >>>>>>> Aljoscha >>>>>>>> On 22 Mar 2016, at 20:03, Vijay Srinivasaraghavan >>>>>>>> <vijikar...@yahoo.com> wrote: >>>>>>>> >>>>>>>> I have tried both log4j logger as well as System.out.println option >>>>>>>> but none of these worked. >>>>>>>> >>>>>>>> From what I have seen so far is the Filesystem streaming connector >>>>>>>> classes are not packaged in the grand jar >>>>>>>> (flink-dist_2.10-1.1-SNAPSHOT.jar) that is copied under >>>>>>>> <FLINK_HOME>/build-target/lib location as part of Flink maven build >>>>>>>> step. >>>>>>>> >>>>>>>> So, I manually copied (overwrite) the compiled class files from >>>>>>>> org.apache.flink.streaming.connectors.fs package to the my "Flink job" >>>>>>>> distribution jar (otherwise it was using standard jars that are >>>>>>>> defined as mvn dependency in Articatory) and then uploaded the jar to >>>>>>>> Job Manager. >>>>>>>> >>>>>>>> Am I missing something? How do I enable logging for the RollingSink >>>>>>>> class? >>>>>>>> >>>>>>>> <dependency> >>>>>>>> <groupId>org.apache.flink</groupId> >>>>>>>> <artifactId>flink-connector-filesystem_2.11</artifactId> >>>>>>>> <version>${flink.version}</version> >>>>>>>> <scope>provided</scope> >>>>>>>> </dependency> >>>>>>>> >>>>>>>> >>>>>>>> On Tuesday, March 22, 2016 3:04 AM, Aljoscha Krettek >>>>>>>> <aljos...@apache.org> wrote: >>>>>>>> >>>>>>>> >>>>>>>> Hi, >>>>>>>> how are you printing the debug statements? >>>>>>>> >>>>>>>> But yeah all the logic of renaming in progress files and cleaning up >>>>>>>> after a failed job happens in restoreState(BucketState state). The >>>>>>>> steps are roughly these: >>>>>>>> >>>>>>>> 1. Move current in-progress file to final location >>>>>>>> 2. truncate the file if necessary (if truncate is not available write >>>>>>>> a .valid-length file) >>>>>>>> 3. Move pending files to final location that where part of the >>>>>>>> checkpoint >>>>>>>> 4. cleanup any leftover pending/in-progress files >>>>>>>> >>>>>>>> Cheers, >>>>>>>> Aljoscha >>>>>>>> >>>>>>>>> On 22 Mar 2016, at 10:08, Vijay Srinivasaraghavan >>>>>>>>> <vijikar...@yahoo.com.INVALID> wrote: >>>>>>>>> >>>>>>>>> Hello, >>>>>>>>> I have enabled checkpoint and I am using RollingSink to sink the data >>>>>>>>> to HDFS (2.7.x) from KafkaConsumer. To simulate failover/recovery, I >>>>>>>>> stopped TaskManager and the job gets rescheduled to other Taskmanager >>>>>>>>> instance. During this momemnt, the current "in-progress" gets closed >>>>>>>>> and renamed to part-0-1 from _part-0-1_in-progress. >>>>>>>>> I was hoping to see the debug statement that I have added to >>>>>>>>> "restoreState" method but none of my debug statement gets printed. I >>>>>>>>> am not sure if the restoreState() method gets invoked during this >>>>>>>>> scenario. Could you please help me understand the flow during >>>>>>>>> "failover" scenario? >>>>>>>>> P.S: Functionally the code appears to be working fine but I am trying >>>>>>>>> to understand the underlying implementation details. public void >>>>>>>>> restoreState(BucketState state) >>>>>>>>> Regards >>>>>>>>> Vijay >