Yes, I have updated on all cluster nodes and restarted entire cluster. 

Do you see any problems with the steps that I followed?

Regards,
Vijay

Sent from my iPhone

> On Mar 23, 2016, at 7:18 AM, Aljoscha Krettek <aljos...@apache.org> wrote:
> 
> Hi,
> did you update the log4j.properties file on all nodes where the TaskManagers 
> run and did you restart the whole cluster?
> 
> Cheers,
> Aljoscha
>> On 23 Mar 2016, at 15:02, Vijay <vijikar...@yahoo.com.INVALID> wrote:
>> 
>> Hi Aljoscha,
>> 
>> I am using standalone flink cluster (3 node). I am running flink job by 
>> submitting/uploading jar through Flink UI.
>> 
>> I have built flink from maven and modified the RollingSink code to add new 
>> debug statements.
>> 
>> I have also packaged the streaming file system connector package (including 
>> RollingSink changes) to the job jar file. Modified changes include both 
>> Sytem.out as well as logger statements.
>> 
>> Updated log4j property file to DEBUG
>> 
>> Regards,
>> Vijay
>> 
>> Sent from my iPhone
>> 
>>> On Mar 23, 2016, at 6:48 AM, Aljoscha Krettek <aljos...@apache.org> wrote:
>>> 
>>> Hi,
>>> what where the steps you took? By the way, are you running this on yarn or 
>>> in standalone mode? How are you starting the Flink job? Do you still don’t 
>>> see DEBUG entries in the log?
>>> 
>>> Cheers,
>>> Aljoscha
>>>> On 23 Mar 2016, at 14:32, Vijay <vijikar...@yahoo.com> wrote:
>>>> 
>>>> I have changed the properties file but it did not help.
>>>> 
>>>> Regards,
>>>> Vijay
>>>> 
>>>> Sent from my iPhone
>>>> 
>>>>> On Mar 23, 2016, at 5:39 AM, Aljoscha Krettek <aljos...@apache.org> wrote:
>>>>> 
>>>>> Ok, then you should be able to change the log level to DEBUG in 
>>>>> conf/log4j.properties.
>>>>> 
>>>>>> On 23 Mar 2016, at 12:41, Vijay <vijikar...@yahoo.com> wrote:
>>>>>> 
>>>>>> I think only the ERROR category gets displayed in the log file
>>>>>> 
>>>>>> Regards,
>>>>>> Vijay
>>>>>> 
>>>>>> Sent from my iPhone
>>>>>> 
>>>>>>> On Mar 23, 2016, at 2:30 AM, Aljoscha Krettek <aljos...@apache.org> 
>>>>>>> wrote:
>>>>>>> 
>>>>>>> Hi,
>>>>>>> are you seeing the regular log output from the RollingSink in the 
>>>>>>> TaskManager logs?
>>>>>>> 
>>>>>>> Cheers,
>>>>>>> Aljoscha
>>>>>>>> On 22 Mar 2016, at 20:03, Vijay Srinivasaraghavan 
>>>>>>>> <vijikar...@yahoo.com> wrote:
>>>>>>>> 
>>>>>>>> I have tried both log4j logger as well as System.out.println option 
>>>>>>>> but none of these worked. 
>>>>>>>> 
>>>>>>>> From what I have seen so far is the Filesystem streaming connector 
>>>>>>>> classes are not packaged in the grand jar 
>>>>>>>> (flink-dist_2.10-1.1-SNAPSHOT.jar) that is copied under 
>>>>>>>> <FLINK_HOME>/build-target/lib location as part of Flink maven build 
>>>>>>>> step.
>>>>>>>> 
>>>>>>>> So, I manually copied (overwrite) the compiled class files from 
>>>>>>>> org.apache.flink.streaming.connectors.fs package to the my "Flink job" 
>>>>>>>> distribution jar (otherwise it was using standard jars that are 
>>>>>>>> defined as mvn dependency in Articatory) and then uploaded the jar to 
>>>>>>>> Job Manager.
>>>>>>>> 
>>>>>>>> Am I missing something? How do I enable logging for the RollingSink 
>>>>>>>> class?
>>>>>>>> 
>>>>>>>> <dependency>
>>>>>>>> <groupId>org.apache.flink</groupId>
>>>>>>>> <artifactId>flink-connector-filesystem_2.11</artifactId>
>>>>>>>> <version>${flink.version}</version>
>>>>>>>> <scope>provided</scope>
>>>>>>>> </dependency>
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On Tuesday, March 22, 2016 3:04 AM, Aljoscha Krettek 
>>>>>>>> <aljos...@apache.org> wrote:
>>>>>>>> 
>>>>>>>> 
>>>>>>>> Hi,
>>>>>>>> how are you printing the debug statements?
>>>>>>>> 
>>>>>>>> But yeah all the logic of renaming in progress files and cleaning up 
>>>>>>>> after a failed job happens in restoreState(BucketState state). The 
>>>>>>>> steps are roughly these:
>>>>>>>> 
>>>>>>>> 1. Move current in-progress file to final location
>>>>>>>> 2. truncate the file if necessary (if truncate is not available write 
>>>>>>>> a .valid-length file)
>>>>>>>> 3. Move pending files to final location that where part of the 
>>>>>>>> checkpoint
>>>>>>>> 4. cleanup any leftover pending/in-progress files
>>>>>>>> 
>>>>>>>> Cheers,
>>>>>>>> Aljoscha
>>>>>>>> 
>>>>>>>>> On 22 Mar 2016, at 10:08, Vijay Srinivasaraghavan 
>>>>>>>>> <vijikar...@yahoo.com.INVALID> wrote:
>>>>>>>>> 
>>>>>>>>> Hello,
>>>>>>>>> I have enabled checkpoint and I am using RollingSink to sink the data 
>>>>>>>>> to HDFS (2.7.x) from KafkaConsumer. To simulate failover/recovery, I 
>>>>>>>>> stopped TaskManager and the job gets rescheduled to other Taskmanager 
>>>>>>>>> instance. During this momemnt, the current "in-progress" gets closed 
>>>>>>>>> and renamed to part-0-1 from _part-0-1_in-progress. 
>>>>>>>>> I was hoping to see the debug statement that I have added to 
>>>>>>>>> "restoreState" method but none of my debug statement gets printed. I 
>>>>>>>>> am not sure if the restoreState() method gets invoked during this 
>>>>>>>>> scenario. Could you please help me understand the flow during 
>>>>>>>>> "failover" scenario?
>>>>>>>>> P.S: Functionally the code appears to be working fine but I am trying 
>>>>>>>>> to understand the underlying implementation details. public void 
>>>>>>>>> restoreState(BucketState state)
>>>>>>>>> Regards
>>>>>>>>> Vijay
> 

Reply via email to