Steve Loughran created HADOOP-16863:
---------------------------------------

             Summary: Report on S3A cached 404 recovery better
                 Key: HADOOP-16863
                 URL: https://issues.apache.org/jira/browse/HADOOP-16863
             Project: Hadoop Common
          Issue Type: Sub-task
          Components: fs/s3
    Affects Versions: 3.3.0
            Reporter: Steve Loughran


A big hadoop -fs copyFromLocal is showing that 404 cacheing is still happening. 

{code}
20/02/13 01:02:18 WARN s3a.S3AFileSystem: Failed to find file 
s3a://dilbert/dogbert/queries_split_1/catberg.q._COPYING_. Either it is not yet 
visible, or it has been deleted.
0/02/13 01:02:18 WARN s3a.S3AFileSystem: Failed to find file 
s3a://dilbert/dogbert/queries_split_1/catberg.q._COPYING_. Either it is not yet 
visible, or it has been deleted.
{noformat}

We are recovering (good) but it's (a) got the people running this code worried 
and (b) shouldn't be happening.


Proposed

* error message to -> to a wiki link to a (new) doc on the topic.
* retried clause to increment counter & if count >1 report on #of attempts and 
duration
* S3A FS.deleteOnExit to avoid all checks
* and review the copyFromLocal to make sure no other probes are happening'



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org

Reply via email to