[jira] [Created] (HDFS-12117) HttpFS does not seem to support SNAPSHOT related methods for WebHDFS REST Interface

2017-07-11 Thread Wellington Chevreuil (JIRA)
Wellington Chevreuil created HDFS-12117:
---

 Summary: HttpFS does not seem to support SNAPSHOT related methods 
for WebHDFS REST Interface
 Key: HDFS-12117
 URL: https://issues.apache.org/jira/browse/HDFS-12117
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: httpfs
Affects Versions: 3.0.0-alpha3
Reporter: Wellington Chevreuil
Assignee: Wellington Chevreuil


Currently, HttpFS is lacking implementation for SNAPSHOT related methods from 
WebHDFS REST interface as defined by WebHDFS documentation [WebHDFS 
documentation|https://archive.cloudera.com/cdh5/cdh/5/hadoop/hadoop-project-dist/hadoop-hdfs/WebHDFS.html#Snapshot_Operations]

I would like to work on this implementation, following the existing design 
approach already implemented by other WebHDFS methods on current HttpFS 
project, so I'll be proposing an initial patch soon for reviews.
 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-12118) Ozone: Ozone shell: Add more testing for volume shell commands

2017-07-11 Thread Yiqun Lin (JIRA)
Yiqun Lin created HDFS-12118:


 Summary: Ozone: Ozone shell: Add more testing for volume shell 
commands
 Key: HDFS-12118
 URL: https://issues.apache.org/jira/browse/HDFS-12118
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ozone, tools
Affects Versions: HDFS-7240
Reporter: Yiqun Lin
Assignee: Yiqun Lin


Currently there are missing enough testing to cover all the ozone command. Now 
we have to test this in a manual way to see if commands run okay as expected. 
There will be lots of test cases we should add for all the volume, bucket and 
key commands. In order to easy reviewed, I'd like to separate this work to 
three subtasks. This JIRA is a good start for implementing for volume commands.
Plan to add the unit test to test following command and its available options:

* infoVolume
* updateVolume
* listVolume
* createVolume




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-12119) Inconsistent "Number of Under-Replicated Blocks" shown on HDFS web UI and fsck report

2017-07-11 Thread liuyiyang (JIRA)
liuyiyang created HDFS-12119:


 Summary: Inconsistent "Number of Under-Replicated Blocks"  shown 
on HDFS web UI and fsck report
 Key: HDFS-12119
 URL: https://issues.apache.org/jira/browse/HDFS-12119
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.6.0
Reporter: liuyiyang


Sometimes the information "Number of Under-Replicated Blocks" shown on NameNode 
web UI is inconsistent with the "Under-replicated blocks" information shown in 
fsck report.

It's easy to reproduce such a case as follows:
1、In a cluster with DN1(rack0)、DN2(rack1) and DN3(rack2) which stores a lot of 
blocks, the replication factor is set to 2;
2、Re-allocate racks as DN1(rack0)、DN2(rack1) and DN3(rack1) ;
3、Restart HDFS daemons.
Then you can find inconsistent  "Number of Under-Replicated Blocks" on web ui 
and "Under-replicated blocks" in fsck result.

I dug into the source code and found that  "Number of Under-Replicated Blocks" 
on web ui consists of blocks that have less than target number of replicas and 
blocks that have the right number of replicas, but which the block manager felt 
were badly distributed.  In fsck result, "Under-replicated blocks" are the 
blocks that have less than target number of replicas, and "Mis-replicated 
blocks" are the blocks that are badly distributed.  So the Under-Replicated 
Blocks info on web UI and fsck result may be inconsistent.

Since Under-Replicated Blocks means higer missing risk for blocks, when threre 
is  no blocks that have less than target replicas but a lot of blocks that have 
the right number of blocks but are badlly distributed, the "Number of 
Under-Replicated Blocks" on web UI will be same as number of "Mis-replicated 
blocks", which is misleading for users. 
It would be clear to make "Number of Under-Replicated Blocks" on web UI be 
consistent with "Under-replicated blocks" in fsck result. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-12120) Use new block for pre-RollingUpgrade files` append requests

2017-07-11 Thread Vinayakumar B (JIRA)
Vinayakumar B created HDFS-12120:


 Summary: Use new block for pre-RollingUpgrade files` append 
requests
 Key: HDFS-12120
 URL: https://issues.apache.org/jira/browse/HDFS-12120
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Vinayakumar B


After the RollingUpgrade prepare, append on pre-RU files will re-open the same 
last block and makes changes to it (appending extra data, changing genstamp 
etc).
These changes to the block will not be tracked in Datanodes (either in trash or 
via hardlinks)

This creates problem if RollingUpgrade.Rollback is called.
Since block state and size both changed, after rollback block will be marked 
corrupted.

To avoid this, first time append on pre-RU files can be forced to write to new 
block itself.




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-11944) TestAclsEndToEnd#testCreateEncryptionZone failing very frequently.

2017-07-11 Thread Rushabh S Shah (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11944?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rushabh S Shah resolved HDFS-11944.
---
  Resolution: Invalid
Release Note: 
It was not a flaky test.
Fixed it via HADOOP-14563

> TestAclsEndToEnd#testCreateEncryptionZone failing very frequently.
> --
>
> Key: HDFS-11944
> URL: https://issues.apache.org/jira/browse/HDFS-11944
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: encryption, test
>Affects Versions: 2.8.0
>Reporter: Rushabh S Shah
>
> TestAclsEndToEnd#testCreateEncryptionZone is failing v frequently.
> The way test is written makes very hard to debug.
> Ideally each test case should test only one behavior.
> But in this test case, it reset the dfs state many times in same test case.
> It fails with the following stack trace.
> {noformat}
> Tests run: 9, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 35.17 sec <<< 
> FAILURE! - in org.apache.hadoop.hdfs.TestAclsEndToEnd
> testCreateEncryptionZone(org.apache.hadoop.hdfs.TestAclsEndToEnd)  Time 
> elapsed: 3.844 sec  <<< FAILURE!
> java.lang.AssertionError: Allowed zone creation of zone with blacklisted 
> GENERATE_EEK
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at org.junit.Assert.assertFalse(Assert.java:64)
>   at 
> org.apache.hadoop.hdfs.TestAclsEndToEnd.testCreateEncryptionZone(TestAclsEndToEnd.java:753)
> Results :
> Failed tests: 
>   TestAclsEndToEnd.testCreateEncryptionZone:753 Allowed zone creation of zone 
> with blacklisted GENERATE_EEK
> {noformat}
> It failed in the following pre-commits.
> [HDFS-11885 
> precommit|https://issues.apache.org/jira/browse/HDFS-11885?focusedCommentId=16040117&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-16040117]
> [HDFS-11804 
> precommit|https://issues.apache.org/jira/browse/HDFS-11804?focusedCommentId=16039872&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-16039872]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86

2017-07-11 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/461/

[Jul 10, 2017 10:53:13 AM] (stevel) HADOOP-14634. Remove jline from main Hadoop 
pom.xml. Contributed by Ray
[Jul 10, 2017 9:34:58 PM] (Arun Suresh) YARN-6776. Refactor 
ApplicaitonMasterService to move actual processing
[Jul 11, 2017 12:48:27 AM] (jitendra) HADOOP-10829. Iteration on 
CredentialProviderFactory.serviceLoader is
[Jul 11, 2017 4:30:13 AM] (aajisaka) HADOOP-14638. Replace commons-logging APIs 
with slf4j in StreamPumper.




-1 overall


The following subsystems voted -1:
findbugs unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

FindBugs :

   module:hadoop-hdfs-project/hadoop-hdfs-client 
   Possible exposure of partially initialized object in 
org.apache.hadoop.hdfs.DFSClient.initThreadsNumForStripedReads(int) At 
DFSClient.java:object in 
org.apache.hadoop.hdfs.DFSClient.initThreadsNumForStripedReads(int) At 
DFSClient.java:[line 2888] 
   org.apache.hadoop.hdfs.server.protocol.SlowDiskReports.equals(Object) 
makes inefficient use of keySet iterator instead of entrySet iterator At 
SlowDiskReports.java:keySet iterator instead of entrySet iterator At 
SlowDiskReports.java:[line 105] 

FindBugs :

   module:hadoop-hdfs-project/hadoop-hdfs 
   Possible null pointer dereference in 
org.apache.hadoop.hdfs.qjournal.server.JournalNode.getJournalsStatus() due to 
return value of called method Dereferenced at 
JournalNode.java:org.apache.hadoop.hdfs.qjournal.server.JournalNode.getJournalsStatus()
 due to return value of called method Dereferenced at JournalNode.java:[line 
302] 
   
org.apache.hadoop.hdfs.server.common.HdfsServerConstants$StartupOption.setClusterId(String)
 unconditionally sets the field clusterId At HdfsServerConstants.java:clusterId 
At HdfsServerConstants.java:[line 193] 
   
org.apache.hadoop.hdfs.server.common.HdfsServerConstants$StartupOption.setForce(int)
 unconditionally sets the field force At HdfsServerConstants.java:force At 
HdfsServerConstants.java:[line 217] 
   
org.apache.hadoop.hdfs.server.common.HdfsServerConstants$StartupOption.setForceFormat(boolean)
 unconditionally sets the field isForceFormat At 
HdfsServerConstants.java:isForceFormat At HdfsServerConstants.java:[line 229] 
   
org.apache.hadoop.hdfs.server.common.HdfsServerConstants$StartupOption.setInteractiveFormat(boolean)
 unconditionally sets the field isInteractiveFormat At 
HdfsServerConstants.java:isInteractiveFormat At HdfsServerConstants.java:[line 
237] 
   Possible null pointer dereference in 
org.apache.hadoop.hdfs.server.datanode.DataStorage.linkBlocksHelper(File, File, 
int, HardLink, boolean, File, List) due to return value of called method 
Dereferenced at 
DataStorage.java:org.apache.hadoop.hdfs.server.datanode.DataStorage.linkBlocksHelper(File,
 File, int, HardLink, boolean, File, List) due to return value of called method 
Dereferenced at DataStorage.java:[line 1339] 
   Possible null pointer dereference in 
org.apache.hadoop.hdfs.server.namenode.NNStorageRetentionManager.purgeOldLegacyOIVImages(String,
 long) due to return value of called method Dereferenced at 
NNStorageRetentionManager.java:org.apache.hadoop.hdfs.server.namenode.NNStorageRetentionManager.purgeOldLegacyOIVImages(String,
 long) due to return value of called method Dereferenced at 
NNStorageRetentionManager.java:[line 258] 
   Possible null pointer dereference in 
org.apache.hadoop.hdfs.server.namenode.NNUpgradeUtil$1.visitFile(Path, 
BasicFileAttributes) due to return value of called method Dereferenced at 
NNUpgradeUtil.java:org.apache.hadoop.hdfs.server.namenode.NNUpgradeUtil$1.visitFile(Path,
 BasicFileAttributes) due to return value of called method Dereferenced at 
NNUpgradeUtil.java:[line 133] 
   Useless condition:argv.length >= 1 at this point At DFSAdmin.java:[line 
2085] 
   Useless condition:numBlocks == -1 at this point At 
ImageLoaderCurrent.java:[line 727] 

FindBugs :

   
module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager
 
   Useless object stored in variable removedNullContainers of method 
org.apache.hadoop.yarn.server.nodemanager.NodeStatusUpdaterImpl.removeOrTrackCompletedContainersFromContext(List)
 At NodeStatusUpdaterImpl.java:removedNullContainers of method 
org.apache.hadoop.yarn.server.nodemanager.NodeStatusUpdaterImpl.removeOrTrackCompletedContainersFromContext(List)
 At NodeStatusUpdaterImpl.java:[line 642] 
   
org.apache.hadoop.yarn.server.nodemanager.NodeStatusUpdaterImpl.removeVeryOldStoppedContainersFromCache()
 makes inefficient use of keySet iterator instead of entrySet iterator At 
NodeStatusUpdaterImpl.java:keySet iterator instead of entrySet iterator At

[jira] [Created] (HDFS-12122) TestEditLogTailer is flaky

2017-07-11 Thread John Zhuge (JIRA)
John Zhuge created HDFS-12122:
-

 Summary: TestEditLogTailer is flaky
 Key: HDFS-12122
 URL: https://issues.apache.org/jira/browse/HDFS-12122
 Project: Hadoop HDFS
  Issue Type: Test
  Components: namenode
Affects Versions: 3.0.0-alpha4
Reporter: John Zhuge
Priority: Minor


https://builds.apache.org/job/PreCommit-HDFS-Build/20225/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
{noformat}
Tests run: 12, Failures: 5, Errors: 0, Skipped: 0, Time elapsed: 46.901 sec <<< 
FAILURE! - in org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer
testTriggersLogRollsForAllStandbyNN[0](org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer)
  Time elapsed: 10.728 sec  <<< FAILURE!
java.lang.AssertionError: Test resulted in an unexpected exit
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1957)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1944)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1937)
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer.testTriggersLogRollsForAllStandbyNN(TestEditLogTailer.java:245)

testNN1TriggersLogRolls[1](org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer)
  Time elapsed: 0.394 sec  <<< FAILURE!
java.lang.AssertionError: Test resulted in an unexpected exit
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1957)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1944)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1937)
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer.testStandbyTriggersLogRolls(TestEditLogTailer.java:201)
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer.testNN1TriggersLogRolls(TestEditLogTailer.java:154)

testNN2TriggersLogRolls[1](org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer)
  Time elapsed: 1.287 sec  <<< FAILURE!
java.lang.AssertionError: Test resulted in an unexpected exit
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1957)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1944)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1937)
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer.testStandbyTriggersLogRolls(TestEditLogTailer.java:201)
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer.testNN2TriggersLogRolls(TestEditLogTailer.java:159)

testNN0TriggersLogRolls[1](org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer)
  Time elapsed: 1.286 sec  <<< FAILURE!
java.lang.AssertionError: Test resulted in an unexpected exit
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1957)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1944)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1937)
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer.testStandbyTriggersLogRolls(TestEditLogTailer.java:201)
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer.testNN0TriggersLogRolls(TestEditLogTailer.java:149)

testTriggersLogRollsForAllStandbyNN[1](org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer)
  Time elapsed: 10.469 sec  <<< FAILURE!
java.lang.AssertionError: Test resulted in an unexpected exit
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1957)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1944)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1937)
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestEditLogTailer.testTriggersLogRollsForAllStandbyNN(TestEditLogTailer.java:245)
{noformat}




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-12123) Ozone: OzoneClient: Abstraction of OzoneClient and default implementation

2017-07-11 Thread Nandakumar (JIRA)
Nandakumar created HDFS-12123:
-

 Summary: Ozone: OzoneClient: Abstraction of OzoneClient and 
default implementation
 Key: HDFS-12123
 URL: https://issues.apache.org/jira/browse/HDFS-12123
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ozone
Reporter: Nandakumar
Assignee: Nandakumar


{{OzoneClient}} interface defines all the client operations supported by Ozone. 
{{OzoneClientImpl}} will have the default implementation, it should connects to 
KSM, SCM and DataNode through RPC to execute client calls.
Similarly we should have a client implementation which implements 
{{OzoneClient}} and uses REST protocol to execute client calls.
This will provide lots of flexibility to Ozone applications, when applications 
are running inside the cluster, they can use RPC protocol, but when running 
from outside the cluster, the same applications can speak REST protocol to 
communicate with Ozone.




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-12121) Audit logging of unsuccessful namesystem calls should not be done while holding fslock.

2017-07-11 Thread Rushabh S Shah (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-12121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rushabh S Shah resolved HDFS-12121.
---
Resolution: Duplicate

Dup of HDFS-11246

> Audit logging of unsuccessful namesystem calls should not be done while 
> holding fslock.
> ---
>
> Key: HDFS-12121
> URL: https://issues.apache.org/jira/browse/HDFS-12121
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.7.0
>Reporter: Rushabh S Shah
>Assignee: Rushabh S Shah
>
> Currently we are audit logging the unsuccessful namesystem calls while 
> holding fslock.
> Following is just one example.
> {code:title=FSNamesystem.java|borderStyle=solid}
> HdfsFileStatus getFileInfo(final String src, boolean resolveLink)
> throws IOException {
> final String operationName = "getfileinfo";
> checkOperation(OperationCategory.READ);
> HdfsFileStatus stat = null;
> readLock();
> try {
>   checkOperation(OperationCategory.READ);
>   stat = FSDirStatAndListingOp.getFileInfo(dir, src, resolveLink);
> } catch (AccessControlException e) {
>   logAuditEvent(false, operationName, src); //< audit logging while 
> holding fs read lock.
>   throw e;
> } finally {
>   readUnlock(operationName);
> }
> logAuditEvent(true, operationName, src);
> return stat;
>   }
> {code}
> We should avoid doing this.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-9590) NPE in Storage$StorageDirectory#unlock()

2017-07-11 Thread Xiao Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9590?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiao Chen resolved HDFS-9590.
-
Resolution: Cannot Reproduce

Haven't seen this since, closing...

> NPE in Storage$StorageDirectory#unlock()
> 
>
> Key: HDFS-9590
> URL: https://issues.apache.org/jira/browse/HDFS-9590
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Xiao Chen
>Assignee: Xiao Chen
> Attachments: HDFS-9590.01.patch
>
>
> The code looks to be possible to have race conditions in multiple-threaded 
> runs.
> {code}
> public void unlock() throws IOException {
>   if (this.lock == null)
> return;
>   this.lock.release();
>   lock.channel().close();
>   lock = null;
> }
> {code}
> This is called in a handful of places, and I don't see any protection. Shall 
> we add some synchronization mechanism? Not sure if I missed any design 
> assumptions here.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-12121) Audit logging of unsuccessful should not be done while holding fslock.

2017-07-11 Thread Rushabh S Shah (JIRA)
Rushabh S Shah created HDFS-12121:
-

 Summary: Audit logging of unsuccessful should not be done while 
holding fslock.
 Key: HDFS-12121
 URL: https://issues.apache.org/jira/browse/HDFS-12121
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.7.0
Reporter: Rushabh S Shah
Assignee: Rushabh S Shah


Currently we are audit logging the unsuccessful calls while holding fslock.
Following is just one example.
{code:title=FSNamesystem.java|borderStyle=solid}
HdfsFileStatus getFileInfo(final String src, boolean resolveLink)
throws IOException {
final String operationName = "getfileinfo";
checkOperation(OperationCategory.READ);
HdfsFileStatus stat = null;
readLock();
try {
  checkOperation(OperationCategory.READ);
  stat = FSDirStatAndListingOp.getFileInfo(dir, src, resolveLink);
} catch (AccessControlException e) {
  logAuditEvent(false, operationName, src); //< audit logging while 
holding fs read lock.
  throw e;
} finally {
  readUnlock(operationName);
}
logAuditEvent(true, operationName, src);
return stat;
  }
{code}

We should avoid doing this.




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-12124) Namenode should throw a RetryStartFileException exception instead of making a synchronous call to kms incase of cache miss.

2017-07-11 Thread Rushabh S Shah (JIRA)
Rushabh S Shah created HDFS-12124:
-

 Summary: Namenode should throw a RetryStartFileException exception 
instead of making a synchronous call to kms incase of cache miss.
 Key: HDFS-12124
 URL: https://issues.apache.org/jira/browse/HDFS-12124
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: kms
Affects Versions: 2.7.0
Reporter: Rushabh S Shah
Assignee: Rushabh S Shah


The namenode should throw a RetryStartFileException exception instead of making 
a synchronous call to kms incase of cache miss. Other kms issues (like kms 
getting stalled, issues talking to underlying kms) will block all the namenode 
handler threads.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/ppc64le

2017-07-11 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/372/

[Jul 10, 2017 9:34:58 PM] (Arun Suresh) YARN-6776. Refactor 
ApplicaitonMasterService to move actual processing
[Jul 11, 2017 12:48:27 AM] (jitendra) HADOOP-10829. Iteration on 
CredentialProviderFactory.serviceLoader is
[Jul 11, 2017 4:30:13 AM] (aajisaka) HADOOP-14638. Replace commons-logging APIs 
with slf4j in StreamPumper.
[Jul 11, 2017 9:22:44 AM] (sunilg) YARN-6714. IllegalStateException while 
handling APP_ATTEMPT_REMOVED
[Jul 11, 2017 12:40:11 PM] (yqlin) HDFS-12085. Reconfigure namenode heartbeat 
interval fails if the




-1 overall


The following subsystems voted -1:
compile mvninstall unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc javac


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

Failed junit tests :

   hadoop.hdfs.TestEncryptionZonesWithKMS 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailureWithRandomECPolicy 
   hadoop.hdfs.server.namenode.TestDecommissioningStatus 
   hadoop.hdfs.tools.offlineImageViewer.TestOfflineImageViewer 
   hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure 
   hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure090 
   hadoop.hdfs.server.namenode.ha.TestStandbyCheckpoints 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure080 
   hadoop.hdfs.web.TestWebHdfsTimeouts 
   hadoop.hdfs.server.datanode.fsdataset.impl.TestSpaceReservation 
   hadoop.hdfs.TestDFSStripedInputStreamWithRandomECPolicy 
   hadoop.yarn.server.nodemanager.recovery.TestNMLeveldbStateStoreService 
   hadoop.yarn.server.nodemanager.containermanager.TestContainerManager 
   hadoop.yarn.server.nodemanager.TestNodeManagerShutdown 
   hadoop.yarn.server.timeline.TestRollingLevelDB 
   hadoop.yarn.server.timeline.TestTimelineDataManager 
   hadoop.yarn.server.timeline.TestLeveldbTimelineStore 
   hadoop.yarn.server.timeline.recovery.TestLeveldbTimelineStateStore 
   hadoop.yarn.server.timeline.TestRollingLevelDBTimelineStore 
   
hadoop.yarn.server.applicationhistoryservice.TestApplicationHistoryServer 
   hadoop.yarn.server.resourcemanager.monitor.TestSchedulingMonitor 
   hadoop.yarn.server.resourcemanager.TestRMEmbeddedElector 
   hadoop.yarn.server.resourcemanager.security.TestDelegationTokenRenewer 
   
hadoop.yarn.server.resourcemanager.scheduler.capacity.TestIncreaseAllocationExpirer
 
   hadoop.yarn.server.resourcemanager.recovery.TestLeveldbRMStateStore 
   hadoop.yarn.server.resourcemanager.TestRMRestart 
   
hadoop.yarn.server.resourcemanager.TestOpportunisticContainerAllocatorAMService 
   hadoop.yarn.server.TestDiskFailures 
   hadoop.yarn.server.TestMiniYarnClusterNodeUtilization 
   hadoop.yarn.server.TestContainerManagerSecurity 
   hadoop.yarn.client.api.impl.TestNMClient 
   hadoop.yarn.server.timeline.TestLevelDBCacheTimelineStore 
   hadoop.yarn.server.timeline.TestOverrideTimelineStoreYarnClient 
   hadoop.yarn.server.timeline.TestEntityGroupFSTimelineStore 
   hadoop.yarn.applications.distributedshell.TestDistributedShell 
   hadoop.mapred.TestShuffleHandler 
   hadoop.mapreduce.v2.hs.TestHistoryServerLeveldbStateStoreService 
   hadoop.mapreduce.TestMRJobClient 
   hadoop.tools.TestDistCpSystem 

Timed out junit tests :

   org.apache.hadoop.hdfs.TestLeaseRecovery2 
   org.apache.hadoop.hdfs.qjournal.client.TestQJMWithFaults 
   org.apache.hadoop.hdfs.server.datanode.TestFsDatasetCache 
   org.apache.hadoop.yarn.server.resourcemanager.TestRMStoreCommands 
   
org.apache.hadoop.yarn.server.resourcemanager.recovery.TestZKRMStateStore 
   
org.apache.hadoop.yarn.server.resourcemanager.TestSubmitApplicationWithRMHA 
   
org.apache.hadoop.yarn.server.resourcemanager.TestKillApplicationWithRMHA 
  

   mvninstall:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/372/artifact/out/patch-mvninstall-root.txt
  [620K]

   compile:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/372/artifact/out/patch-compile-root.txt
  [20K]

   cc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/372/artifact/out/patch-compile-root.txt
  [20K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/372/artifact/out/patch-compile-root.txt
  [20K]

   unit:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/372/artifact/out/patch-unit-hadoop-assemblies.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/372/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
  [688K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/372/artifact/out/patch-unit-hadoop-yarn-p

[jira] [Created] (HDFS-12125) Document the missing -removePolicy command of ec

2017-07-11 Thread Wenxin He (JIRA)
Wenxin He created HDFS-12125:


 Summary: Document the missing -removePolicy command of ec
 Key: HDFS-12125
 URL: https://issues.apache.org/jira/browse/HDFS-12125
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation, erasure-coding
Affects Versions: 3.0.0-alpha4
Reporter: Wenxin He
Assignee: Wenxin He


Document the missing command -removePolicy in HDFSErasureCoding.md and 
HDFSCommands.md and regroup the ec commands to improve the user experience.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-12109) "fs" java.net.UnknownHostException when HA NameNode is used

2017-07-11 Thread Luigi Di Fraia (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-12109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Luigi Di Fraia resolved HDFS-12109.
---
Resolution: Not A Bug

> "fs" java.net.UnknownHostException when HA NameNode is used
> ---
>
> Key: HDFS-12109
> URL: https://issues.apache.org/jira/browse/HDFS-12109
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 2.8.0
> Environment: [hadoop@namenode01 ~]$ cat /etc/redhat-release
> CentOS Linux release 7.3.1611 (Core)
> [hadoop@namenode01 ~]$ uname -a
> Linux namenode01 3.10.0-514.10.2.el7.x86_64 #1 SMP Fri Mar 3 00:04:05 UTC 
> 2017 x86_64 x86_64 x86_64 GNU/Linux
> [hadoop@namenode01 ~]$ java -version
> java version "1.8.0_131"
> Java(TM) SE Runtime Environment (build 1.8.0_131-b11)
> Java HotSpot(TM) 64-Bit Server VM (build 25.131-b11, mixed mode)
>Reporter: Luigi Di Fraia
>
> After setting up an HA NameNode configuration, the following invocation of 
> "fs" fails:
> [hadoop@namenode01 ~]$ /usr/local/hadoop/bin/hdfs dfs -ls /
> -ls: java.net.UnknownHostException: saccluster
> It works if properties are defined as per below:
> /usr/local/hadoop/bin/hdfs dfs -Ddfs.nameservices=saccluster 
> -Ddfs.client.failover.proxy.provider.saccluster=org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider
>  -Ddfs.ha.namenodes.saccluster=namenode01,namenode02 
> -Ddfs.namenode.rpc-address.saccluster.namenode01=namenode01:8020 
> -Ddfs.namenode.rpc-address.saccluster.namenode02=namenode02:8020 -ls /
> These properties are defined in /usr/local/hadoop/etc/hadoop/hdfs-site.xml as 
> per below:
> 
> dfs.nameservices
> saccluster
> 
> 
> dfs.ha.namenodes.saccluster
> namenode01,namenode02
> 
> 
> dfs.namenode.rpc-address.saccluster.namenode01
> namenode01:8020
> 
> 
> dfs.namenode.rpc-address.saccluster.namenode02
> namenode02:8020
> 
> 
> dfs.namenode.http-address.saccluster.namenode01
> namenode01:50070
> 
> 
> dfs.namenode.http-address.saccluster.namenode02
> namenode02:50070
> 
> 
> dfs.namenode.shared.edits.dir
> 
> qjournal://namenode01:8485;namenode02:8485;datanode01:8485/saccluster
> 
> 
> dfs.client.failover.proxy.provider.mycluster
> 
> org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider
> 
> In /usr/local/hadoop/etc/hadoop/core-site.xml the default FS is defined as 
> per below:
> 
> fs.defaultFS
> hdfs://saccluster
> 
> In /usr/local/hadoop/etc/hadoop/hadoop-env.sh the following export is defined:
> export HADOOP_CONF_DIR="/usr/local/hadoop/etc/hadoop"
> Is "fs" trying to read these properties from somewhere else, such as a 
> separate client configuration file?
> Apologies if I am missing something obvious here.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org