[jira] [Reopened] (HDFS-14158) Checkpointer ignores configured time period > 5 minutes

2019-01-16 Thread Timo Walter (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDFS-14158?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Timo Walter reopened HDFS-14158:


The explanation of [~ajisakaa] doesn't make sense to me. In my opinion, the 
issue is already relevant for the functionality.

> Checkpointer ignores configured time period > 5 minutes
> ---
>
> Key: HDFS-14158
> URL: https://issues.apache.org/jira/browse/HDFS-14158
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.8.1
>Reporter: Timo Walter
>Priority: Minor
>  Labels: checkpoint, hdfs, namenode
>
> The checkpointer always triggers a checkpoint every 5 minutes and ignores the 
> flag "*dfs.namenode.checkpoint.period*", if its greater than 5 minutes.
> See the code below (in Checkpointer.java):
> {code:java}
> //Main work loop of the Checkpointer
> public void run() {
>   // Check the size of the edit log once every 5 minutes.
>   long periodMSec = 5 * 60;   // 5 minutes
>   if(checkpointConf.getPeriod() < periodMSec) {
> periodMSec = checkpointConf.getPeriod();
>   }
> {code}
> If the configured period ("*dfs.namenode.checkpoint.period*") is lower than 5 
> minutes, you choose use the configured one. But it always ignores it, if it's 
> greater than 5 minutes.
>  
> In my opinion, the if-expression should be:
> {code:java}
> if(checkpointConf.getPeriod() > periodMSec) {
> periodMSec = checkpointConf.getPeriod();
>   }
> {code}
>  
> Then "*dfs.namenode.checkpoint.period*" won't get ignored.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: [VOTE] Release Apache Hadoop 3.2.0 - RC1

2019-01-16 Thread Sunil G
Thanks everyone for helping to vote this release!

With 7 binding votes, 10 non-binding votes and no veto, this vote stands
 passed,
I'm going to work on staging the release.

Thanks,
Sunil

On Tue, Jan 15, 2019 at 9:59 PM Weiwei Yang  wrote:

> +1 (binding)
>
> - Setup a cluster, run teragen/terasort jobs
> - Verified general readability of documentation (titles/navigations)
> - Run some simple yarn commands: app/applicationattempt/container
> - Checked restful APIs: RM cluster/metrics/scheduler/nodes, NM
> node/apps/container
> - Verified simple failover scenario
> - Submitted distributed shell apps with affinity/anti-affinity constraints
> - Configured conf based node attribute provider, alter attribute values
> and verified the change
> - Verified CLI add/list/remove node-attributes, submitted app with simple
> node-attribute constraint
>
> --
> Weiwei
>


Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86

2019-01-16 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/

[Jan 15, 2019 3:59:59 AM] (aajisaka) YARN-9179. Fix NPE in 
AbstractYarnScheduler#updateNewContainerInfo.
[Jan 15, 2019 1:15:18 PM] (stevel) HADOOP-16019. ZKDelegationTokenSecretManager 
won't log exception message
[Jan 15, 2019 11:11:16 PM] (bharat) HDDS-978. Fix typo in doc : Client > S3 
section. Contributed by  Dinesh
[Jan 16, 2019 1:28:08 AM] (aajisaka) HADOOP-15941. Addendum patch. Contributed 
by Takanobu Asanuma.




-1 overall


The following subsystems voted -1:
asflicense findbugs hadolint pathlen unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

Failed junit tests :

   hadoop.hdfs.server.balancer.TestBalancerWithMultipleNameNodes 
   hadoop.hdfs.web.TestWebHdfsTimeouts 
  

   cc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/diff-compile-cc-root.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/diff-compile-javac-root.txt
  [336K]

   checkstyle:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/diff-checkstyle-root.txt
  [17M]

   hadolint:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/diff-patch-hadolint.txt
  [4.0K]

   pathlen:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/pathlen.txt
  [12K]

   pylint:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/diff-patch-pylint.txt
  [60K]

   shellcheck:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/diff-patch-shellcheck.txt
  [20K]

   shelldocs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/diff-patch-shelldocs.txt
  [12K]

   whitespace:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/whitespace-eol.txt
  [9.3M]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/whitespace-tabs.txt
  [1.1M]

   findbugs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-hdds_client.txt
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-hdds_container-service.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-hdds_framework.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-hdds_server-scm.txt
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-hdds_tools.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-ozone_client.txt
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-ozone_common.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-ozone_objectstore-service.txt
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-ozone_ozone-manager.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-ozone_ozonefs.txt
  [12K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-ozone_s3gateway.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/branch-findbugs-hadoop-ozone_tools.txt
  [8.0K]

   javadoc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/diff-javadoc-javadoc-root.txt
  [752K]

   unit:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
  [328K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
  [84K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-applications_hadoop-yarn-applications-distributedshell.txt
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1018/artifact/out/patch-unit-hadoop-mapreduce-project_hadoop-mapreduce-client_hadoop-mapreduce-client-jobclient.txt
  [84K]
   
https://builds.apache

Re: [VOTE] Release Apache Hadoop 3.2.0 - RC1

2019-01-16 Thread Rohith Sharma K S
+1 (binding) Apologies for late vote!

- Built from source for hbase profile=2.0 and deployed RM HA cluster along
with ATSv2.0 enabled.
- Ran sample jobs and verified all ATSv2.0 data are publishing.
- Verified UI2 for ATSv2.0 data!

-Rohith Sharma K S

On Wed, 16 Jan 2019 at 07:00, Sunil G  wrote:

> Thanks everyone for helping to vote this release!
>
> With 7 binding votes, 10 non-binding votes and no veto, this vote stands
>  passed,
> I'm going to work on staging the release.
>
> Thanks,
> Sunil
>
> On Tue, Jan 15, 2019 at 9:59 PM Weiwei Yang  wrote:
>
> > +1 (binding)
> >
> > - Setup a cluster, run teragen/terasort jobs
> > - Verified general readability of documentation (titles/navigations)
> > - Run some simple yarn commands: app/applicationattempt/container
> > - Checked restful APIs: RM cluster/metrics/scheduler/nodes, NM
> > node/apps/container
> > - Verified simple failover scenario
> > - Submitted distributed shell apps with affinity/anti-affinity
> constraints
> > - Configured conf based node attribute provider, alter attribute values
> > and verified the change
> > - Verified CLI add/list/remove node-attributes, submitted app with simple
> > node-attribute constraint
> >
> > --
> > Weiwei
> >
>


Re: [VOTE] Release Apache Hadoop 3.2.0 - RC1

2019-01-16 Thread Naganarasimha Garla
+1 (binding) Apologies for one more late vote!

- Built from source and also installed the release tar.
- Verified all Node Labels and node .
-  Ran sample jobs and verified UI

Thanks and Regards,
+ Naga

On Wed, Jan 16, 2019 at 10:34 PM Rohith Sharma K S <
rohithsharm...@apache.org> wrote:

> +1 (binding) Apologies for late vote!
>
> - Built from source for hbase profile=2.0 and deployed RM HA cluster along
> with ATSv2.0 enabled.
> - Ran sample jobs and verified all ATSv2.0 data are publishing.
> - Verified UI2 for ATSv2.0 data!
>
> -Rohith Sharma K S
>
> On Wed, 16 Jan 2019 at 07:00, Sunil G  wrote:
>
> > Thanks everyone for helping to vote this release!
> >
> > With 7 binding votes, 10 non-binding votes and no veto, this vote stands
> >  passed,
> > I'm going to work on staging the release.
> >
> > Thanks,
> > Sunil
> >
> > On Tue, Jan 15, 2019 at 9:59 PM Weiwei Yang  wrote:
> >
> > > +1 (binding)
> > >
> > > - Setup a cluster, run teragen/terasort jobs
> > > - Verified general readability of documentation (titles/navigations)
> > > - Run some simple yarn commands: app/applicationattempt/container
> > > - Checked restful APIs: RM cluster/metrics/scheduler/nodes, NM
> > > node/apps/container
> > > - Verified simple failover scenario
> > > - Submitted distributed shell apps with affinity/anti-affinity
> > constraints
> > > - Configured conf based node attribute provider, alter attribute values
> > > and verified the change
> > > - Verified CLI add/list/remove node-attributes, submitted app with
> simple
> > > node-attribute constraint
> > >
> > > --
> > > Weiwei
> > >
> >
>


[jira] [Created] (HDFS-14211) [Consistent Observer Reads] Allow for configurable "always msync" mode

2019-01-16 Thread Erik Krogen (JIRA)
Erik Krogen created HDFS-14211:
--

 Summary: [Consistent Observer Reads] Allow for configurable 
"always msync" mode
 Key: HDFS-14211
 URL: https://issues.apache.org/jira/browse/HDFS-14211
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Reporter: Erik Krogen


To allow for reads to be serviced from an ObserverNode (see HDFS-12943) in a 
consistent way, an {{msync}} API was introduced (HDFS-13688) to allow for a 
client to fetch the latest transaction ID from the Active NN, thereby ensuring 
that subsequent reads from the ObserverNode will be up-to-date with the current 
state of the Active.

Using this properly, however, requires application-side changes: for examples, 
a NodeManager should call {{msync}} before localizing the resources for a 
client, since it received notification of the existence of those resources via 
communicate which is out-of-band to HDFS and thus could potentially attempt to 
localize them prior to the availability of those resources on the ObserverNode.

Until such application-side changes can be made, which will be a longer-term 
effort, we need to provide a mechanism for unchanged clients to utilize the 
ObserverNode without exposing such a client to inconsistencies. This is 
essentially phase 3 of the roadmap outlined in the [design 
document|https://issues.apache.org/jira/secure/attachment/12915990/ConsistentReadsFromStandbyNode.pdf]
 for HDFS-12943.

The design document proposes some heuristics based on understanding of how 
common applications (e.g. MR) use HDFS for resources. As an initial pass, we 
can simply have a flag which tells a client to call {{msync}} before _every 
single_ read operation. This may seem counterintuitive, as it turns every read 
operation into two RPCs: {{msync}} to the Active following by an actual read 
operation to the Observer. However, the {{msync}} operation is extremely 
lightweight, as it does not acquire the {{FSNamesystemLock}}, and in 
experiments we have found that this approach can easily scale to well over 
100,000 {{msync}} operations per second on the Active (while still servicing 
approx. 10,000 write op/s). Combined with the fast-path edit log tailing for 
standby/observer nodes (HDFS-13150), this "always msync" approach should 
introduce only a few ms of extra latency to each read call.

Below are some experimental results collected from experiments which convert a 
normal RPC workload into one in which all read operations are turned into an 
{{msync}}. The baseline is a workload of 1.5k write op/s and 25k read op/s.

||Rate Multiplier|2|4|6|8||
||RPC Queue Avg Time (ms)|14.2|53.2|110.4|125.3||
||RPC Queue NumOps Avg (k)|51.4|102.3|147.8|177.9||
||RPC Queue NumOps Max (k)|148.8|269.5|306.3|312.4||

Results are promising up to between 4x and 6x of the baseline workload, which 
is approx. 100-150k read op/s.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-14212) `hdfs crypto -listZones` should list encrypted zones from all namespaces

2019-01-16 Thread venkata naidu udamala (JIRA)
venkata naidu udamala created HDFS-14212:


 Summary: `hdfs crypto -listZones` should list encrypted zones from 
all namespaces
 Key: HDFS-14212
 URL: https://issues.apache.org/jira/browse/HDFS-14212
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: venkata naidu udamala


`hdfs crypto -listZones` is listing the encryption zones only from the default 
namespace. there is no way to list the encryption zones from other namespaces. 
It would be helpful if there is a way to list encryption zones from all 
namespaces or take any option to specify the namespace in the command



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-14212) `hdfs crypto -listZones` should list encrypted zones from all namespaces

2019-01-16 Thread venkata naidu udamala (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDFS-14212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

venkata naidu udamala resolved HDFS-14212.
--
  Resolution: Fixed
Release Note: This can be done using "fs" option  `hdfs crypto -fs  
hdfs:namespace1 -listZones`

> `hdfs crypto -listZones` should list encrypted zones from all namespaces
> 
>
> Key: HDFS-14212
> URL: https://issues.apache.org/jira/browse/HDFS-14212
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: venkata naidu udamala
>Priority: Major
>
> `hdfs crypto -listZones` is listing the encryption zones only from the 
> default namespace. there is no way to list the encryption zones from other 
> namespaces. It would be helpful if there is a way to list encryption zones 
> from all namespaces or take any option to specify the namespace in the command



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-981) Block allocation should involve pipeline selection and then container selection

2019-01-16 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-981:


 Summary: Block allocation should involve pipeline selection and 
then container selection
 Key: HDDS-981
 URL: https://issues.apache.org/jira/browse/HDDS-981
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently SCM maintains a list of preallocated containers and allocates blocks 
from these containers. This approach does not work well with dynamics of the 
cluster where new nodes are being added and pipelines are destroyed. New 
containers are not created until all the preallocated containers are exhausted.

The Jira aims to establish a criteria in block allocation where first a 
pipeline is selected amongst the available pipelines and then a container is 
selected in that pipeline. In order to handle the dynamics of the cluster a 
fixed interval pipeline creator job can be launched which creates pipelines in 
the system.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org