Apache Hadoop qbt Report: branch2.10+JDK7 on Linux/x86
For more details, see https://builds.apache.org/job/hadoop-qbt-branch-2.10-java7-linux-x86/755/ [Jul 21, 2020 1:02:41 PM] (iwasakims) YARN-6515. Fix warnings from Spotbugs in hadoop-yarn-server-nodemanager. -1 overall The following subsystems voted -1: asflicense findbugs hadolint jshint pathlen unit xml The following subsystems voted -1 but were configured to be filtered/ignored: cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace The following subsystems are considered long running: (runtime bigger than 1h 0m 0s) unit Specific tests: XML : Parsing Error(s): hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/empty-configuration.xml hadoop-tools/hadoop-azure/src/config/checkstyle-suppressions.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/public/crossdomain.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/public/crossdomain.xml findbugs : module:hadoop-yarn-project/hadoop-yarn Boxed value is unboxed and then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:[line 335] org.apache.hadoop.yarn.state.StateMachineFactory.generateStateGraph(String) makes inefficient use of keySet iterator instead of entrySet iterator At StateMachineFactory.java:keySet iterator instead of entrySet iterator At StateMachineFactory.java:[line 505] findbugs : module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common org.apache.hadoop.yarn.state.StateMachineFactory.generateStateGraph(String) makes inefficient use of keySet iterator instead of entrySet iterator At StateMachineFactory.java:keySet iterator instead of entrySet iterator At StateMachineFactory.java:[line 505] findbugs : module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server Boxed value is unboxed and then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:[line 335] findbugs : module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase Boxed value is unboxed and then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:[line 335] findbugs : module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client Boxed value is unboxed and then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:[line 335] findbugs : module:hadoop-yarn-project Boxed value is unboxed and then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:[line 335] org.apache.hadoop.yarn.state.StateMachineFactory.generateStateGraph(String) makes inefficient use of keySet iterator instead of entrySet iterator At StateMachineFactory.java:keySet iterator instead of entrySet iterator At StateMachineFactory.java:[line 505] findbugs : module:hadoop-mapreduce-project/hadoop-mapreduce-client Primitive is boxed to call Long.compareTo(Long):Long.compareTo(Long): use Long.compare(long, long) instead At JVMId.java
Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86_64
For more details, see https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/211/ [Jul 21, 2020 5:43:05 AM] (Shashikant Banerjee) HDFS-15470. Added more unit tests to validate rename behaviour across snapshots. Contributed by Shashikant Banerjee. [Jul 21, 2020 6:06:24 AM] (noreply) HDFS-15479. Ordered snapshot deletion: make it a configurable feature (#2156) [Jul 21, 2020 4:00:07 PM] (Iñigo Goiri) HDFS-15246. ArrayIndexOfboundsException in BlockManager CreateLocatedBlock. Contributed by Hemanth Boyina. [Jul 21, 2020 4:18:54 PM] (noreply) HADOOP-17092. ABFS: Making AzureADAuthenticator.getToken() throw HttpException [Jul 21, 2020 4:22:38 PM] (noreply) Hadoop 17132. ABFS: Fix Rename and Delete Idempotency check trigger -1 overall The following subsystems voted -1: asflicense findbugs pathlen unit xml The following subsystems voted -1 but were configured to be filtered/ignored: cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace The following subsystems are considered long running: (runtime bigger than 1h 0m 0s) unit Specific tests: XML : Parsing Error(s): hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-excerpt.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-missing-tags.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-missing-tags2.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-sample-output.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/resources/fair-scheduler-invalid.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/resources/yarn-site-with-invalid-allocation-file-ref.xml findbugs : module:hadoop-common-project/hadoop-common Dead store to $L5 in org.apache.hadoop.ipc.Server$ConnectionManager.decrUserConnections(String) At Server.java:org.apache.hadoop.ipc.Server$ConnectionManager.decrUserConnections(String) At Server.java:[line 3729] Dead store to $L5 in org.apache.hadoop.ipc.Server$ConnectionManager.incrUserConnections(String) At Server.java:org.apache.hadoop.ipc.Server$ConnectionManager.incrUserConnections(String) At Server.java:[line 3717] findbugs : module:hadoop-common-project Dead store to $L5 in org.apache.hadoop.ipc.Server$ConnectionManager.decrUserConnections(String) At Server.java:org.apache.hadoop.ipc.Server$ConnectionManager.decrUserConnections(String) At Server.java:[line 3729] Dead store to $L5 in org.apache.hadoop.ipc.Server$ConnectionManager.incrUserConnections(String) At Server.java:org.apache.hadoop.ipc.Server$ConnectionManager.incrUserConnections(String) At Server.java:[line 3717] findbugs : module:hadoop-hdfs-project/hadoop-hdfs Method org.apache.hadoop.hdfs.server.datanode.checker.DatasetVolumeChecker$ResultHandler.onSuccess(Object) overrides the nullness annotation of parameter $L1 in an incompatible way At DatasetVolumeChecker.java:annotation of parameter $L1 in an incompatible way At DatasetVolumeChecker.java:[line 322] Method org.apache.hadoop.hdfs.server.datanode.checker.DatasetVolumeChecker$ResultHandler.onSuccess(VolumeCheckResult) overrides the nullness annotation of parameter result in an incompatible way At DatasetVolumeChecker.java:annotation of parameter result in an incompatible way At DatasetVolumeChecker.java:[lines 358-376] Method org.apache.hadoop.hdfs.server.datanode.checker.ThrottledAsyncChecker$2.onSuccess(Object) overrides the nullness annotation of parameter result in an incompatible way At ThrottledAsyncChecker.java:annotation of parameter result in an incompatible way At ThrottledAsyncChecker.java:[lines 170-175] Dead store to $L8 in org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.incrOpCount(FSEditLogOpCodes, EnumMap, Step, StartupProgress$Counter) At FSEditLogLoader.java:org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.incrOpCount(FSEditLogOpCodes, EnumMap, Step, StartupProgress$Counter) At FSEditLogLoader.java:[line 1241] findbugs : module:hadoop-hdfs-project Method org.apache.hadoop.hdfs.server.datanode.checker.DatasetVolumeChecker$ResultHandler.onSuccess(Object) overrides the nullness annotation of parameter $L1 in an incompatible way At DatasetVolumeChecker.java:annotation of parameter $L1 in an incompatible way At DatasetVolumeChecker.java:[line 322] Method org.apache.hadoop.hdfs.server.datanode.checker.DatasetVolumeChecker$ResultHandler.onSuccess(VolumeCheckResult) overrides the nullness annotation of parameter result in an incompatible way At DatasetVolume
[ANNOUNCE] New Apache Hadoop PMC member : Ayush Saxena
Hi all, I am very glad to announce that Ayush Saxena was voted to join Apache Hadoop PMC. Congratulations Ayush! Well deserved and thank you for your dedication to the project. Please keep up the good work. Thanks, -Vinay
Re: [ANNOUNCE] New Apache Hadoop PMC member : Ayush Saxena
Congrats On Thu, Jul 23, 2020 at 1:10 AM Vinayakumar B wrote: > Hi all, > > I am very glad to announce that Ayush Saxena was voted to join Apache > Hadoop PMC. > > Congratulations Ayush! Well deserved and thank you for your dedication to > the project. Please keep up the good work. > > Thanks, > -Vinay >
Re: [ANNOUNCE] New Apache Hadoop PMC member : Ayush Saxena
Congratulations, Ayush.Keep up the good work. Thank you./Sree On Wednesday, July 22, 2020, 12:48:55 PM PDT, Team AMR wrote: Congrats On Thu, Jul 23, 2020 at 1:10 AM Vinayakumar B wrote: > Hi all, > > I am very glad to announce that Ayush Saxena was voted to join Apache > Hadoop PMC. > > Congratulations Ayush! Well deserved and thank you for your dedication to > the project. Please keep up the good work. > > Thanks, > -Vinay >
[jira] [Created] (HDFS-15489) Documentation link is broken for Apache Hadoop
Namit Maheshwari created HDFS-15489: --- Summary: Documentation link is broken for Apache Hadoop Key: HDFS-15489 URL: https://issues.apache.org/jira/browse/HDFS-15489 Project: Hadoop HDFS Issue Type: Bug Components: documentation Reporter: Namit Maheshwari Attachments: DocumentLinkBroken.mov Please see attached video and screenshots [^DocumentLinkBroken.mov] -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
Re: [ANNOUNCE] New Apache Hadoop PMC member : Ayush Saxena
Congratulations Ayush! Regards, He Xiaoqiao On Thu, Jul 23, 2020 at 9:18 AM Sheng Liu wrote: > Congrats, and thanks for your help. > > Sree Vaddi 于2020年7月23日周四 上午3:59写道: > > > Congratulations, Ayush.Keep up the good work. > > > > > > Thank you./Sree > > > > > > > > On Wednesday, July 22, 2020, 12:48:55 PM PDT, Team AMR < > > teamamr.apa...@gmail.com> wrote: > > > > Congrats > > > > > > On Thu, Jul 23, 2020 at 1:10 AM Vinayakumar B > > wrote: > > > > > Hi all, > > > > > > I am very glad to announce that Ayush Saxena was voted to join Apache > > > Hadoop PMC. > > > > > > Congratulations Ayush! Well deserved and thank you for your dedication > to > > > the project. Please keep up the good work. > > > > > > Thanks, > > > -Vinay > > > > > >
[jira] [Created] (HDFS-15490) Address checkstyle issues reported with HDFS-15480
Shashikant Banerjee created HDFS-15490: -- Summary: Address checkstyle issues reported with HDFS-15480 Key: HDFS-15490 URL: https://issues.apache.org/jira/browse/HDFS-15490 Project: Hadoop HDFS Issue Type: Sub-task Reporter: Shashikant Banerjee -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86_64
For more details, see https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/212/ [Jul 22, 2020 4:40:20 AM] (noreply) HADOOP-17138. Fix spotbugs warnings surfaced after upgrade to 4.0.6. (#2155) [Jul 22, 2020 6:29:10 AM] (noreply) HDFS-15478: When Empty mount points, we are assigning fallback link to self. But it should not use full URI for target fs. (#2160). Contributed by Uma Maheswara Rao G. [Jul 22, 2020 3:39:49 PM] (Akira Ajisaka) HADOOP-17147. Dead link in hadoop-kms/index.md.vm. Contributed by Xieming Li. [Jul 22, 2020 5:22:30 PM] (noreply) HADOOP-17113. Adding ReadAhead Counters in ABFS (#2154) [Jul 22, 2020 11:16:27 PM] (noreply) HDFS-15480. Ordered snapshot deletion: record snapshot deletion in XAttr (#2163) [Error replacing 'FILE' - Workspace is not accessible] - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
[jira] [Created] (HDFS-15491) Journal transactions batched in sync metrics are missing from Journal Node JMX
Mohit Saindane created HDFS-15491: - Summary: Journal transactions batched in sync metrics are missing from Journal Node JMX Key: HDFS-15491 URL: https://issues.apache.org/jira/browse/HDFS-15491 Project: Hadoop HDFS Issue Type: Bug Components: 3.1.1, hdfs, journal-node, qjm Affects Versions: 3.1.1 Reporter: Mohit Saindane Fix For: 3.1.1 Attachments: JournalNode metrics issue.png I have enable the HA in HDP3.1.5 multinode cluster, I was trying to access the REST response given by Journal Node on http 8480 port, but as per [Apache Hadoop 3.1.1 Documentation|[https://hadoop.apache.org/docs/r3.1.1/hadoop-project-dist/hadoop-common/Metrics.html#JournalNode]] I found that some of the Journal Node specific metrics are missing. (metrics like NumTransactionsBatchedInSync_(60,300,3600)_sNumOps and NumTransactionsBatchedInSync_Num_s_(50/75/90/95/99)_thPercentileLatencyMicros), So checked the _JournalMetrics.java_ code given in _hadoop-hdfs-3.1.1.3.1.5.0-152-sources.jar_ as well as __ [Hadoop Open Source Code|[https://github.com/apache/hadoop/blob/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalMetrics.java]] found that the following code is missing : @Metric("Journal transactions batched in sync") final MutableQuantiles[] numTransactionsBatchedInSync; -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org