[jira] [Created] (KUDU-3019) Kudu client hangs in Deffered.join()
Zoltán Borók-Nagy created KUDU-3019: --- Summary: Kudu client hangs in Deffered.join() Key: KUDU-3019 URL: https://issues.apache.org/jira/browse/KUDU-3019 Project: Kudu Issue Type: Bug Reporter: Zoltán Borók-Nagy In Impala we've seen the Kudu client hanging with the following stack trace: {noformat} Thread 53015: (state = BLOCKED) - java.lang.Object.wait(long) @bci=0 (Compiled frame; information may be imprecise) - java.lang.Object.wait() @bci=2, line=502 (Compiled frame) - com.stumbleupon.async.Deferred.doJoin(boolean, long) @bci=77, line=1122 (Compiled frame) - com.stumbleupon.async.Deferred.join() @bci=3, line=1006 (Compiled frame) - org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred) @bci=1, line=340 (Compiled frame) - org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, line=212 (Compiled frame) - org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer) @bci=32, line=115 (Compiled frame) ...{noformat} The client hangs in Deferred.join(): [https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423] To at least mitigate the problem, maybe Deferred.join(long timeout) could be used instead. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (KUDU-3019) Kudu client hangs in Deffered.join()
[ https://issues.apache.org/jira/browse/KUDU-3019?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zoltán Borók-Nagy updated KUDU-3019: Description: In Impala we've seen the Kudu client hanging with the following stack trace: {noformat} Thread 53015: (state = BLOCKED) - java.lang.Object.wait(long) @bci=0 (Compiled frame; information may be imprecise) - java.lang.Object.wait() @bci=2, line=502 (Compiled frame) - com.stumbleupon.async.Deferred.doJoin(boolean, long) @bci=77, line=1122 (Compiled frame) - com.stumbleupon.async.Deferred.join() @bci=3, line=1006 (Compiled frame) - org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred) @bci=1, line=340 (Compiled frame) - org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, line=212 (Compiled frame) - org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer) @bci=32, line=115 (Compiled frame) - org.apache.impala.planner.SingleNodePlanner.createScanNode(org.apache.impala.analysis.TableRef, org.apache.impala.analysis.AggregateInfo, org.apache.impala.analysis.Analyzer) @bci=252, line=1312 (Compiled frame) ...{noformat} The client hangs in Deferred.join(): [https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423] To at least mitigate the problem, maybe Deferred.join(long timeout) could be used instead. was: In Impala we've seen the Kudu client hanging with the following stack trace: {noformat} Thread 53015: (state = BLOCKED) - java.lang.Object.wait(long) @bci=0 (Compiled frame; information may be imprecise) - java.lang.Object.wait() @bci=2, line=502 (Compiled frame) - com.stumbleupon.async.Deferred.doJoin(boolean, long) @bci=77, line=1122 (Compiled frame) - com.stumbleupon.async.Deferred.join() @bci=3, line=1006 (Compiled frame) - org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred) @bci=1, line=340 (Compiled frame) - org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, line=212 (Compiled frame) - org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer) @bci=32, line=115 (Compiled frame) ...{noformat} The client hangs in Deferred.join(): [https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423] To at least mitigate the problem, maybe Deferred.join(long timeout) could be used instead. > Kudu client hangs in Deffered.join() > > > Key: KUDU-3019 > URL: https://issues.apache.org/jira/browse/KUDU-3019 > Project: Kudu > Issue Type: Bug >Reporter: Zoltán Borók-Nagy >Priority: Major > > In Impala we've seen the Kudu client hanging with the following stack trace: > {noformat} > Thread 53015: (state = BLOCKED) > - java.lang.Object.wait(long) @bci=0 (Compiled frame; information may be > imprecise) > - java.lang.Object.wait() @bci=2, line=502 (Compiled frame) > - com.stumbleupon.async.Deferred.doJoin(boolean, long) @bci=77, line=1122 > (Compiled frame) > - com.stumbleupon.async.Deferred.join() @bci=3, line=1006 (Compiled frame) > - > org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred) > @bci=1, line=340 (Compiled frame) > - org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, > line=212 (Compiled frame) > - > org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer) > @bci=32, line=115 (Compiled frame) > - > org.apache.impala.planner.SingleNodePlanner.createScanNode(org.apache.impala.analysis.TableRef, > org.apache.impala.analysis.AggregateInfo, > org.apache.impala.analysis.Analyzer) @bci=252, line=1312 (Compiled frame) > ...{noformat} > The client hangs in Deferred.join(): > [https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423] > To at least mitigate the problem, maybe Deferred.join(long timeout) could be > used instead. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (KUDU-3019) Kudu client hangs in Deffered.join()
[ https://issues.apache.org/jira/browse/KUDU-3019?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zoltán Borók-Nagy updated KUDU-3019: Affects Version/s: 1.6.0 > Kudu client hangs in Deffered.join() > > > Key: KUDU-3019 > URL: https://issues.apache.org/jira/browse/KUDU-3019 > Project: Kudu > Issue Type: Bug >Affects Versions: 1.6.0 >Reporter: Zoltán Borók-Nagy >Priority: Major > > In Impala we've seen the Kudu client hanging with the following stack trace: > {noformat} > Thread 53015: (state = BLOCKED) > - java.lang.Object.wait(long) @bci=0 (Compiled frame; information may be > imprecise) > - java.lang.Object.wait() @bci=2, line=502 (Compiled frame) > - com.stumbleupon.async.Deferred.doJoin(boolean, long) @bci=77, line=1122 > (Compiled frame) > - com.stumbleupon.async.Deferred.join() @bci=3, line=1006 (Compiled frame) > - > org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred) > @bci=1, line=340 (Compiled frame) > - org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, > line=212 (Compiled frame) > - > org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer) > @bci=32, line=115 (Compiled frame) > - > org.apache.impala.planner.SingleNodePlanner.createScanNode(org.apache.impala.analysis.TableRef, > org.apache.impala.analysis.AggregateInfo, > org.apache.impala.analysis.Analyzer) @bci=252, line=1312 (Compiled frame) > ...{noformat} > The client hangs in Deferred.join(): > [https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423] > To at least mitigate the problem, maybe Deferred.join(long timeout) could be > used instead. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (KUDU-3019) Kudu client hangs in Deffered.join()
[ https://issues.apache.org/jira/browse/KUDU-3019?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16995540#comment-16995540 ] Zoltán Borók-Nagy commented on KUDU-3019: - Thanks for looking at it, [~tlipcon]. It's version 1.6.0. The timeouts are the Impala defaults, which means 18ms for "defaultAdminOperationTimeoutMs" and "defaultOperationTimeoutMs" for the Kudu client. The query hangs for hours in that RPC call. I'm seeing it at a customer, not able to repro. > Kudu client hangs in Deffered.join() > > > Key: KUDU-3019 > URL: https://issues.apache.org/jira/browse/KUDU-3019 > Project: Kudu > Issue Type: Bug >Affects Versions: 1.6.0 >Reporter: Zoltán Borók-Nagy >Priority: Major > > In Impala we've seen the Kudu client hanging with the following stack trace: > {noformat} > Thread 53015: (state = BLOCKED) > - java.lang.Object.wait(long) @bci=0 (Compiled frame; information may be > imprecise) > - java.lang.Object.wait() @bci=2, line=502 (Compiled frame) > - com.stumbleupon.async.Deferred.doJoin(boolean, long) @bci=77, line=1122 > (Compiled frame) > - com.stumbleupon.async.Deferred.join() @bci=3, line=1006 (Compiled frame) > - > org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred) > @bci=1, line=340 (Compiled frame) > - org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, > line=212 (Compiled frame) > - > org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer) > @bci=32, line=115 (Compiled frame) > - > org.apache.impala.planner.SingleNodePlanner.createScanNode(org.apache.impala.analysis.TableRef, > org.apache.impala.analysis.AggregateInfo, > org.apache.impala.analysis.Analyzer) @bci=252, line=1312 (Compiled frame) > ...{noformat} > The client hangs in Deferred.join(): > [https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423] > To at least mitigate the problem, maybe Deferred.join(long timeout) could be > used instead. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (KUDU-3041) Kudu Java client shade is incomplete
Ismaël Mejía created KUDU-3041: -- Summary: Kudu Java client shade is incomplete Key: KUDU-3041 URL: https://issues.apache.org/jira/browse/KUDU-3041 Project: Kudu Issue Type: Bug Components: client Affects Versions: 1.11.1 Reporter: Ismaël Mejía While working on an update of the Kudu integration on Apache Beam BEAM-5086 We found this issue. We use [tool to test for linkage errors|https://github.com/GoogleCloudPlatform/cloud-opensource-java] and it reports the classes that are missing but required by other classes. This is the result for the kudu-client case: {code:java} Class javax.servlet.ServletOutputStream is not found; referenced by 1 class file org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet (kudu-client-1.11.1.jar) Class javax.servlet.http.HttpServlet is not found; referenced by 1 class file org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet (kudu-client-1.11.1.jar) Class javax.servlet.ServletException is not found; referenced by 1 class file org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet (kudu-client-1.11.1.jar) Class javax.servlet.ServletConfig is not found; referenced by 1 class file org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet (kudu-client-1.11.1.jar) Class javax.servlet.http.HttpServletRequest is not found; referenced by 1 class file org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet (kudu-client-1.11.1.jar) Class javax.servlet.http.HttpServletResponse is not found; referenced by 1 class file org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet (kudu-client-1.11.1.jar) Class org.jboss.marshalling.ByteInput is not found; referenced by 4 class files org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ChannelBufferByteInput (kudu-client-1.11.1.jar) org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.LimitingByteInput (kudu-client-1.11.1.jar) org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ChannelBufferByteInput (beam-vendor-grpc-1_21_0-0.1.jar) org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.LimitingByteInput (beam-vendor-grpc-1_21_0-0.1.jar) Class org.jboss.marshalling.ByteOutput is not found; referenced by 2 class files org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ChannelBufferByteOutput (kudu-client-1.11.1.jar) org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ChannelBufferByteOutput (beam-vendor-grpc-1_21_0-0.1.jar) Class org.jboss.marshalling.Unmarshaller is not found; referenced by 8 class files org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.CompatibleMarshallingDecoder (kudu-client-1.11.1.jar) org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ContextBoundUnmarshallerProvider (kudu-client-1.11.1.jar) org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.MarshallingDecoder (kudu-client-1.11.1.jar) org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ThreadLocalUnmarshallerProvider (kudu-client-1.11.1.jar) org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.MarshallingDecoder (beam-vendor-grpc-1_21_0-0.1.jar) org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.CompatibleMarshallingDecoder (beam-vendor-grpc-1_21_0-0.1.jar) org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ThreadLocalUnmarshallerProvider (beam-vendor-grpc-1_21_0-0.1.jar) org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ContextBoundUnmarshallerProvider (beam-vendor-grpc-1_21_0-0.1.jar) Class org.jboss.marshalling.Marshaller is not found; referenced by 6 class files org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.CompatibleMarshallingEncoder (kudu-client-1.11.1.jar) org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.MarshallingEncoder (kudu-client-1.11.1.jar) org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ThreadLocalMarshallerProvider (kudu-client-1.11.1.jar) org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.CompatibleMarshallingEncoder (beam-vendor-grpc-1_21_0-0.1.jar) org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.MarshallingEncoder (beam-vendor-grpc-1_21_0-0.1.jar) org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ThreadLocalMarshallerProvider (beam-vendor-grpc-1_21_0-0.1.jar) Class org.jboss.marshalling.MarshallerFactory is not found; referenced by 8 class files org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.DefaultMarshallerProvider (kudu-client-1.11.1.jar
[jira] [Commented] (KUDU-3041) Kudu Java client shade is incomplete
[ https://issues.apache.org/jira/browse/KUDU-3041?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014973#comment-17014973 ] Ismaël Mejía commented on KUDU-3041: Please ignore the `beam-vendor-grpc-1_21_0-0.1.jar` those are issues on our own shading. > Kudu Java client shade is incomplete > > > Key: KUDU-3041 > URL: https://issues.apache.org/jira/browse/KUDU-3041 > Project: Kudu > Issue Type: Bug > Components: client >Affects Versions: 1.11.1 >Reporter: Ismaël Mejía >Priority: Major > > While working on an update of the Kudu integration on Apache Beam BEAM-5086 > We found this issue. We use [tool to test for linkage > errors|https://github.com/GoogleCloudPlatform/cloud-opensource-java] and it > reports the classes that are missing but required by other classes. > This is the result for the kudu-client case: > {code:java} > Class javax.servlet.ServletOutputStream is not found; > referenced by 1 class file > > org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet > (kudu-client-1.11.1.jar) > Class javax.servlet.http.HttpServlet is not found; > referenced by 1 class file > > org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet > (kudu-client-1.11.1.jar) > Class javax.servlet.ServletException is not found; > referenced by 1 class file > > org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet > (kudu-client-1.11.1.jar) > Class javax.servlet.ServletConfig is not found; > referenced by 1 class file > > org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet > (kudu-client-1.11.1.jar) > Class javax.servlet.http.HttpServletRequest is not found; > referenced by 1 class file > > org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet > (kudu-client-1.11.1.jar) > Class javax.servlet.http.HttpServletResponse is not found; > referenced by 1 class file > > org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet > (kudu-client-1.11.1.jar) > Class org.jboss.marshalling.ByteInput is not found; > referenced by 4 class files > > org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ChannelBufferByteInput > (kudu-client-1.11.1.jar) > > org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.LimitingByteInput > (kudu-client-1.11.1.jar) > > org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ChannelBufferByteInput > (beam-vendor-grpc-1_21_0-0.1.jar) > > org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.LimitingByteInput > (beam-vendor-grpc-1_21_0-0.1.jar) > Class org.jboss.marshalling.ByteOutput is not found; > referenced by 2 class files > > org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ChannelBufferByteOutput > (kudu-client-1.11.1.jar) > > org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ChannelBufferByteOutput > (beam-vendor-grpc-1_21_0-0.1.jar) > Class org.jboss.marshalling.Unmarshaller is not found; > referenced by 8 class files > > org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.CompatibleMarshallingDecoder > (kudu-client-1.11.1.jar) > > org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ContextBoundUnmarshallerProvider > (kudu-client-1.11.1.jar) > > org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.MarshallingDecoder > (kudu-client-1.11.1.jar) > > org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ThreadLocalUnmarshallerProvider > (kudu-client-1.11.1.jar) > > org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.MarshallingDecoder > (beam-vendor-grpc-1_21_0-0.1.jar) > > org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.CompatibleMarshallingDecoder > (beam-vendor-grpc-1_21_0-0.1.jar) > > org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ThreadLocalUnmarshallerProvider > (beam-vendor-grpc-1_21_0-0.1.jar) > > org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ContextBoundUnmarshallerProvider > (beam-vendor-grpc-1_21_0-0.1.jar) > Class org.jboss.marshalling.Marshaller is not found; > referenced by 6 class files > > org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.CompatibleMarshallingEncoder > (kudu-client-1.11.1.jar) > > org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.Mars
[jira] [Created] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20)
João Boto created KUDU-3174: --- Summary: KuduTestHarness not working on focal (ubuntu 20) Key: KUDU-3174 URL: https://issues.apache.org/jira/browse/KUDU-3174 Project: Kudu Issue Type: Bug Affects Versions: 1.11.1, 1.12.0 Reporter: João Boto when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting this error: {code:java} // code placeholder java.io.EOFException at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) at org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) at org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) at org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) at org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) at org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} tested with 1.11 and 1.12 on local and on travis (with focal dist) -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20)
[ https://issues.apache.org/jira/browse/KUDU-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] João Boto updated KUDU-3174: Description: when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting this error: {code:java} // code placeholder java.io.EOFException at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) at org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) at org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) at org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) at org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) at org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} tested with 1.11 and 1.12 on local and on travis (with focal dist) both with jdk8 and jdk11 was: when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting this error: {code:java} // code placeholder java.io.EOFException at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) at org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) at org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) at org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) at org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) at org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} tested with 1.11 and 1.12 on local and on travis (with focal dist) > KuduTestHarness not working on focal (ubuntu 20) > > > Key: KUDU-3174 > URL: https://issues.apache.org/jira/browse/KUDU-3174 > Project: Kudu > Issue Type: Bug >Affects Versions: 1.12.0, 1.11.1 >Reporter: João Boto >Priority: Major > > when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im > getting this error: > {code:java} > // code placeholder > java.io.EOFException > at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) > at > org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) > at > org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) > at > org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) > at > org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) > at > org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} > tested with 1.11 and 1.12 on local and on travis (with focal dist) > both with jdk8 and jdk11 > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Closed] (KUDU-3175) .
[ https://issues.apache.org/jira/browse/KUDU-3175?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] João Boto closed KUDU-3175. --- Resolution: Invalid > . > - > > Key: KUDU-3175 > URL: https://issues.apache.org/jira/browse/KUDU-3175 > Project: Kudu > Issue Type: Improvement >Reporter: João Boto >Priority: Major > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (KUDU-3175) upgrade flink to 1.11.1
João Boto created KUDU-3175: --- Summary: upgrade flink to 1.11.1 Key: KUDU-3175 URL: https://issues.apache.org/jira/browse/KUDU-3175 Project: Kudu Issue Type: Improvement Reporter: João Boto -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (KUDU-3175) .
[ https://issues.apache.org/jira/browse/KUDU-3175?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] João Boto updated KUDU-3175: Summary: . (was: upgrade flink to 1.11.1) > . > - > > Key: KUDU-3175 > URL: https://issues.apache.org/jira/browse/KUDU-3175 > Project: Kudu > Issue Type: Improvement >Reporter: João Boto >Priority: Major > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20)
[ https://issues.apache.org/jira/browse/KUDU-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] João Boto updated KUDU-3174: Description: when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting this error: {code:java} java.io.EOFException at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) at org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) at org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) at org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) at org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) at org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} tested with 1.11 and 1.12 on local and on travis (with focal dist) both with jdk8 and jdk11 was: when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting this error: {code:java} // code placeholder java.io.EOFException at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) at org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) at org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) at org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) at org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) at org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} tested with 1.11 and 1.12 on local and on travis (with focal dist) both with jdk8 and jdk11 > KuduTestHarness not working on focal (ubuntu 20) > > > Key: KUDU-3174 > URL: https://issues.apache.org/jira/browse/KUDU-3174 > Project: Kudu > Issue Type: Bug >Affects Versions: 1.12.0, 1.11.1 >Reporter: João Boto >Priority: Major > > when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im > getting this error: > {code:java} > java.io.EOFException > at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) > at > org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) > at > org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) > at > org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) > at > org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) > at > org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} > tested with 1.11 and 1.12 on local and on travis (with focal dist) > both with jdk8 and jdk11 > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20 LTS)
[ https://issues.apache.org/jira/browse/KUDU-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] João Boto updated KUDU-3174: Summary: KuduTestHarness not working on focal (ubuntu 20 LTS) (was: KuduTestHarness not working on focal (ubuntu 20)) > KuduTestHarness not working on focal (ubuntu 20 LTS) > > > Key: KUDU-3174 > URL: https://issues.apache.org/jira/browse/KUDU-3174 > Project: Kudu > Issue Type: Bug >Affects Versions: 1.12.0, 1.11.1 >Reporter: João Boto >Priority: Major > > when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im > getting this error: > {code:java} > java.io.EOFException > at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) > at > org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) > at > org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) > at > org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) > at > org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) > at > org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} > tested with 1.11 and 1.12 on local and on travis (with focal dist) > both with jdk8 and jdk11 > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20 LTS)
[ https://issues.apache.org/jira/browse/KUDU-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17185755#comment-17185755 ] João Boto commented on KUDU-3174: - [~adar] maybe could you check > KuduTestHarness not working on focal (ubuntu 20 LTS) > > > Key: KUDU-3174 > URL: https://issues.apache.org/jira/browse/KUDU-3174 > Project: Kudu > Issue Type: Bug >Affects Versions: 1.12.0, 1.11.1 >Reporter: João Boto >Priority: Major > > when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im > getting this error: > {code:java} > java.io.EOFException > at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) > at > org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) > at > org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) > at > org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) > at > org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) > at > org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} > tested with 1.11 and 1.12 on local and on travis (with focal dist) > both with jdk8 and jdk11 > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20 LTS)
[ https://issues.apache.org/jira/browse/KUDU-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] João Boto updated KUDU-3174: Description: when executing tests with KuduTestHarness on ubuntu 20 LTS (focal fossa) Im getting this error: {code:java} java.io.EOFException at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) at org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) at org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) at org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) at org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) at org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} tested with 1.11 and 1.12 on local and on travis (with focal dist) both with jdk8 and jdk11 was: when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting this error: {code:java} java.io.EOFException at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) at org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) at org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) at org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) at org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) at org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} tested with 1.11 and 1.12 on local and on travis (with focal dist) both with jdk8 and jdk11 > KuduTestHarness not working on focal (ubuntu 20 LTS) > > > Key: KUDU-3174 > URL: https://issues.apache.org/jira/browse/KUDU-3174 > Project: Kudu > Issue Type: Bug >Affects Versions: 1.12.0, 1.11.1 >Reporter: João Boto >Priority: Major > > when executing tests with KuduTestHarness on ubuntu 20 LTS (focal fossa) Im > getting this error: > {code:java} > java.io.EOFException > at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397) > at > org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162) > at > org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235) > at > org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72) > at > org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697) > at > org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code} > tested with 1.11 and 1.12 on local and on travis (with focal dist) > both with jdk8 and jdk11 > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Assigned] (KUDU-1698) Kudu C++ client: add a new unit test to make sure default_rpc_timeout and session timeout are separate entities
[ https://issues.apache.org/jira/browse/KUDU-1698?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-1698: Assignee: Bakai Ádám > Kudu C++ client: add a new unit test to make sure default_rpc_timeout and > session timeout are separate entities > --- > > Key: KUDU-1698 > URL: https://issues.apache.org/jira/browse/KUDU-1698 > Project: Kudu > Issue Type: Task > Components: client, test >Reporter: Alexey Serbin >Assignee: Bakai Ádám >Priority: Minor > Labels: newbie > > We need a new unit test that makes sure there is a difference between > top-level operation timeout and per-call RPC timeout in Kudu C++ client > library. Prior to change introduced in > 5195ce573850653e0e53094cdd35a1da93d33444 it was the same (which was a bug). > The test should: > * set per-call RPC timeout when creating KuduClient object > * set KuduSession::SetTimeoutMillis() for the target session: the value > should be 2 times of per-call RPC timeout or such. > * create a tablet with replication factor of 2 at least. > * find current tablet replica leader and pause it (send SIGSTOP) > * make a write into the table > * make sure the write operation was successful > Prior to change introduced in 5195ce573850653e0e53094cdd35a1da93d33444 such a > test would fail because the C++ client used the full operation deadline on > every RPC call. > I.e., it would wait till the call to current leader times out, and that would > consume time budget of the whole operation. Once RPC timeout is less thatn > the timeout for the whole write operation, the call to the frozen tablet > server should timeout, and the client should re-discover a new tablet > replicate leader and complete the write operation successfully. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3414) Client tries with the same address over and over in case of negotiation timeout
Bakai Ádám created KUDU-3414: Summary: Client tries with the same address over and over in case of negotiation timeout Key: KUDU-3414 URL: https://issues.apache.org/jira/browse/KUDU-3414 Project: Kudu Issue Type: Bug Components: client Reporter: Bakai Ádám Reproduction: 1. Start new master and tablet servers. 2. Create a client and create a table and do a succesfull write operation with it. 3. Pause the leader tablet server. 4. Now if you attempt to write with the same client, it will timeout, and it will try negotiation with the same server until the session times out. (If a new client is created, then the write operation is succesful.) KUDU-1580 works on a similar issue, the error is retriable for now, but [https://gerrit.cloudera.org/#/c/6926/] states that "the Kudu C++ client retries an RPC with other tablet replica if the connection negotiation with current replica timed out." which doesn't happen right now. (Didn't check if the current behaviour is a regression or not.) Here is a reproduction test code into client-test.cc: {code:java} TEST_F(ClientTest, TestDefaultRPCTimeoutSessionTimeoutDifferent) { cluster_->Shutdown(); FLAGS_raft_heartbeat_interval_ms = 50; FLAGS_leader_failure_exp_backoff_max_delta_ms = 1000; const string table_name = "TestTimeout"; unique_ptr ext_cluster_; ext_cluster_.reset(new ExternalMiniCluster( ExternalMiniClusterOptions())); ext_cluster_->Start(); ext_cluster_->AddMaster(); ext_cluster_->AddTabletServer(); ext_cluster_->AddTabletServer(); ext_cluster_->AddTabletServer(); ext_cluster_->AddTabletServer(); ext_cluster_->AddTabletServer(); ext_cluster_->WaitForTabletServerCount(5,MonoDelta::FromSeconds(1)); client::sp::shared_ptr client; client_= nullptr; KuduClientBuilder builder; builder.num_reactors(1); builder.connection_negotiation_timeout(MonoDelta::FromMilliseconds(1500)); builder.default_rpc_timeout(MonoDelta::FromMilliseconds(500)); ext_cluster_->CreateClient(&builder,&client); unique_ptr table_creator(client->NewTableCreator()); table_creator->table_name(table_name) .schema(&schema_) .num_replicas(5) .set_range_partition_columns({ "key" }) .timeout(MonoDelta::FromSeconds(60)) .Create(); shared_ptr table; client->OpenTable(table_name,&table); unique_ptr insert; shared_ptr session = client->NewSession(); session->SetTimeoutMillis(6); insert = BuildTestInsert(table.get(), 1); ASSERT_OK(session->Apply(insert.release())); session->Flush(); KuduScanTokenBuilder builder2(table.get()); vector tokens; ASSERT_OK(builder2.Build(&tokens)); const KuduTablet& tablet = tokens[0]->tablet(); for (const auto& a: tablet.replicas()) { if (a->is_leader()) { ext_cluster_->tablet_server_by_uuid(a->ts().uuid())->Pause(); //ext_cluster_->tablet_server_by_uuid(a->ts().uuid())->Shutdown(); break; } } LOG(INFO)<<"paused"; //ext_cluster_->CreateClient(&builder,&client); session = client->NewSession(); session->SetTimeoutMillis(6); insert = BuildTestInsert(table.get(), 2); ASSERT_OK(session->Apply(insert.release())); } {code} (Don't forget to add the declarations and includes to the start of the file:) {code:java} #include "kudu/mini-cluster/external_mini_cluster.h" using kudu::cluster::ExternalMiniCluster; using kudu::cluster::ExternalMiniClusterOptions; {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3414) Client tries with the same address over and over in case of negotiation timeout
[ https://issues.apache.org/jira/browse/KUDU-3414?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3414: - Description: Reproduction: 1. Start new master and tablet servers. 2. Create a client and create a table and do a succesfull write operation with it. 3. Pause the leader tablet server. 4. Now if you attempt to write with the same client, it will timeout, and it will try negotiation with the same server until the session times out. (If a new client is created, then the write operation is succesful.) KUDU-1580 works on a similar issue, the error is retriable for now, but [https://gerrit.cloudera.org/#/c/6926/] states that "the Kudu C++ client retries an RPC with other tablet replica if the connection negotiation with current replica timed out." which doesn't happen right now. (Didn't check if the current behaviour is a regression or not.) Here is a reproduction test code into client-test.cc: {code:java} TEST_F(ClientTest, TestDefaultRPCTimeoutSessionTimeoutDifferent) { cluster_->Shutdown(); FLAGS_raft_heartbeat_interval_ms = 50; FLAGS_leader_failure_exp_backoff_max_delta_ms = 1000; const string table_name = "TestTimeout"; unique_ptr ext_cluster_; ext_cluster_.reset(new ExternalMiniCluster( ExternalMiniClusterOptions())); ext_cluster_->Start(); ext_cluster_->AddMaster(); ext_cluster_->AddTabletServer(); ext_cluster_->AddTabletServer(); ext_cluster_->AddTabletServer(); ext_cluster_->AddTabletServer(); ext_cluster_->AddTabletServer(); ext_cluster_->WaitForTabletServerCount(5,MonoDelta::FromSeconds(1)); client::sp::shared_ptr client; client_= nullptr; KuduClientBuilder builder; builder.num_reactors(1); builder.connection_negotiation_timeout(MonoDelta::FromMilliseconds(1500)); builder.default_rpc_timeout(MonoDelta::FromMilliseconds(500)); ext_cluster_->CreateClient(&builder,&client); unique_ptr table_creator(client->NewTableCreator()); table_creator->table_name(table_name) .schema(&schema_) .num_replicas(5) .set_range_partition_columns({ "key" }) .timeout(MonoDelta::FromSeconds(60)) .Create(); shared_ptr table; client->OpenTable(table_name,&table); unique_ptr insert; shared_ptr session = client->NewSession(); session->SetTimeoutMillis(6); insert = BuildTestInsert(table.get(), 1); ASSERT_OK(session->Apply(insert.release())); session->Flush(); KuduScanTokenBuilder builder2(table.get()); vector tokens; ASSERT_OK(builder2.Build(&tokens)); const KuduTablet& tablet = tokens[0]->tablet(); for (const auto& a: tablet.replicas()) { if (a->is_leader()) { ext_cluster_->tablet_server_by_uuid(a->ts().uuid())->Pause(); //ext_cluster_->tablet_server_by_uuid(a->ts().uuid())->Shutdown(); break; } } LOG(INFO)<<"paused"; //ext_cluster_->CreateClient(&builder,&client); session = client->NewSession(); session->SetTimeoutMillis(6); insert = BuildTestInsert(table.get(), 2); ASSERT_OK(session->Apply(insert.release())); } {code} (Don't forget to add the declarations and includes to the start of the [file:|file:///]) {code:java} #include "kudu/mini-cluster/external_mini_cluster.h" using kudu::cluster::ExternalMiniCluster; using kudu::cluster::ExternalMiniClusterOptions; {code} The logic about deciding when to try different tablet is in: RetriableRpcStatus WriteRpc::AnalyzeResponse(const Status& rpc_cb_status) was: Reproduction: 1. Start new master and tablet servers. 2. Create a client and create a table and do a succesfull write operation with it. 3. Pause the leader tablet server. 4. Now if you attempt to write with the same client, it will timeout, and it will try negotiation with the same server until the session times out. (If a new client is created, then the write operation is succesful.) KUDU-1580 works on a similar issue, the error is retriable for now, but [https://gerrit.cloudera.org/#/c/6926/] states that "the Kudu C++ client retries an RPC with other tablet replica if the connection negotiation with current replica timed out." which doesn't happen right now. (Didn't check if the current behaviour is a regression or not.) Here is a reproduction test code into client-test.cc: {code:java} TEST_F(ClientTest, TestDefaultRPCTimeoutSessionTimeoutDifferent) { cluster_->Shutdown(); FLAGS_raft_heartbeat_interval_ms = 50; FLAGS_leader_failure_exp_backoff_max_delta_ms = 1000; const string table_name = "TestTimeout"; unique_ptr ext_cluster_;
[jira] [Commented] (KUDU-1698) Kudu C++ client: add a new unit test to make sure default_rpc_timeout and session timeout are separate entities
[ https://issues.apache.org/jira/browse/KUDU-1698?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17625692#comment-17625692 ] Bakai Ádám commented on KUDU-1698: -- I tried to recreate the exact steps in a test, but it failed, because the client didn't rediscover, but rather tried over and over again until session timeout. I talked with [~aserbin] and we came to the conclusion to create a new issue for the not rediscovering behaviour, and test the session and rpc timeout in an ohter way. The new issue is: KUDU-3414 . The new idea to test the separate entity property: * Make the tablet lookup artificially slow by adding latency. * See that the rpc is timing out but retries. * Remove the artifical delay * Check that the operation was succesful in the end, and tablet look up happened twice. > Kudu C++ client: add a new unit test to make sure default_rpc_timeout and > session timeout are separate entities > --- > > Key: KUDU-1698 > URL: https://issues.apache.org/jira/browse/KUDU-1698 > Project: Kudu > Issue Type: Task > Components: client, test >Reporter: Alexey Serbin >Assignee: Bakai Ádám >Priority: Minor > Labels: newbie > > We need a new unit test that makes sure there is a difference between > top-level operation timeout and per-call RPC timeout in Kudu C++ client > library. Prior to change introduced in > 5195ce573850653e0e53094cdd35a1da93d33444 it was the same (which was a bug). > The test should: > * set per-call RPC timeout when creating KuduClient object > * set KuduSession::SetTimeoutMillis() for the target session: the value > should be 2 times of per-call RPC timeout or such. > * create a tablet with replication factor of 2 at least. > * find current tablet replica leader and pause it (send SIGSTOP) > * make a write into the table > * make sure the write operation was successful > Prior to change introduced in 5195ce573850653e0e53094cdd35a1da93d33444 such a > test would fail because the C++ client used the full operation deadline on > every RPC call. > I.e., it would wait till the call to current leader times out, and that would > consume time budget of the whole operation. Once RPC timeout is less thatn > the timeout for the whole write operation, the call to the frozen tablet > server should timeout, and the client should re-discover a new tablet > replicate leader and complete the write operation successfully. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (KUDU-3387) Enable TCP keepalive for connections created by Java client
[ https://issues.apache.org/jira/browse/KUDU-3387?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17628948#comment-17628948 ] Bakai Ádám commented on KUDU-3387: -- In the commit 57dda5d4868d29f68de4aa0ac516ca390333e6be , the possibility was explored, and the timeout is enabled, but the timeout is not set: AsyncKuduClient.java:~3030 {code:java} + // Unfortunately there is no way to override the keep-alive timeout in + // Java since the JRE doesn't expose any way to call setsockopt() with + // TCP_KEEPIDLE. And of course the default timeout is >2h. Sigh. + b.option(ChannelOption.SO_KEEPALIVE, true); {code} As far as I understand netty doesn't expose the Socket itself, but since then it is possible to change the timeout interval on some platform with ExtendedSocketOptions (https://docs.oracle.com/en/java/javase/11/docs/api/jdk.net/jdk/net/ExtendedSocketOptions.html). Here is an example how to set it: https://stackoverflow.com/questions/22472844/how-to-set-socket-option-tcp-keepcnt-tcp-keepidle-tcp-keepintvl-in-java-or-n So in theory It may be possible to set options on the SocketChannel (on Mac and Linux, source: https://bugs.openjdk.org/browse/JDK-8194298) , but not with io.netty.channel.socket.nio.NioSocketChannel . These are the options that can be configured: https://javadoc.io/static/io.netty/netty-transport/5.0.0.Alpha2/io/netty/channel/socket/SocketChannelConfig.html . This is the highest version of netty I could find at the date of the comment. So, in short: the TCP keepalive is set to true already, but the timeout interval is not explicitly set, and it can be very high (2 hours). Now Java support changing the keepalive time on Mac and Linux(keep in mind the current solution is Linux-only (Socket::SetTcpKeepAlive)), but netty doesn't support it, so it can not be changed. > Enable TCP keepalive for connections created by Java client > --- > > Key: KUDU-3387 > URL: https://issues.apache.org/jira/browse/KUDU-3387 > Project: Kudu > Issue Type: Improvement > Components: client, java >Reporter: Alexey Serbin >Priority: Major > Labels: newbie > > https://github.com/apache/kudu/commit/89c02fded7595b4712b465bfb939e4f3035b2e75 > enabled TCP keepalive for outbound connections; and the C++ client since > then is able to detect a dropped off connection when server didn't send a TCP > RST. > It's would be great to enable TCP keepalive for connections created by Java > client as well. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (KUDU-3300) Include the full path of the container in the error message
[ https://issues.apache.org/jira/browse/KUDU-3300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3300: Assignee: Bakai Ádám > Include the full path of the container in the error message > --- > > Key: KUDU-3300 > URL: https://issues.apache.org/jira/browse/KUDU-3300 > Project: Kudu > Issue Type: Improvement > Components: cfile >Reporter: Abhishek >Assignee: Bakai Ádám >Priority: Minor > Labels: newbie > > If there are multiple data directories configured, having the linux path to > the full container file will help to locate the file without having to search > for the file > Check failed: _s.ok() Bad status: Corruption: Failed to load FS layout: Could > not open container 26f5cbd97dfe4cb98f49bb0a6a494e8f: Invalid magic number: > Expected: kuducntr, found: \000\000\020\001\030▒▒▒ -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-2981) Push predicate evaluation into more CFile decoders
[ https://issues.apache.org/jira/browse/KUDU-2981?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-2981: - Labels: performance (was: newbie performance) > Push predicate evaluation into more CFile decoders > -- > > Key: KUDU-2981 > URL: https://issues.apache.org/jira/browse/KUDU-2981 > Project: Kudu > Issue Type: Improvement > Components: cfile, perf >Reporter: Bankim Bhavsar >Priority: Major > Labels: performance > > Commit c0f3727 added an optimization to push predicate evaluation into the > CFile decoders without fully materializing the contents of each cblock. It > did this with dictionary-encoded blocks, but the optimization can be applied > to any other encoding types too. > KUDU-736 also notes that we may be able to apply some predicates on > bitshuffled data. > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3421) Make commane line error message more useful
Bakai Ádám created KUDU-3421: Summary: Make commane line error message more useful Key: KUDU-3421 URL: https://issues.apache.org/jira/browse/KUDU-3421 Project: Kudu Issue Type: Improvement Components: CLI Reporter: Bakai Ádám If the user runs this command for example: {code:java} /data/9/abakai/kudu/build/debug% /data/9/abakai/kudu/build/debug/bin/kudu-tserver --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 --tserver_master_addrs=127.0.0.1:18764 --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www --memory_limit_hard_bytes=72050161527 -block_manager log {code} it works fine. But if the user adds "-logtostderr true", it will throw an error: {code:java} [abakai@va1022]/data/9/abakai/kudu/build/debug% /data/9/abakai/kudu/build/debug/bin/kudu-tserver --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 --tserver_master_addrs=127.0.0.1:18764 --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www --memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code} Which doesn't really help to identify which command line option is incorrect (and maybe how is it incorrect) -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3421) Make commane line error message more useful
[ https://issues.apache.org/jira/browse/KUDU-3421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3421: - Description: If the user runs this command for example: {code:java} /data/9/abakai/kudu/build/debug% /data/9/abakai/kudu/build/debug/bin/kudu-tserver --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 --tserver_master_addrs=127.0.0.1:18764 --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www --memory_limit_hard_bytes=72050161527 -block_manager log {code} it works fine. But if the user adds "-logtostderr true", it will throw an error: {code:java} [abakai@va1022]/data/9/abakai/kudu/build/debug% /data/9/abakai/kudu/build/debug/bin/kudu-tserver --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 --tserver_master_addrs=127.0.0.1:18764 --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www --memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code} Which doesn't really help to identify which command line option is incorrect (and maybe how is it incorrect) (By the way the problem with the second command is that logtostderr doesn't need true or false option, it's existence indicates that logtostderr should be switched to true) was: If the user runs this command for example: {code:java} /data/9/abakai/kudu/build/debug% /data/9/abakai/kudu/build/debug/bin/kudu-tserver --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 --tserver_master_addrs=127.0.0.1:18764 --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www --memory_limit_hard_bytes=72050161527 -block_manager log {code} it works fine. But if the user adds "-logtostderr true", it will throw an error: {code:java} [abakai@va1022]/data/9/abakai/kudu/build/debug% /data/9/abakai/kudu/build/debug/bin/kudu-tserver --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 --tserver_master_addrs=127.0.0.1:18764 --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www --memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code} Which doesn't really help to identify which command line option is incorrect (and maybe how is it incorrect) > Make commane line error message more useful > --- > > Key: KUDU-3421 > URL: https://issues.apache.org/jira/browse/KUDU-3421 > Project: Kudu > Issue Type: Improvement > Components: CLI >Reporter: Bakai Ádám >Priority: Major > > If the user runs this command for example: > {code:java} > /data/9/abakai/kudu/build/debug% > /data/9/abakai/kudu/build/debug/bin/kudu-tserver > --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data > --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal > --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log > --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync > --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 > --tserver_master_addrs=127.0.0.1:18764 > --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www > --memory_limit_hard_bytes=72050161527 -block_manager log {code} > it works fine. But if the user adds "-logtostderr true", it will throw an > error: > {code:java} > [abakai@va1022]/data/9/abakai/kudu/build/debug% > /data/9/abakai/kudu/build/debug/bin/kudu-tserver > --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data > --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal > --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log > --rpc_bind_addresses=127.0.0.1:19874 --t
[jira] [Commented] (KUDU-3210) Support FIPS approved mode
[ https://issues.apache.org/jira/browse/KUDU-3210?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17631907#comment-17631907 ] Bakai Ádám commented on KUDU-3210: -- On OpenSSL 3, FIPS_mode is removed ([https://wiki.openssl.org/index.php/OpenSSL_3.0#Upgrading_from_the_OpenSSL_2.0_FIPS_Object_Module)] and we use the system openssl, it may be a good idea to create a solution that works on both openssl versions. > Support FIPS approved mode > -- > > Key: KUDU-3210 > URL: https://issues.apache.org/jira/browse/KUDU-3210 > Project: Kudu > Issue Type: Improvement >Reporter: Attila Bukor >Assignee: Attila Bukor >Priority: Major > Fix For: 1.14.0 > > > FIPS 140-2 is a standard used to approve cryptographic modules. Some versions > of OpenSSL support a "FIPS mode" where only approved algorithms and key sizes > are enabled. Kudu should be able to run when FIPS mode is enabled and should > provide a way for admins to require that FIPS mode is enabled. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3421) Make commane line error message more useful
[ https://issues.apache.org/jira/browse/KUDU-3421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3421: - Labels: newbie (was: ) > Make commane line error message more useful > --- > > Key: KUDU-3421 > URL: https://issues.apache.org/jira/browse/KUDU-3421 > Project: Kudu > Issue Type: Improvement > Components: CLI >Reporter: Bakai Ádám >Priority: Major > Labels: newbie > > If the user runs this command for example: > {code:java} > /data/9/abakai/kudu/build/debug% > /data/9/abakai/kudu/build/debug/bin/kudu-tserver > --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data > --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal > --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log > --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync > --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 > --tserver_master_addrs=127.0.0.1:18764 > --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www > --memory_limit_hard_bytes=72050161527 -block_manager log {code} > it works fine. But if the user adds "-logtostderr true", it will throw an > error: > {code:java} > [abakai@va1022]/data/9/abakai/kudu/build/debug% > /data/9/abakai/kudu/build/debug/bin/kudu-tserver > --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data > --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal > --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log > --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync > --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 > --tserver_master_addrs=127.0.0.1:18764 > --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www > --memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true > usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code} > Which doesn't really help to identify which command line option is incorrect > (and maybe how is it incorrect) > (By the way the problem with the second command is that logtostderr doesn't > need true or false option, it's existence indicates that logtostderr should > be switched to true) -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3421) Make command line error message more useful
[ https://issues.apache.org/jira/browse/KUDU-3421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3421: - Summary: Make command line error message more useful (was: Make commane line error message more useful) > Make command line error message more useful > --- > > Key: KUDU-3421 > URL: https://issues.apache.org/jira/browse/KUDU-3421 > Project: Kudu > Issue Type: Improvement > Components: CLI >Reporter: Bakai Ádám >Priority: Major > Labels: newbie > > If the user runs this command for example: > {code:java} > /data/9/abakai/kudu/build/debug% > /data/9/abakai/kudu/build/debug/bin/kudu-tserver > --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data > --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal > --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log > --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync > --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 > --tserver_master_addrs=127.0.0.1:18764 > --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www > --memory_limit_hard_bytes=72050161527 -block_manager log {code} > it works fine. But if the user adds "-logtostderr true", it will throw an > error: > {code:java} > [abakai@va1022]/data/9/abakai/kudu/build/debug% > /data/9/abakai/kudu/build/debug/bin/kudu-tserver > --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data > --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal > --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log > --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync > --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 > --tserver_master_addrs=127.0.0.1:18764 > --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www > --memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true > usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code} > Which doesn't really help to identify which command line option is incorrect > (and maybe how is it incorrect) > (By the way the problem with the second command is that logtostderr doesn't > need true or false option, it's existence indicates that logtostderr should > be switched to true) -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Closed] (KUDU-3421) Make command line error message more useful
[ https://issues.apache.org/jira/browse/KUDU-3421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám closed KUDU-3421. > Make command line error message more useful > --- > > Key: KUDU-3421 > URL: https://issues.apache.org/jira/browse/KUDU-3421 > Project: Kudu > Issue Type: Improvement > Components: CLI >Reporter: Bakai Ádám >Priority: Major > Labels: newbie > Fix For: 1.17.0 > > > If the user runs this command for example: > {code:java} > /data/9/abakai/kudu/build/debug% > /data/9/abakai/kudu/build/debug/bin/kudu-tserver > --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data > --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal > --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log > --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync > --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 > --tserver_master_addrs=127.0.0.1:18764 > --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www > --memory_limit_hard_bytes=72050161527 -block_manager log {code} > it works fine. But if the user adds "-logtostderr true", it will throw an > error: > {code:java} > [abakai@va1022]/data/9/abakai/kudu/build/debug% > /data/9/abakai/kudu/build/debug/bin/kudu-tserver > --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data > --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal > --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log > --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync > --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 > --tserver_master_addrs=127.0.0.1:18764 > --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www > --memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true > usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code} > Which doesn't really help to identify which command line option is incorrect > (and maybe how is it incorrect) > (By the way the problem with the second command is that logtostderr doesn't > need true or false option, it's existence indicates that logtostderr should > be switched to true) -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (KUDU-3300) Include the full path of the container in the error message
[ https://issues.apache.org/jira/browse/KUDU-3300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3300: Assignee: Abhishek (was: Bakai Ádám) > Include the full path of the container in the error message > --- > > Key: KUDU-3300 > URL: https://issues.apache.org/jira/browse/KUDU-3300 > Project: Kudu > Issue Type: Improvement > Components: cfile >Reporter: Abhishek >Assignee: Abhishek >Priority: Minor > Labels: newbie > Fix For: 1.17.0 > > > If there are multiple data directories configured, having the linux path to > the full container file will help to locate the file without having to search > for the file > Check failed: _s.ok() Bad status: Corruption: Failed to load FS layout: Could > not open container 26f5cbd97dfe4cb98f49bb0a6a494e8f: Invalid magic number: > Expected: kuducntr, found: \000\000\020\001\030▒▒▒ -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (KUDU-3300) Include the full path of the container in the error message
[ https://issues.apache.org/jira/browse/KUDU-3300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám resolved KUDU-3300. -- Fix Version/s: 1.17.0 Resolution: Fixed > Include the full path of the container in the error message > --- > > Key: KUDU-3300 > URL: https://issues.apache.org/jira/browse/KUDU-3300 > Project: Kudu > Issue Type: Improvement > Components: cfile >Reporter: Abhishek >Assignee: Bakai Ádám >Priority: Minor > Labels: newbie > Fix For: 1.17.0 > > > If there are multiple data directories configured, having the linux path to > the full container file will help to locate the file without having to search > for the file > Check failed: _s.ok() Bad status: Corruption: Failed to load FS layout: Could > not open container 26f5cbd97dfe4cb98f49bb0a6a494e8f: Invalid magic number: > Expected: kuducntr, found: \000\000\020\001\030▒▒▒ -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (KUDU-3292) Show non-default flags on varz Web UI
[ https://issues.apache.org/jira/browse/KUDU-3292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3292: Assignee: Bakai Ádám (was: Zoltan Chovan) > Show non-default flags on varz Web UI > - > > Key: KUDU-3292 > URL: https://issues.apache.org/jira/browse/KUDU-3292 > Project: Kudu > Issue Type: Improvement > Components: ui >Reporter: Grant Henke >Assignee: Bakai Ádám >Priority: Minor > Labels: beginner, newbie, newbie++, trivial > > Currently each Kudu server has a /varz webpage (the Flags tab) showing all of > the flags set on the server. It would be a nice usability change to include a > seperate section showing only the non-default flags. This should be super > straigtforward given we have the ability to get all the non-default flags via > GetNonDefaultFlags or GetNonDefaultFlagsMap in flags.cc. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (KUDU-1698) Kudu C++ client: add a new unit test to make sure default_rpc_timeout and session timeout are separate entities
[ https://issues.apache.org/jira/browse/KUDU-1698?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám resolved KUDU-1698. -- Fix Version/s: n/a Resolution: Fixed > Kudu C++ client: add a new unit test to make sure default_rpc_timeout and > session timeout are separate entities > --- > > Key: KUDU-1698 > URL: https://issues.apache.org/jira/browse/KUDU-1698 > Project: Kudu > Issue Type: Task > Components: client, test >Reporter: Alexey Serbin >Assignee: Bakai Ádám >Priority: Minor > Labels: newbie > Fix For: n/a > > > We need a new unit test that makes sure there is a difference between > top-level operation timeout and per-call RPC timeout in Kudu C++ client > library. Prior to change introduced in > 5195ce573850653e0e53094cdd35a1da93d33444 it was the same (which was a bug). > The test should: > * set per-call RPC timeout when creating KuduClient object > * set KuduSession::SetTimeoutMillis() for the target session: the value > should be 2 times of per-call RPC timeout or such. > * create a tablet with replication factor of 2 at least. > * find current tablet replica leader and pause it (send SIGSTOP) > * make a write into the table > * make sure the write operation was successful > Prior to change introduced in 5195ce573850653e0e53094cdd35a1da93d33444 such a > test would fail because the C++ client used the full operation deadline on > every RPC call. > I.e., it would wait till the call to current leader times out, and that would > consume time budget of the whole operation. Once RPC timeout is less thatn > the timeout for the whole write operation, the call to the frozen tablet > server should timeout, and the client should re-discover a new tablet > replicate leader and complete the write operation successfully. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (KUDU-3292) Show non-default flags on varz Web UI
[ https://issues.apache.org/jira/browse/KUDU-3292?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17674833#comment-17674833 ] Bakai Ádám commented on KUDU-3292: -- [~laiyingchun] I'm not familiar with impala, and you are right, it looks more informative, but refactoring it is a way bigger task then what this issue was about originally. > Show non-default flags on varz Web UI > - > > Key: KUDU-3292 > URL: https://issues.apache.org/jira/browse/KUDU-3292 > Project: Kudu > Issue Type: Improvement > Components: ui >Reporter: Grant Henke >Assignee: Bakai Ádám >Priority: Minor > Labels: beginner, newbie, newbie++, trivial > Attachments: image-2023-01-10-11-57-13-209.png > > > Currently each Kudu server has a /varz webpage (the Flags tab) showing all of > the flags set on the server. It would be a nice usability change to include a > seperate section showing only the non-default flags. This should be super > straigtforward given we have the ability to get all the non-default flags via > GetNonDefaultFlags or GetNonDefaultFlagsMap in flags.cc. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (KUDU-3292) Show non-default flags on varz Web UI
[ https://issues.apache.org/jira/browse/KUDU-3292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám resolved KUDU-3292. -- Fix Version/s: n/a Resolution: Fixed > Show non-default flags on varz Web UI > - > > Key: KUDU-3292 > URL: https://issues.apache.org/jira/browse/KUDU-3292 > Project: Kudu > Issue Type: Improvement > Components: ui >Reporter: Grant Henke >Assignee: Bakai Ádám >Priority: Minor > Labels: beginner, newbie, newbie++, trivial > Fix For: n/a > > Attachments: image-2023-01-10-11-57-13-209.png > > > Currently each Kudu server has a /varz webpage (the Flags tab) showing all of > the flags set on the server. It would be a nice usability change to include a > seperate section showing only the non-default flags. This should be super > straigtforward given we have the ability to get all the non-default flags via > GetNonDefaultFlags or GetNonDefaultFlagsMap in flags.cc. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (KUDU-3431) There is no binary package suitable for osx-aarch
[ https://issues.apache.org/jira/browse/KUDU-3431?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17677407#comment-17677407 ] Bakai Ádám commented on KUDU-3431: -- [~healchow] I created a short documentary review request about how to run the java examples on M1 Mac, it may help: [https://gerrit.cloudera.org/#/c/19418/] If it didn't help in your case, please describe it more in details where it causes problem that the MAC ARM64 jar does not exist. > There is no binary package suitable for osx-aarch > - > > Key: KUDU-3431 > URL: https://issues.apache.org/jira/browse/KUDU-3431 > Project: Kudu > Issue Type: Bug >Affects Versions: 1.15.0, 1.16.0 > Environment: MacBook M1 chip, OS Ventura 13.1 >Reporter: Heal Chow >Priority: Major > Attachments: image-2022-12-29-20-43-52-324.png > > > My development machine is MacBook of Apple M1 chip. During the development > process, I found that kudu binary does not adapt to osx aarch_ 64 > architecture. > > In this repository, I didn't find what I wanted. > [https://repo.maven.apache.org/maven2/org/apache/kudu/kudu-binary/1.16.0/] > !image-2022-12-29-20-43-52-324.png|width=577,height=287! -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3434) Couldn't close scanner : Service unavailable: reactor is shutting down (error 58) warning at big scans
Bakai Ádám created KUDU-3434: Summary: Couldn't close scanner : Service unavailable: reactor is shutting down (error 58) warning at big scans Key: KUDU-3434 URL: https://issues.apache.org/jira/browse/KUDU-3434 Project: Kudu Issue Type: Bug Components: client Environment: MAC Arm64 development environment Reporter: Bakai Ádám I ran {code:java} ./bin/kudu table scan localhost:8764 test_table {code} The test_table has 372000 and the command takes about 1.90579 seconds to execute. Most of the time it ends like this: {code:java} ...(rows)... (int32 key=371998, int32 integer_val=743996, int32 non_null_with_default=1859990, bool another_val=NULL) (int32 key=371999, int32 integer_val=743998, int32 non_null_with_default=1859995, bool another_val=NULL) Total count 372000 cost 1.90579 seconds W20230117 13:23:30.705157 1238496 client.cc:1955] Couldn't close scanner a92a2d1cc6394cc3837d67286ce25575: Service unavailable: reactor is shutting down (error 58) {code} If I run with the java-example test table(150 rows, around 0.003 seconds to execute), it never shows the warning. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3436) build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1 ARM64
Bakai Ádám created KUDU-3436: Summary: build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1 ARM64 Key: KUDU-3436 URL: https://issues.apache.org/jira/browse/KUDU-3436 Project: Kudu Issue Type: Bug Reporter: Bakai Ádám {code:java} build_mini_cluster_binaries.sh {code} returns the following error: {code:java} Traceback (most recent call last): File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 503, in main() File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 500, in main relocate_deps(target_src, target_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 408, in relocate_deps return relocate_deps_macos(target_src, target_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 397, in relocate_deps_macos copy_file(dep_src, dep_dst) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 287, in copy_file shutil.copyfile(src, dest) File "/opt/homebrew/Cellar/python@2/2.7.18/Frameworks/Python.framework/Versions/2.7/lib/python2.7/shutil.py", line 96, in copyfile with open(src, 'rb') as fsrc: IOError: [Errno 2] No such file or directory: u'/usr/lib/libc++abi.dylib' {code} After further investigation, it looks like libc++abi.dylib is in the uninstrumented lib, but otool -L always gives back a path for /usr/lib/libc++abi.dylib . Simply adding the dylib into the PAT_MACOS_LIB_EXCLUDE list doesn't work: it creates a jar file, but the binaries can not be started. It is probably due to the changes in how dynamic linking works in newer MacOS: https://stackoverflow.com/questions/70581876/macos-dynamic-linker-reports-it-loaded-library-which-doesnt-exist -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3436) build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1
[ https://issues.apache.org/jira/browse/KUDU-3436?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3436: - Description: {code:java} build_mini_cluster_binaries.sh {code} returns the following error: {code:java} Traceback (most recent call last): File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 503, in main() File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 500, in main relocate_deps(target_src, target_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 408, in relocate_deps return relocate_deps_macos(target_src, target_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 397, in relocate_deps_macos copy_file(dep_src, dep_dst) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 287, in copy_file shutil.copyfile(src, dest) File "/opt/homebrew/Cellar/python@2/2.7.18/Frameworks/Python.framework/Versions/2.7/lib/python2.7/shutil.py", line 96, in copyfile with open(src, 'rb') as fsrc: IOError: [Errno 2] No such file or directory: u'/usr/lib/libc++abi.dylib' {code} After further investigation, it looks like libc+{+}abi.dylib is in the uninstrumented lib, but otool -L always gives back a path for /usr/lib/libc{+}+abi.dylib . Simply adding the dylib into the PAT_MACOS_LIB_EXCLUDE list doesn't work: it creates a jar file, but the binaries can not be started. It is probably due to the changes in how dynamic linking works in newer MacOS: [https://stackoverflow.com/questions/70581876/macos-dynamic-linker-reports-it-loaded-library-which-doesnt-exist] It happens both on ARM64 and X86 was: {code:java} build_mini_cluster_binaries.sh {code} returns the following error: {code:java} Traceback (most recent call last): File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 503, in main() File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 500, in main relocate_deps(target_src, target_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 408, in relocate_deps return relocate_deps_macos(target_src, target_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 398, in relocate_deps_macos relocate_deps_macos(dep_src, dep_dst, config) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 397, in relocate_deps_macos copy_file(dep_src, dep_dst) File "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", line 287, in copy_file shutil.copyfile(src, dest) File "/opt/homebrew/Cellar/python@2/2.7.18/Frameworks/Python.framework/Versions/2.7/lib/python2.7/shutil.py", line 96, in copyfile with open(src, 'rb') as fsrc: IOError: [Errno 2] No such file or dir
[jira] [Updated] (KUDU-3436) build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1
[ https://issues.apache.org/jira/browse/KUDU-3436?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3436: - Summary: build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1 (was: build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1 ARM64) > build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1 > - > > Key: KUDU-3436 > URL: https://issues.apache.org/jira/browse/KUDU-3436 > Project: Kudu > Issue Type: Bug >Reporter: Bakai Ádám >Priority: Major > > > {code:java} > build_mini_cluster_binaries.sh {code} > returns the following error: > {code:java} > Traceback (most recent call last): > File > "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", > line 503, in > main() > File > "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", > line 500, in main > relocate_deps(target_src, target_dst, config) > File > "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", > line 408, in relocate_deps > return relocate_deps_macos(target_src, target_dst, config) > File > "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", > line 398, in relocate_deps_macos > relocate_deps_macos(dep_src, dep_dst, config) > File > "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", > line 398, in relocate_deps_macos > relocate_deps_macos(dep_src, dep_dst, config) > File > "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", > line 398, in relocate_deps_macos > relocate_deps_macos(dep_src, dep_dst, config) > File > "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", > line 398, in relocate_deps_macos > relocate_deps_macos(dep_src, dep_dst, config) > File > "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", > line 397, in relocate_deps_macos > copy_file(dep_src, dep_dst) > File > "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py", > line 287, in copy_file > shutil.copyfile(src, dest) > File > "/opt/homebrew/Cellar/python@2/2.7.18/Frameworks/Python.framework/Versions/2.7/lib/python2.7/shutil.py", > line 96, in copyfile > with open(src, 'rb') as fsrc: > IOError: [Errno 2] No such file or directory: u'/usr/lib/libc++abi.dylib' > {code} > After further investigation, it looks like libc++abi.dylib is in the > uninstrumented lib, but otool -L always gives back a path for > /usr/lib/libc++abi.dylib . Simply adding the dylib into the > PAT_MACOS_LIB_EXCLUDE list doesn't work: it creates a jar file, but the > binaries can not be started. > It is probably due to the changes in how dynamic linking works in newer > MacOS: > https://stackoverflow.com/questions/70581876/macos-dynamic-linker-reports-it-loaded-library-which-doesnt-exist > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3437) Starting Master server runs into error first time
Bakai Ádám created KUDU-3437: Summary: Starting Master server runs into error first time Key: KUDU-3437 URL: https://issues.apache.org/jira/browse/KUDU-3437 Project: Kudu Issue Type: Bug Components: cfile Environment: Macos 13.0.1 Arm64, fastdebug build Reporter: Bakai Ádám Reproduction script: {code:java} #Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in your path to run this script. export PATH=$PATH:./bin/stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data master-3/wal master-3/log; start_kudu.sh -m 3echo just to make sure it returns three masters as expected kudu master list localhost:8764,localhost:8766,localhost:8768echo This returns error: kudu-master --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=8771echo just to make sure it returns three masters kudu master list localhost:8764,localhost:8766,localhost:8768 echo Now it should be running fine: kudu-master --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=8771 &sleep 2 echo just to make sure it returns four masters kudu master list localhost:8764,localhost:8766,localhost:8768 {code} My error for the first run: {code:java} F20230127 15:58:28.923543 360774 cache.cc:579] Check failed: IsGTest() Metrics should only be set once per Cache *** Check failure stack trace: *** @ 0x1025950d4 google::LogMessageFatal::~LogMessageFatal() @ 0x102591930 google::LogMessageFatal::~LogMessageFatal() @ 0x10b56c4a8 kudu::(anonymous namespace)::ShardedCache<>::SetMetrics() @ 0x103522c1c kudu::cfile::BlockCache::StartInstrumentation() @ 0x101249230 kudu::master::Master::Init() @ 0x10126cbf0 kudu::master::RunMasterServer() @ 0x1005eae64 kudu::master::MasterMain() @ 0x1005ead10 main {code} If I uncomment the line {code:java} CHECK(IsGTest()) << "Metrics should only be set once per Cache"; {code} from kudu/util/cache.cc:579 ( CHECK(IsGTest()) << "Metrics should only be set once per Cache"<
[jira] [Updated] (KUDU-3437) Starting Master server runs into error first time
[ https://issues.apache.org/jira/browse/KUDU-3437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3437: - Description: Reproduction script: {noformat} #Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in your path to run this script. export PATH=$PATH:./bin/ stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data master-3/wal master-3/log; start_kudu.sh -m 3 echo just to make sure it returns three masters as expected kudu master list localhost:8764,localhost:8766,localhost:8768 echo This returns error: kudu-master --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=8771 echo just to make sure it returns three masters kudu master list localhost:8764,localhost:8766,localhost:8768 echo Now it should be running fine: kudu-master --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=8771 & sleep 2 echo just to make sure it returns four masters kudu master list localhost:8764,localhost:8766,localhost:8768 {noformat} My error for the first run: {code:java} F20230127 15:58:28.923543 360774 cache.cc:579] Check failed: IsGTest() Metrics should only be set once per Cache *** Check failure stack trace: *** @ 0x1025950d4 google::LogMessageFatal::~LogMessageFatal() @ 0x102591930 google::LogMessageFatal::~LogMessageFatal() @ 0x10b56c4a8 kudu::(anonymous namespace)::ShardedCache<>::SetMetrics() @ 0x103522c1c kudu::cfile::BlockCache::StartInstrumentation() @ 0x101249230 kudu::master::Master::Init() @ 0x10126cbf0 kudu::master::RunMasterServer() @ 0x1005eae64 kudu::master::MasterMain() @ 0x1005ead10 main {code} If I uncomment the line {code:java} CHECK(IsGTest()) << "Metrics should only be set once per Cache"; {code} from kudu/util/cache.cc:579 ( CHECK(IsGTest()) << "Metrics should only be set once per Cache"<::SetMetrics() @ 0x103522c1c kudu::cfile::BlockCache::StartInstrumentation() @ 0x101249230 kudu::master::Master::Init() @ 0x10126cbf0 kudu::master::RunMasterServer() @ 0x1005eae64 kudu::master::MasterMain() @ 0x1005ead10 main {code} If I uncomment the line {code:java} CHECK(IsGTest()) << "Metrics should only be set once per Cache"; {code} from kudu/util/cache.cc:579 ( CHECK(IsGTest()) << "Metrics should only be set once per Cache"< Starting Master server runs into error first time > - > > Key: KUDU-3437 > URL: https://issues.apache.org/jira/browse/KUDU-3437 > Project: Kudu > Issue Type: Bug > Components: cfile > Environment: Macos 13.0.1 Arm64, fastdebug build >Reporter: Bakai Ádám >Priority: Major > > Reproduction script: > {noformat} > #Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in > your path to run this script. > export PATH=$PATH:./bin/ > stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data master-3/wal > master-3/log; start_kudu.sh -m 3 > echo just to make sure it returns three masters as expected > kudu master list localhost:8764,localhost:8766,localhost:8768 > echo This returns error: > kudu-master > --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 > --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal > --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 > --time_source=system_unsync --unlock_unsafe_flags > --webserver_interface=127.0.0.1 --webserver_port=8771 > echo just to make sure it returns three masters > kudu master list localhost:8764,localhost:8766,localhost:8768 > echo Now it should be running fine: > kudu-master > --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 > --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal > --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 > --time_source=system_unsync --unlock_unsafe_flags > --webserver_interface=127.0.0.1 --webserver_port=8771 & > sleep 2 > echo just to make sure it returns four masters > kudu master list localhost:8764,localhost:8766,localhost:8768 >
[jira] [Updated] (KUDU-3437) Starting Master server runs into error first time
[ https://issues.apache.org/jira/browse/KUDU-3437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3437: - Description: Reproduction script: {noformat} #Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in your path to run this script. export PATH=$PATH:./bin/ stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data master-3/wal master-3/log; start_kudu.sh -m 3 echo just to make sure it returns three masters as expected kudu master list localhost:8764,localhost:8766,localhost:8768 echo This returns error: kudu-master --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=8771 echo just to make sure it returns three masters kudu master list localhost:8764,localhost:8766,localhost:8768 echo Now it should be running fine: kudu-master --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=8771 & sleep 2 echo just to make sure it returns four masters kudu master list localhost:8764,localhost:8766,localhost:8768 {noformat} My error for the first run: {code:java} F20230127 15:58:28.923543 360774 cache.cc:579] Check failed: IsGTest() Metrics should only be set once per Cache *** Check failure stack trace: *** @ 0x1025950d4 google::LogMessageFatal::~LogMessageFatal() @ 0x102591930 google::LogMessageFatal::~LogMessageFatal() @ 0x10b56c4a8 kudu::(anonymous namespace)::ShardedCache<>::SetMetrics() @ 0x103522c1c kudu::cfile::BlockCache::StartInstrumentation() @ 0x101249230 kudu::master::Master::Init() @ 0x10126cbf0 kudu::master::RunMasterServer() @ 0x1005eae64 kudu::master::MasterMain() @ 0x1005ead10 main {code} If I uncomment the line {code:java} CHECK(IsGTest()) << "Metrics should only be set once per Cache"; {code} from kudu/util/cache.cc:579 it seems to work fine. was: Reproduction script: {noformat} #Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in your path to run this script. export PATH=$PATH:./bin/ stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data master-3/wal master-3/log; start_kudu.sh -m 3 echo just to make sure it returns three masters as expected kudu master list localhost:8764,localhost:8766,localhost:8768 echo This returns error: kudu-master --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=8771 echo just to make sure it returns three masters kudu master list localhost:8764,localhost:8766,localhost:8768 echo Now it should be running fine: kudu-master --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=8771 & sleep 2 echo just to make sure it returns four masters kudu master list localhost:8764,localhost:8766,localhost:8768 {noformat} My error for the first run: {code:java} F20230127 15:58:28.923543 360774 cache.cc:579] Check failed: IsGTest() Metrics should only be set once per Cache *** Check failure stack trace: *** @ 0x1025950d4 google::LogMessageFatal::~LogMessageFatal() @ 0x102591930 google::LogMessageFatal::~LogMessageFatal() @ 0x10b56c4a8 kudu::(anonymous namespace)::ShardedCache<>::SetMetrics() @ 0x103522c1c kudu::cfile::BlockCache::StartInstrumentation() @ 0x101249230 kudu::master::Master::Init() @ 0x10126cbf0 kudu::master::RunMasterServer() @ 0x1005eae64 kudu::master::MasterMain() @ 0x1005ead10 main {code} If I uncomment the line {code:java} CHECK(IsGTest()) << "Metrics should only be set once per Cache"; {code} from kudu/util/cache.cc:579 ( CHECK(IsGTest()) << "Metrics should only be set once per Cache"< Starting Master server runs into error first time > - > > Key: KUDU-3437 > URL: https://issues.apache.org/jira/browse/KUDU-3437 > Project: Kud
[jira] [Created] (KUDU-3439) make cpp client example error message for mistyped master easier to understand
Bakai Ádám created KUDU-3439: Summary: make cpp client example error message for mistyped master easier to understand Key: KUDU-3439 URL: https://issues.apache.org/jira/browse/KUDU-3439 Project: Kudu Issue Type: Wish Components: client Reporter: Bakai Ádám If the user mistypes the master address, then he gets a very long error message. It definitely contains the useful part ("Could not connect to the cluster: unable to resolve address for localhosta: nodename nor servname provided, or not known") , but there are so many lines it's kind of hard to find: {noformat} adambakai@abakai-MBP16 examples % ./example localhosta:8764 Running with Kudu client version: kudu 1.17.0-SNAPSHOT (rev 244b21efb87927de795206b6825b64eb76cacfd9-dirty) Long version info: kudu 1.17.0-SNAPSHOT revision 244b21efb87927de795206b6825b64eb76cacfd9-dirty build type FASTDEBUG built by adambakai at 31 Jan 2023 11:06:55 CET on abakai-MBP16 Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/openssl_util.cc Line number: 109 Time: Tue Jan 31 11:15:39 2023 Message: FIPS mode is disabled. Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/security/tls_context.cc Line number: 297 Time: Tue Jan 31 11:15:39 2023 Message: OpenSSL security level is 1 Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/thread.cc Line number: 658 Time: Tue Jan 31 11:15:39 2023 Message: Started thread 740961 - kernel-watchdog: kernel-watcher Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/thread.cc Line number: 658 Time: Tue Jan 31 11:15:39 2023 Message: Started thread 740962 - reactor: rpc reactor Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/thread.cc Line number: 658 Time: Tue Jan 31 11:15:39 2023 Message: Started thread 740963 - reactor: rpc reactor Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/thread.cc Line number: 658 Time: Tue Jan 31 11:15:39 2023 Message: Started thread 740964 - reactor: rpc reactor Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/thread.cc Line number: 658 Time: Tue Jan 31 11:15:39 2023 Message: Started thread 740965 - reactor: rpc reactor Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/cache.cc Line number: 541 Time: Tue Jan 31 11:15:39 2023 Message: Will use 16 shards for recency list cache. Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/mem_tracker.cc Line number: 83 Time: Tue Jan 31 11:15:39 2023 Message: Creating tracker root Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/mem_tracker.cc Line number: 83 Time: Tue Jan 31 11:15:39 2023 Message: Creating tracker ttl-cache-sharded_fifo_cache->root Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/ttl_cache.h Line number: 134 Time: Tue Jan 31 11:15:39 2023 Message: constructed TTL cache 'ttl-cache' with capacity of 1048576 Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/mem_tracker.cc Line number: 87 Time: Tue Jan 31 11:15:39 2023 Message: Destroying tracker ttl-cache-sharded_fifo_cache->root Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/rpc/messenger.cc Line number: 185 Time: Tue Jan 31 11:15:39 2023 Message: shutting down messenger client Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/rpc/reactor.cc Line number: 236 Time: Tue Jan 31 11:15:39 2023 Message: client_R000: shutting down Reactor thread. Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/rpc/reactor.cc Line number: 236 Time: Tue Jan 31 11:15:39 2023 Message: client_R001: shutting down Reactor thread. Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/rpc/reactor.cc Line number: 251 Time: Tue Jan 31 11:15:39 2023 Message: client_R000: tearing down outbound TCP connections... Received log message from Kudu client library Severity: 0 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/rpc/reactor.cc Line number: 236
[jira] [Assigned] (KUDU-3437) Starting Master server runs into error first time
[ https://issues.apache.org/jira/browse/KUDU-3437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3437: Assignee: Bakai Ádám > Starting Master server runs into error first time > - > > Key: KUDU-3437 > URL: https://issues.apache.org/jira/browse/KUDU-3437 > Project: Kudu > Issue Type: Bug > Components: cfile > Environment: Macos 13.0.1 Arm64, fastdebug build >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > > Reproduction script: > {noformat} > #Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in > your path to run this script. > export PATH=$PATH:./bin/ > stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data master-3/wal > master-3/log; start_kudu.sh -m 3 > echo just to make sure it returns three masters as expected > kudu master list localhost:8764,localhost:8766,localhost:8768 > echo This returns error: > kudu-master > --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 > --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal > --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 > --time_source=system_unsync --unlock_unsafe_flags > --webserver_interface=127.0.0.1 --webserver_port=8771 > echo just to make sure it returns three masters > kudu master list localhost:8764,localhost:8766,localhost:8768 > echo Now it should be running fine: > kudu-master > --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 > --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal > --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 > --time_source=system_unsync --unlock_unsafe_flags > --webserver_interface=127.0.0.1 --webserver_port=8771 & > sleep 2 > echo just to make sure it returns four masters > kudu master list localhost:8764,localhost:8766,localhost:8768 > {noformat} > My error for the first run: > {code:java} > F20230127 15:58:28.923543 360774 cache.cc:579] Check failed: IsGTest() > Metrics should only be set once per Cache > *** Check failure stack trace: *** > @ 0x1025950d4 google::LogMessageFatal::~LogMessageFatal() > @ 0x102591930 google::LogMessageFatal::~LogMessageFatal() > @ 0x10b56c4a8 kudu::(anonymous > namespace)::ShardedCache<>::SetMetrics() > @ 0x103522c1c kudu::cfile::BlockCache::StartInstrumentation() > @ 0x101249230 kudu::master::Master::Init() > @ 0x10126cbf0 kudu::master::RunMasterServer() > @ 0x1005eae64 kudu::master::MasterMain() > @ 0x1005ead10 main {code} > If I uncomment the line > {code:java} > CHECK(IsGTest()) << "Metrics should only be set once per Cache"; > {code} > from kudu/util/cache.cc:579 it seems to work fine. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3491) MiniDumpExceptionHandler assert randomly fails
Bakai Ádám created KUDU-3491: Summary: MiniDumpExceptionHandler assert randomly fails Key: KUDU-3491 URL: https://issues.apache.org/jira/browse/KUDU-3491 Project: Kudu Issue Type: Bug Affects Versions: 1.15.0 Reporter: Bakai Ádám When starting master, this error randomly happens on my system: {code:java} + exec /opt/cloudera/parcels/CDH-7.2.18-1.cdh7.2.18.p0.43161468/lib/kudu/sbin/kudu-master --master_addresses=abakai-1.abakai.root.hwx.site,abakai-2.abakai.root.hwx.site,abakai-3.abakai.root.hwx.site --location_mapping_cmd=/var/run/cloudera-scm-agent/process/126-kudu-KUDU_MASTER/topology.py --flagfile=/var/run/cloudera-scm-agent/process/126-kudu-KUDU_MASTER/gflagfile F20230717 10:37:23.626719 101405 minidump.cc:273] Check failed: 0 == MinidumpExceptionHandler::current_num_instances_.fetch_add(1) (0 vs. 1) *** Check failure stack trace: *** Wrote minidump to /var/log/kudu/minidumps/kudu-master/1664cde8-f41a-4b7f-121578b0-55ba68a6.dmp *** Aborted at 1689590243 (unix time) try "date -d @1689590243" if you are using GNU date *** PC: @ 0x0 (unknown) *** SIGABRT (@0x9c2900018c1d) received by PID 101405 (TID 0x7f204a7cda00) from PID 101405; stack trace: *** @ 0xe4df76 google::(anonymous namespace)::FailureSignalHandler() @ 0x7f204a19e6d0 (unknown) @ 0x7f20483a6277 __GI_raise @ 0x7f20483a7968 __GI_abort @ 0xcb50af kudu::AbortFailureFunction() @ 0xe4296d google::LogMessage::Fail() @ 0xe4584a google::LogMessage::SendToLog() @ 0xe4249e google::LogMessage::Flush() @ 0xe439d9 google::LogMessageFatal::~LogMessageFatal() @ 0x319a0a1 kudu::MinidumpExceptionHandler::RegisterMinidumpExceptionHandler() @ 0x319a107 kudu::MinidumpExceptionHandler::MinidumpExceptionHandler() @ 0x12821f0 kudu::server::ServerBase::ServerBase() @ 0x124577e kudu::kserver::KuduServer::KuduServer() @ 0xddc9ce kudu::master::Master::Master() @ 0xd5732b kudu::master::RunMasterServer() @ 0xd51d8a kudu::master::MasterMain() @ 0x7f2048392445 __libc_start_main @ 0xd51b64 (unknown) {code} Version information: {code:java} [root@abakai-1 ~]# /opt/cloudera/parcels/CDH-7.2.18-1.cdh7.2.18.p0.43161468/lib/kudu/sbin-release/kudu-master -version kudu 1.15.0.7.2.18.0-205 revision 7e222133b1a13ce6c212ffb32d8ceaa0c6a8545a build type RELEASE built by None at 14 Jul 2023 06:51:13 UTC on re-centos-slave-large-2wczs build id 1178567 {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3505) kudu ksck fails if healthy master in healthy cluster is started after the command
Bakai Ádám created KUDU-3505: Summary: kudu ksck fails if healthy master in healthy cluster is started after the command Key: KUDU-3505 URL: https://issues.apache.org/jira/browse/KUDU-3505 Project: Kudu Issue Type: Bug Components: master Reporter: Bakai Ádám Environment: single master configuration If the master is not running and the user starts a kudu cluster ksck command then it will try to connect to the master over and over again. Once the master is started, the ksck command is executed and it shows a bunch of errors: {code:java} adambakai@abakai-MBP16 d % kudu cluster ksck localhost:8764 -ksck_format plain_full Master Summary UUID | Address | Status --++- f41052a1ba8242d49ee5e16c0d60558a | localhost:8764 | HEALTHY All reported replicas are: A = f41052a1ba8242d49ee5e16c0d60558a The consensus matrix is: Config source | Replicas | Current term | Config index | Committed? ---+--+--+--+ A | A* | 16 | -1 | YesFlags of checked categories for Master: Flag | Value | Master -+-+- builtin_ntp_servers | 0.pool.ntp.org,1.pool.ntp.org,2.pool.ntp.org,3.pool.ntp.org | all 1 server(s) checked time_source | system_unsync | all 1 server(s) checkedTablet Server Summary UUID | Address | Status | Location | Tablet Leaders | Active Scanners --++-+--++- 1938796538bf483f9bcd133e29aa645b | 127.0.0.1:9878 | HEALTHY | | 0 | 0 8080a72aeb714c5087b8c515f21b1735 | 127.0.0.1:9870 | HEALTHY | | 1 | 0 9f86252d00814cb3ae0ef6858ee31a02 | 127.0.0.1:9874 | HEALTHY | | 0 | 0 c23de9c2b3e1448fa8dde2bb1a292388 | 127.0.0.1:9872 | HEALTHY | | 0 | 0 fb700997c9274a9d8287eb3c765606d2 | 127.0.0.1:9876 | HEALTHY | | 0 | 0Tablet Server Location Summary Location | Count --+- | 5Flags of checked categories for Tablet Server: Flag | Value | Tablet Server -+-+- builtin_ntp_servers | 0.pool.ntp.org,1.pool.ntp.org,2.pool.ntp.org,3.pool.ntp.org | all 5 server(s) checked time_source | system_unsync | all 5 server(s) checkedVersion Summary Version | Servers -+- 1.18.0-SNAPSHOT | all 6 server(s) checkedTablet Summary Tablet 5d87f015c3a2438c8cec6e84796f9ecb of table 'db.test_table' is healthy. 8080a72aeb714c5087b8c515f21b1735 (127.0.0.1:9870): RUNNING [LEADER] c23de9c2b3e1448fa8dde2bb1a292388 (127.0.0.1:9872): RUNNING 9f86252d00814cb3ae0ef6858ee31a02 (127.0.0.1:9874): RUNNING All reported replicas are: A = 8080a72aeb714c5087b8c515f21b1735 B = c23de9c2b3e1448fa8dde2bb1a292388 C = 9f86252d00814cb3ae0ef6858ee31a02 The consensus matrix is: Config source | Replicas | Current term | Config index | Committed? ---+--+--+--+ master | A* B C | | | Yes A | A* B C | 3 | -1 | Yes B | A* B C | 3 | -1 | Yes C | A* B C | 3 | -1 | YesThe cluster doesn't have any matching system tables Summary by table Name | RF | Status | Total Tablets | Healthy | Recovering | Under-replicated | Unavailable ---++-+---+-++--+- db.test_table | 3 | HEALTHY | 1 | 1 | 0 | 0 | 0Tablet Replica Count Summary Statistic | Replica Count +--- Minimum | 0 First Quartile | 0 Median | 1 Third Quartile | 1 Maximum | 1Tablet Replica Count by Tablet Server UUID | Host | Replica Count --++--- 1938796538bf483f9bcd133e29aa645b | 127.0.0.1:9878 | 0 8080a72aeb714c5087b8c515f21b1735 | 127.0.0.1:9870 | 1 9f86252d00814cb3ae0ef6858ee31a02 | 127.0.0.1:9874 | 1 c23de9c2b3e1448fa8dde2bb1a292388 | 127.0.0.1:9872 | 1 fb700997c9274a9d8287eb3c765606d2 |
[jira] [Created] (KUDU-3547) kudu master add failes if new master port is not the default value
Bakai Ádám created KUDU-3547: Summary: kudu master add failes if new master port is not the default value Key: KUDU-3547 URL: https://issues.apache.org/jira/browse/KUDU-3547 Project: Kudu Issue Type: Bug Components: master Reporter: Bakai Ádám Tried this script on de87aca3dd29177658790cffeb7a1de0c8d69231 : {code:java} PATH=$PATH:./binPATH=$PATH:../../src/kudu/scripts/ cleanup(){kill %1 %2 %3 %4 %5 %6 %7 %8 %9 %10pkill kudu# sudo fuser -k -9 9873/tcpfuser -k -9 8765/tcp pkill kudupkill kudu-masterpkill kudu-tserver../src/kudu/scripts/stop_kudu.sh sleep 1} cleanup rm -rf master-0/ tserver-0/ tserver-1 tserver-2 master-1 master-2create_stuff(){rm -rf $1/data $1/wal $1/log mkdir -p $1/data $1/wal $1/log} create_stuff master-0create_stuff tserver-0create_stuff tserver-1create_stuff tserver-2create_stuff master-1 sync $PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-0/data --fs_wal_dir=$PWD/tserver-0/wal --log_dir=$PWD/tserver-0/log --rpc_bind_addresses=127.0.0.1:9870 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9871 --tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 &$PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-1/data --fs_wal_dir=$PWD/tserver-1/wal --log_dir=$PWD/tserver-1/log --rpc_bind_addresses=127.0.0.1:9872 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9873 --tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 &$PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-2/data --fs_wal_dir=$PWD/tserver-2/wal --log_dir=$PWD/tserver-2/log --rpc_bind_addresses=127.0.0.1:9874 --time_source=system_unsync --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9875 --tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 & sleep 1 $PWD/bin/kudu-master --master_addresses=127.0.0.1:8764 --fs_data_dirs=$PWD/master-0/data --fs_wal_dir=$PWD/master-0/wal --log_dir=$PWD/master-0/log --rpc_bind_addresses=127.0.0.1:8764 --time_source=system_unsync --unlock_unsafe_flags --webserver_port=8765 & #/home/abakai/kudu_1_16_2/kudu/a/b/bin/kudu-master --master_addresses=127.0.0.1:8764 --fs_data_dirs=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/data --fs_wal_dir=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/wal --log_dir=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/log --rpc_bind_addresses=127.0.0.1:8764 --time_source=system_unsync --unlock_unsafe_flags --webserver_port=8765 --webserver_doc_root=/home/abakai/kudu_1_16_2/kudu/a/b/../../www kudu master list localhost:8764 sleep 1kudu master add 127.0.0.1:8764 127.0.0.1:8766 --fs_data_dirs=$PWD/master-1/data --fs_wal_dir=$PWD/master-1/wal -logtostderr -v 99 &>test123 kudu master list localhost:8764 {code} New master is not added to the master quorum. There is a workaround, adding rpc_bind_address to the arguments: {code:java} kudu master add 127.0.0.1:8764 127.0.0.1:8766 --fs_data_dirs=$PWD/master-1/data --fs_wal_dir=$PWD/master-1/wal -logtostderr -v 1 --rpc_bind_addresses=127.0.0.1:8766 {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3547) kudu master add fails if new master port is not the default value
[ https://issues.apache.org/jira/browse/KUDU-3547?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3547: - Summary: kudu master add fails if new master port is not the default value (was: kudu master add failes if new master port is not the default value) > kudu master add fails if new master port is not the default value > - > > Key: KUDU-3547 > URL: https://issues.apache.org/jira/browse/KUDU-3547 > Project: Kudu > Issue Type: Bug > Components: master >Reporter: Bakai Ádám >Priority: Minor > > Tried this script on de87aca3dd29177658790cffeb7a1de0c8d69231 : > {code:java} > PATH=$PATH:./binPATH=$PATH:../../src/kudu/scripts/ > cleanup(){kill %1 %2 %3 %4 %5 %6 %7 %8 %9 %10pkill kudu# sudo fuser -k -9 > 9873/tcpfuser -k -9 8765/tcp > pkill kudupkill kudu-masterpkill kudu-tserver../src/kudu/scripts/stop_kudu.sh > sleep 1} > cleanup > rm -rf master-0/ tserver-0/ tserver-1 tserver-2 master-1 > master-2create_stuff(){rm -rf $1/data $1/wal $1/log > mkdir -p $1/data $1/wal $1/log} > create_stuff master-0create_stuff tserver-0create_stuff tserver-1create_stuff > tserver-2create_stuff master-1 > sync > $PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-0/data > --fs_wal_dir=$PWD/tserver-0/wal --log_dir=$PWD/tserver-0/log > --rpc_bind_addresses=127.0.0.1:9870 --time_source=system_unsync > --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9871 > --tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 > &$PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-1/data > --fs_wal_dir=$PWD/tserver-1/wal --log_dir=$PWD/tserver-1/log > --rpc_bind_addresses=127.0.0.1:9872 --time_source=system_unsync > --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9873 > --tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 > &$PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-2/data > --fs_wal_dir=$PWD/tserver-2/wal --log_dir=$PWD/tserver-2/log > --rpc_bind_addresses=127.0.0.1:9874 --time_source=system_unsync > --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9875 > --tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 & > sleep 1 > $PWD/bin/kudu-master --master_addresses=127.0.0.1:8764 > --fs_data_dirs=$PWD/master-0/data --fs_wal_dir=$PWD/master-0/wal > --log_dir=$PWD/master-0/log --rpc_bind_addresses=127.0.0.1:8764 > --time_source=system_unsync --unlock_unsafe_flags --webserver_port=8765 & > #/home/abakai/kudu_1_16_2/kudu/a/b/bin/kudu-master > --master_addresses=127.0.0.1:8764 > --fs_data_dirs=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/data > --fs_wal_dir=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/wal > --log_dir=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/log > --rpc_bind_addresses=127.0.0.1:8764 --time_source=system_unsync > --unlock_unsafe_flags --webserver_port=8765 > --webserver_doc_root=/home/abakai/kudu_1_16_2/kudu/a/b/../../www > kudu master list localhost:8764 > sleep 1kudu master add 127.0.0.1:8764 127.0.0.1:8766 > --fs_data_dirs=$PWD/master-1/data --fs_wal_dir=$PWD/master-1/wal -logtostderr > -v 99 &>test123 > kudu master list localhost:8764 > {code} > New master is not added to the master quorum. > There is a workaround, adding rpc_bind_address to the arguments: > {code:java} > kudu master add 127.0.0.1:8764 127.0.0.1:8766 > --fs_data_dirs=$PWD/master-1/data --fs_wal_dir=$PWD/master-1/wal -logtostderr > -v 1 --rpc_bind_addresses=127.0.0.1:8766 > {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3548) kudu master add times out if new master argument is an ip address that can be resolved
Bakai Ádám created KUDU-3548: Summary: kudu master add times out if new master argument is an ip address that can be resolved Key: KUDU-3548 URL: https://issues.apache.org/jira/browse/KUDU-3548 Project: Kudu Issue Type: Bug Reporter: Bakai Ádám It is produced on kudu on de87aca3dd29177658790cffeb7a1de0c8d69231: {code:java} kudu master add 127.0.0.1:8764 127.0.0.1 --fs_data_dirs=$PWD/master-1/data --fs_wal_dir=$PWD/master-1/wal -logtostderr -v 1 {code} In CheckMasterVoterAndHealthy the elements in master.registration().rpc_addresses() are: 127.0.0.1:8764 and adam-7070-ubuntu:7051 (adam-7070 is my hostname), and the new_master_hp.ToString() is 127.0.0.1:7051 which won't match to adam-7070-ubuntu:7051 even though they are the same. This makes CheckMasterVoterAndHealthy time out, but the new master still gets into the raft quorum. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-2203) java.io.FileNotFoundException when trying to initialize MiniKuduCluster
Nacho García Fernández created KUDU-2203: Summary: java.io.FileNotFoundException when trying to initialize MiniKuduCluster Key: KUDU-2203 URL: https://issues.apache.org/jira/browse/KUDU-2203 Project: Kudu Issue Type: Bug Components: java, test Affects Versions: 1.5.0 Reporter: Nacho García Fernández I'm getting the following error when I try to create a new instance of MiniKuduCluster: {code:java} java.io.FileNotFoundException: Cannot find binary kudu-master in binary directory null at org.apache.kudu.client.TestUtils.findBinary(TestUtils.java:159) at org.apache.kudu.client.MiniKuduCluster.startMasters(MiniKuduCluster.java:210) at org.apache.kudu.client.MiniKuduCluster.startCluster(MiniKuduCluster.java:153) at org.apache.kudu.client.MiniKuduCluster.start(MiniKuduCluster.java:117) at org.apache.kudu.client.MiniKuduCluster.access$300(MiniKuduCluster.java:50) at org.apache.kudu.client.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:661) at org.apache.kudu.client.BaseKuduTest.doSetup(BaseKuduTest.java:113) at org.apache.kudu.client.BaseKuduTest.setUpBeforeClass(BaseKuduTest.java:76) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.junit.runner.JUnitCore.run(JUnitCore.java:137) at com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:68) at com.intellij.rt.execution.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:47) at com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:242) at com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:70) {code} This is happening when a base class try to extend from BaseKuduTest (which internally instantiates a MiniKuduCluster). My simple test class: {code:java} import org.apache.kudu.client.BaseKuduTest; import org.junit.Before; import org.junit.Test; public class KuduInputFormatTest extends BaseKuduTest { @Before public void initialize() throws Exception { } @Test public void test() throws Exception { System.out.println("The error occurred before this message is printed"); } {code} Current POM dependencies: {noformat} org.apache.kudu kudu-client ${kudu.version} org.apache.kudu kudu-client ${kudu.version} test-jar test {noformat} where ${kudu.version} is 1.5.0 -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Updated] (KUDU-2203) java.io.FileNotFoundException when trying to initialize MiniKuduCluster
[ https://issues.apache.org/jira/browse/KUDU-2203?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nacho García Fernández updated KUDU-2203: - Description: I'm getting the following error when I try to create a new instance of MiniKuduCluster: {code:java} java.io.FileNotFoundException: Cannot find binary kudu-master in binary directory null at org.apache.kudu.client.TestUtils.findBinary(TestUtils.java:159) at org.apache.kudu.client.MiniKuduCluster.startMasters(MiniKuduCluster.java:210) at org.apache.kudu.client.MiniKuduCluster.startCluster(MiniKuduCluster.java:153) at org.apache.kudu.client.MiniKuduCluster.start(MiniKuduCluster.java:117) at org.apache.kudu.client.MiniKuduCluster.access$300(MiniKuduCluster.java:50) at org.apache.kudu.client.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:661) at org.apache.kudu.client.BaseKuduTest.doSetup(BaseKuduTest.java:113) at org.apache.kudu.client.BaseKuduTest.setUpBeforeClass(BaseKuduTest.java:76) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.junit.runner.JUnitCore.run(JUnitCore.java:137) at com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:68) at com.intellij.rt.execution.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:47) at com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:242) at com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:70) {code} This is happening when a base class try to extend from BaseKuduTest (which internally instantiates a MiniKuduCluster). My simple test class: {code:java} import org.apache.kudu.client.BaseKuduTest; import org.junit.Before; import org.junit.Test; public class KuduInputFormatTest extends BaseKuduTest { @Before public void initialize() throws Exception { } @Test public void test() throws Exception { System.out.println("The error occurs before this message is printed"); } {code} Current POM dependencies: {noformat} org.apache.kudu kudu-client ${kudu.version} org.apache.kudu kudu-client ${kudu.version} test-jar test {noformat} where kudu.version is 1.5.0 was: I'm getting the following error when I try to create a new instance of MiniKuduCluster: {code:java} java.io.FileNotFoundException: Cannot find binary kudu-master in binary directory null at org.apache.kudu.client.TestUtils.findBinary(TestUtils.java:159) at org.apache.kudu.client.MiniKuduCluster.startMasters(MiniKuduCluster.java:210) at org.apache.kudu.client.MiniKuduCluster.startCluster(MiniKuduCluster.java:153) at org.apache.kudu.client.MiniKuduCluster.start(MiniKuduCluster.java:117) at org.apache.kudu.client.MiniKuduCluster.access$300(MiniKuduCluster.java:50) at org.apache.kudu.client.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:661) at org.apache.kudu.client.BaseKuduTest.doSetup(BaseKuduTest.java:113) at org.apache.kudu.client.BaseKuduTest.setUpBeforeClass(BaseKuduTest.java:76) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters
[jira] [Commented] (KUDU-2203) java.io.FileNotFoundException when trying to initialize MiniKuduCluster
[ https://issues.apache.org/jira/browse/KUDU-2203?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16224480#comment-16224480 ] Nacho García Fernández commented on KUDU-2203: -- Hi Jean-Daniel. Thanks, it was useful. Is there any plan to provide an embedded kudu driver to perform tests without needing a real kudu deployment? (I can move this question to Stackoverflow if needed). > java.io.FileNotFoundException when trying to initialize MiniKuduCluster > > > Key: KUDU-2203 > URL: https://issues.apache.org/jira/browse/KUDU-2203 > Project: Kudu > Issue Type: Bug > Components: java, test >Affects Versions: 1.5.0 >Reporter: Nacho García Fernández > > I'm getting the following error when I try to create a new instance of > MiniKuduCluster: > {code:java} > java.io.FileNotFoundException: Cannot find binary kudu-master in binary > directory null > at org.apache.kudu.client.TestUtils.findBinary(TestUtils.java:159) > at > org.apache.kudu.client.MiniKuduCluster.startMasters(MiniKuduCluster.java:210) > at > org.apache.kudu.client.MiniKuduCluster.startCluster(MiniKuduCluster.java:153) > at > org.apache.kudu.client.MiniKuduCluster.start(MiniKuduCluster.java:117) > at > org.apache.kudu.client.MiniKuduCluster.access$300(MiniKuduCluster.java:50) > at > org.apache.kudu.client.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:661) > at org.apache.kudu.client.BaseKuduTest.doSetup(BaseKuduTest.java:113) > at > org.apache.kudu.client.BaseKuduTest.setUpBeforeClass(BaseKuduTest.java:76) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) > at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) > at > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) > at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) > at org.junit.runners.ParentRunner.run(ParentRunner.java:363) > at org.junit.runner.JUnitCore.run(JUnitCore.java:137) > at > com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:68) > at > com.intellij.rt.execution.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:47) > at > com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:242) > at > com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:70) > {code} > This is happening when a base class try to extend from BaseKuduTest (which > internally instantiates a MiniKuduCluster). > My simple test class: > {code:java} > import org.apache.kudu.client.BaseKuduTest; > import org.junit.Before; > import org.junit.Test; > public class KuduInputFormatTest extends BaseKuduTest { > @Before > public void initialize() throws Exception { > } > @Test > public void test() throws Exception { > System.out.println("The error occurs before this message is printed"); > } > {code} > Current POM dependencies: > {noformat} > > org.apache.kudu > kudu-client > ${kudu.version} > > > org.apache.kudu > kudu-client > ${kudu.version} > test-jar > test > > {noformat} > where kudu.version is 1.5.0 -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0
Nacho García Fernández created KUDU-2223: Summary: Failed to add distributed masters: Unable to start Master at index 0 Key: KUDU-2223 URL: https://issues.apache.org/jira/browse/KUDU-2223 Project: Kudu Issue Type: Bug Components: build, java Affects Versions: 1.5.0 Reporter: Nacho García Fernández After successfuly building Kudu on my OSX, I try to mvn verify in the java submodule, but I get the following exception: {code:java} [INFO] --- [INFO] T E S T S [INFO] --- [INFO] Running org.apache.kudu.client.TestAlterTable [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.42 s <<< FAILURE! - in org.apache.kudu.client.TestAlterTable [ERROR] org.apache.kudu.client.TestAlterTable Time elapsed: 0.42 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 [INFO] Running org.apache.kudu.client.TestAsyncKuduClient [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.644 s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduClient [ERROR] org.apache.kudu.client.TestAsyncKuduClient Time elapsed: 0.644 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAsyncKuduClient.setUpBeforeClass(TestAsyncKuduClient.java:45) [INFO] Running org.apache.kudu.client.TestAsyncKuduSession [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.478 s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduSession [ERROR] org.apache.kudu.client.TestAsyncKuduSession Time elapsed: 0.478 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAsyncKuduSession.setUpBeforeClass(TestAsyncKuduSession.java:59) [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquire [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.601 s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquire [ERROR] org.apache.kudu.client.TestAuthnTokenReacquire Time elapsed: 0.601 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAuthnTokenReacquire.setUpBeforeClass(TestAuthnTokenReacquire.java:55) [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquireOpen [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.865 s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquireOpen [ERROR] org.apache.kudu.client.TestAuthnTokenReacquireOpen Time elapsed: 0.865 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to start a single Master: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAuthnTokenReacquireOpen.setUpBeforeClass(TestAuthnTokenReacquireOpen.java:58) [INFO] Running org.apache.kudu.client.TestBitSet [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.061 s - in org.apache.kudu.client.TestBitSet [INFO] Running org.apache.kudu.client.TestBytes [INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.077 s - in org.apache.kudu.client.TestBytes [INFO] Running org.apache.kudu.client.TestClientFailoverSupport [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.399 s <<< FAILURE! - in org.apache.kudu.client.TestClientFailoverSupport [ERROR] org.apache.kudu.client.TestClientFailoverSupport Time elapsed: 0.399 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestClientFailoverSupport.setUpBeforeClass(TestClientFailoverSupport.java:35) [INFO] Running org.apache.kudu.client.TestColumnRangePredicate [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.106 s - in org.apache.kudu.client.TestColumnRangePredicate [INFO] Running org.apache.kudu.client.Tes
[jira] [Updated] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0
[ https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nacho García Fernández updated KUDU-2223: - Description: After successfully building Kudu on my OSX, I try to mvn verify in the java submodule, but I get the following exception: {code:java} [INFO] --- [INFO] T E S T S [INFO] --- [INFO] Running org.apache.kudu.client.TestAlterTable [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.42 s <<< FAILURE! - in org.apache.kudu.client.TestAlterTable [ERROR] org.apache.kudu.client.TestAlterTable Time elapsed: 0.42 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 [INFO] Running org.apache.kudu.client.TestAsyncKuduClient [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.644 s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduClient [ERROR] org.apache.kudu.client.TestAsyncKuduClient Time elapsed: 0.644 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAsyncKuduClient.setUpBeforeClass(TestAsyncKuduClient.java:45) [INFO] Running org.apache.kudu.client.TestAsyncKuduSession [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.478 s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduSession [ERROR] org.apache.kudu.client.TestAsyncKuduSession Time elapsed: 0.478 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAsyncKuduSession.setUpBeforeClass(TestAsyncKuduSession.java:59) [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquire [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.601 s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquire [ERROR] org.apache.kudu.client.TestAuthnTokenReacquire Time elapsed: 0.601 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAuthnTokenReacquire.setUpBeforeClass(TestAuthnTokenReacquire.java:55) [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquireOpen [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.865 s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquireOpen [ERROR] org.apache.kudu.client.TestAuthnTokenReacquireOpen Time elapsed: 0.865 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to start a single Master: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAuthnTokenReacquireOpen.setUpBeforeClass(TestAuthnTokenReacquireOpen.java:58) [INFO] Running org.apache.kudu.client.TestBitSet [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.061 s - in org.apache.kudu.client.TestBitSet [INFO] Running org.apache.kudu.client.TestBytes [INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.077 s - in org.apache.kudu.client.TestBytes [INFO] Running org.apache.kudu.client.TestClientFailoverSupport [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.399 s <<< FAILURE! - in org.apache.kudu.client.TestClientFailoverSupport [ERROR] org.apache.kudu.client.TestClientFailoverSupport Time elapsed: 0.399 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestClientFailoverSupport.setUpBeforeClass(TestClientFailoverSupport.java:35) [INFO] Running org.apache.kudu.client.TestColumnRangePredicate [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.106 s - in org.apache.kudu.client.TestColumnRangePredicate [INFO] Running org.apache.kudu.client.TestConnectionCache [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.43 s <<< FAILURE! - in org.apache.kudu.client.TestConnectionCache [ERROR] test(org.apache.kudu.client.TestConnectio
[jira] [Updated] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0
[ https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nacho García Fernández updated KUDU-2223: - Description: After successfully building Kudu on my OSX, I try to run the mvn verify command in the java submodule, but I get the following exception: {code:java} [INFO] --- [INFO] T E S T S [INFO] --- [INFO] Running org.apache.kudu.client.TestAlterTable [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.42 s <<< FAILURE! - in org.apache.kudu.client.TestAlterTable [ERROR] org.apache.kudu.client.TestAlterTable Time elapsed: 0.42 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 [INFO] Running org.apache.kudu.client.TestAsyncKuduClient [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.644 s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduClient [ERROR] org.apache.kudu.client.TestAsyncKuduClient Time elapsed: 0.644 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAsyncKuduClient.setUpBeforeClass(TestAsyncKuduClient.java:45) [INFO] Running org.apache.kudu.client.TestAsyncKuduSession [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.478 s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduSession [ERROR] org.apache.kudu.client.TestAsyncKuduSession Time elapsed: 0.478 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAsyncKuduSession.setUpBeforeClass(TestAsyncKuduSession.java:59) [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquire [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.601 s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquire [ERROR] org.apache.kudu.client.TestAuthnTokenReacquire Time elapsed: 0.601 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAuthnTokenReacquire.setUpBeforeClass(TestAuthnTokenReacquire.java:55) [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquireOpen [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.865 s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquireOpen [ERROR] org.apache.kudu.client.TestAuthnTokenReacquireOpen Time elapsed: 0.865 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to start a single Master: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAuthnTokenReacquireOpen.setUpBeforeClass(TestAuthnTokenReacquireOpen.java:58) [INFO] Running org.apache.kudu.client.TestBitSet [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.061 s - in org.apache.kudu.client.TestBitSet [INFO] Running org.apache.kudu.client.TestBytes [INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.077 s - in org.apache.kudu.client.TestBytes [INFO] Running org.apache.kudu.client.TestClientFailoverSupport [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.399 s <<< FAILURE! - in org.apache.kudu.client.TestClientFailoverSupport [ERROR] org.apache.kudu.client.TestClientFailoverSupport Time elapsed: 0.399 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestClientFailoverSupport.setUpBeforeClass(TestClientFailoverSupport.java:35) [INFO] Running org.apache.kudu.client.TestColumnRangePredicate [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.106 s - in org.apache.kudu.client.TestColumnRangePredicate [INFO] Running org.apache.kudu.client.TestConnectionCache [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.43 s <<< FAILURE! - in org.apache.kudu.client.TestConnectionCache [ERROR] test(org.apache.kudu.clie
[jira] [Updated] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0
[ https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nacho García Fernández updated KUDU-2223: - Description: After successfully building Kudu on my OSX, I try to run the mvn verify command in the java submodule, but I get the following exception: {code:java} [INFO] --- [INFO] T E S T S [INFO] --- [INFO] Running org.apache.kudu.client.TestAlterTable [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.42 s <<< FAILURE! - in org.apache.kudu.client.TestAlterTable [ERROR] org.apache.kudu.client.TestAlterTable Time elapsed: 0.42 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 [INFO] Running org.apache.kudu.client.TestAsyncKuduClient [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.644 s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduClient [ERROR] org.apache.kudu.client.TestAsyncKuduClient Time elapsed: 0.644 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAsyncKuduClient.setUpBeforeClass(TestAsyncKuduClient.java:45) [INFO] Running org.apache.kudu.client.TestAsyncKuduSession [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.478 s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduSession [ERROR] org.apache.kudu.client.TestAsyncKuduSession Time elapsed: 0.478 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAsyncKuduSession.setUpBeforeClass(TestAsyncKuduSession.java:59) [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquire [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.601 s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquire [ERROR] org.apache.kudu.client.TestAuthnTokenReacquire Time elapsed: 0.601 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAuthnTokenReacquire.setUpBeforeClass(TestAuthnTokenReacquire.java:55) [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquireOpen [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.865 s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquireOpen [ERROR] org.apache.kudu.client.TestAuthnTokenReacquireOpen Time elapsed: 0.865 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to start a single Master: /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestAuthnTokenReacquireOpen.setUpBeforeClass(TestAuthnTokenReacquireOpen.java:58) [INFO] Running org.apache.kudu.client.TestBitSet [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.061 s - in org.apache.kudu.client.TestBitSet [INFO] Running org.apache.kudu.client.TestBytes [INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.077 s - in org.apache.kudu.client.TestBytes [INFO] Running org.apache.kudu.client.TestClientFailoverSupport [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.399 s <<< FAILURE! - in org.apache.kudu.client.TestClientFailoverSupport [ERROR] org.apache.kudu.client.TestClientFailoverSupport Time elapsed: 0.399 s <<< ERROR! org.apache.kudu.client.NonRecoverableException: Failed to add distributed masters: Unable to start Master at index 0: /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited on signal 6 at org.apache.kudu.client.TestClientFailoverSupport.setUpBeforeClass(TestClientFailoverSupport.java:35) [INFO] Running org.apache.kudu.client.TestColumnRangePredicate [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.106 s - in org.apache.kudu.client.TestColumnRangePredicate [INFO] Running org.apache.kudu.client.TestConnectionCache [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.43 s <<< FAILURE! - in org.apache.kudu.client.TestConnectionCache [ERROR] test(org.apache.kudu.client.TestConnectionCache) Time elapsed: 0.375 s <<<
[jira] [Updated] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0
[ https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nacho García Fernández updated KUDU-2223: - Affects Version/s: (was: 1.5.0) 1.6.0 > Failed to add distributed masters: Unable to start Master at index 0 > > > Key: KUDU-2223 > URL: https://issues.apache.org/jira/browse/KUDU-2223 > Project: Kudu > Issue Type: Bug > Components: build, java >Affects Versions: 1.6.0 >Reporter: Nacho García Fernández > > After successfully building Kudu on my OSX, I try to run the mvn verify > command in the java submodule, but I get the following exception: > {code:java} > [INFO] --- > [INFO] T E S T S > [INFO] --- > [INFO] Running org.apache.kudu.client.TestAlterTable > [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.42 > s <<< FAILURE! - in org.apache.kudu.client.TestAlterTable > [ERROR] org.apache.kudu.client.TestAlterTable Time elapsed: 0.42 s <<< > ERROR! > org.apache.kudu.client.NonRecoverableException: Failed to add distributed > masters: Unable to start Master at index 0: > /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited > on signal 6 > [INFO] Running org.apache.kudu.client.TestAsyncKuduClient > [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.644 > s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduClient > [ERROR] org.apache.kudu.client.TestAsyncKuduClient Time elapsed: 0.644 s > <<< ERROR! > org.apache.kudu.client.NonRecoverableException: Failed to add distributed > masters: Unable to start Master at index 0: > /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited > on signal 6 > at > org.apache.kudu.client.TestAsyncKuduClient.setUpBeforeClass(TestAsyncKuduClient.java:45) > [INFO] Running org.apache.kudu.client.TestAsyncKuduSession > [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.478 > s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduSession > [ERROR] org.apache.kudu.client.TestAsyncKuduSession Time elapsed: 0.478 s > <<< ERROR! > org.apache.kudu.client.NonRecoverableException: Failed to add distributed > masters: Unable to start Master at index 0: > /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited > on signal 6 > at > org.apache.kudu.client.TestAsyncKuduSession.setUpBeforeClass(TestAsyncKuduSession.java:59) > [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquire > [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.601 > s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquire > [ERROR] org.apache.kudu.client.TestAuthnTokenReacquire Time elapsed: 0.601 s > <<< ERROR! > org.apache.kudu.client.NonRecoverableException: Failed to add distributed > masters: Unable to start Master at index 0: > /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited > on signal 6 > at > org.apache.kudu.client.TestAuthnTokenReacquire.setUpBeforeClass(TestAuthnTokenReacquire.java:55) > [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquireOpen > [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.865 > s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquireOpen > [ERROR] org.apache.kudu.client.TestAuthnTokenReacquireOpen Time elapsed: > 0.865 s <<< ERROR! > org.apache.kudu.client.NonRecoverableException: Failed to start a single > Master: /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process > exited on signal 6 > at > org.apache.kudu.client.TestAuthnTokenReacquireOpen.setUpBeforeClass(TestAuthnTokenReacquireOpen.java:58) > [INFO] Running org.apache.kudu.client.TestBitSet > [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.061 > s - in org.apache.kudu.client.TestBitSet > [INFO] Running org.apache.kudu.client.TestBytes > [INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.077 > s - in org.apache.kudu.client.TestBytes > [INFO] Running org.apache.kudu.client.TestClientFailoverSupport > [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.399 > s <<< FAILURE! - in org.apache.kudu.client.TestClientFailoverSupport > [ERROR] org.apache.kudu.client.TestClientFailoverSupport Time elapsed: 0.399 > s <<<
[jira] [Commented] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0
[ https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16261314#comment-16261314 ] Nacho García Fernández commented on KUDU-2223: -- Example of surefire logs: {code:java} 19:58:04.892 [INFO - Time-limited test] (MiniKuduCluster.java:145) Starting process: [/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu, test, mini_cluster, --serialization=pb] 19:58:05.054 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: create_cluster { num_masters: 3 num_tservers: 0 enable_kerberos: false } 19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:123) Response: 19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: start_cluster { } 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 19:58:05.079746 3212440512 external_mini_cluster.cc:771] Running /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --fs_wal_dir=/tmp/kudutest-214522293/minicluster-data/master-0/wal 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/master-0/data 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --block_manager=file 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --webserver_interface=localhost 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --ipki_ca_key_size=1024 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --tsk_num_rsa_bits=512 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --rpc_bind_addresses=127.0.0.1:11030 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --webserver_interface=127.0.0.1 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --webserver_port=0 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --never_fsync 19:58:05.082 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --ipki_server_key_size=1024 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --enable_minidumps=false 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --redact=none 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --metrics_log_interval_ms=1000 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --logtostderr 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --logbuflevel=-1 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --log_dir=/tmp/kudutest-214522293/minicluster-data/master-0/logs 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --server_dump_info_path=/tmp/kudutest-214522293/minicluster-data/master-0/data/info.pb 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --server_dump_info_format=pb 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --rpc_server_allow_ephemeral_ports 19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --unlock_experimental_flags 19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --unlock_unsafe_flags 19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --master_addresses=127.0.0.1:11030,127.0.0.1:11031,127.0.0.1:11032 with env {} 19:58:05.095 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) WARNING: Logging before InitGoogleLogging() is written to STDERR 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.095039 3212440512 flags.cc:391] Enabled unsafe flag: --rpc_server_allow_ephemeral_ports=true 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.095845 3212440512 flags.cc:391] Enabled unsafe flag: --never_fsync=true 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.096652 3212440512 flags.cc:391] Enabled experimental flag: --ipki_ca_key_size=1024 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.096664 3212440512 flags.cc:391] Enabled experimental flag: --ipki_server_key_size=1024 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.096679 3212440512 flags.cc:391] Enabled experimental flag: --tsk_num_rsa_bits=512 19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 19:58:05.097530 3212440512 master_main.cc:66] Master server non-default flags: 19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/master-0/data 19:58:05.097 [INFO - cluster stderr pri
[jira] [Comment Edited] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0
[ https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16261314#comment-16261314 ] Nacho García Fernández edited comment on KUDU-2223 at 11/21/17 7:11 PM: Example of surefire logs: {code:java} 19:58:04.892 [INFO - Time-limited test] (MiniKuduCluster.java:145) Starting process: [/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu, test, mini_cluster, --serialization=pb] 19:58:05.054 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: create_cluster { num_masters: 3 num_tservers: 0 enable_kerberos: false } 19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:123) Response: 19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: start_cluster { } 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 19:58:05.079746 3212440512 external_mini_cluster.cc:771] Running /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --fs_wal_dir=/tmp/kudutest-214522293/minicluster-data/master-0/wal 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/master-0/data 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --block_manager=file 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --webserver_interface=localhost 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --ipki_ca_key_size=1024 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --tsk_num_rsa_bits=512 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --rpc_bind_addresses=127.0.0.1:11030 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --webserver_interface=127.0.0.1 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --webserver_port=0 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --never_fsync 19:58:05.082 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --ipki_server_key_size=1024 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --enable_minidumps=false 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --redact=none 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --metrics_log_interval_ms=1000 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --logtostderr 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --logbuflevel=-1 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --log_dir=/tmp/kudutest-214522293/minicluster-data/master-0/logs 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --server_dump_info_path=/tmp/kudutest-214522293/minicluster-data/master-0/data/info.pb 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --server_dump_info_format=pb 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --rpc_server_allow_ephemeral_ports 19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --unlock_experimental_flags 19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --unlock_unsafe_flags 19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --master_addresses=127.0.0.1:11030,127.0.0.1:11031,127.0.0.1:11032 with env {} 19:58:05.095 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) WARNING: Logging before InitGoogleLogging() is written to STDERR 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.095039 3212440512 flags.cc:391] Enabled unsafe flag: --rpc_server_allow_ephemeral_ports=true 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.095845 3212440512 flags.cc:391] Enabled unsafe flag: --never_fsync=true 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.096652 3212440512 flags.cc:391] Enabled experimental flag: --ipki_ca_key_size=1024 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.096664 3212440512 flags.cc:391] Enabled experimental flag: --ipki_server_key_size=1024 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.096679 3212440512 flags.cc:391] Enabled experimental flag: --tsk_num_rsa_bits=512 19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 19:58:05.097530 3212440512 master_main.cc:66] Master server non-default flags: 19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/mast
[jira] [Comment Edited] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0
[ https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16261314#comment-16261314 ] Nacho García Fernández edited comment on KUDU-2223 at 11/21/17 7:12 PM: Example of surefire logs: {code:java} 19:58:04.892 [INFO - Time-limited test] (MiniKuduCluster.java:145) Starting process: [/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu, test, mini_cluster, --serialization=pb] 19:58:05.054 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: create_cluster { num_masters: 3 num_tservers: 0 enable_kerberos: false } 19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:123) Response: 19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: start_cluster { } 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 19:58:05.079746 3212440512 external_mini_cluster.cc:771] Running /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --fs_wal_dir=/tmp/kudutest-214522293/minicluster-data/master-0/wal 19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/master-0/data 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --block_manager=file 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --webserver_interface=localhost 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --ipki_ca_key_size=1024 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --tsk_num_rsa_bits=512 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --rpc_bind_addresses=127.0.0.1:11030 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --webserver_interface=127.0.0.1 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --webserver_port=0 19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --never_fsync 19:58:05.082 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --ipki_server_key_size=1024 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --enable_minidumps=false 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --redact=none 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --metrics_log_interval_ms=1000 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --logtostderr 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --logbuflevel=-1 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --log_dir=/tmp/kudutest-214522293/minicluster-data/master-0/logs 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --server_dump_info_path=/tmp/kudutest-214522293/minicluster-data/master-0/data/info.pb 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --server_dump_info_format=pb 19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --rpc_server_allow_ephemeral_ports 19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --unlock_experimental_flags 19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --unlock_unsafe_flags 19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --master_addresses=127.0.0.1:11030,127.0.0.1:11031,127.0.0.1:11032 with env {} 19:58:05.095 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) WARNING: Logging before InitGoogleLogging() is written to STDERR 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.095039 3212440512 flags.cc:391] Enabled unsafe flag: --rpc_server_allow_ephemeral_ports=true 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.095845 3212440512 flags.cc:391] Enabled unsafe flag: --never_fsync=true 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.096652 3212440512 flags.cc:391] Enabled experimental flag: --ipki_ca_key_size=1024 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.096664 3212440512 flags.cc:391] Enabled experimental flag: --ipki_server_key_size=1024 19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 19:58:05.096679 3212440512 flags.cc:391] Enabled experimental flag: --tsk_num_rsa_bits=512 19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 19:58:05.097530 3212440512 master_main.cc:66] Master server non-default flags: 19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) --fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/mast
[jira] [Commented] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0
[ https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16262426#comment-16262426 ] Nacho García Fernández commented on KUDU-2223: -- I've just run the tests one more time after removing the /tmp/master directory. However, some of them are still failing with the same exception: {code:java} 0xNacho ~/dev/github/kudu/java master ● mvn test [INFO] Scanning for projects... [WARNING] [WARNING] Some problems were encountered while building the effective model for org.apache.kudu:kudu-spark2-tools_2.11:jar:1.6.0-SNAPSHOT [WARNING] 'artifactId' contains an expression but should be a constant. @ org.apache.kudu:kudu-${spark.version.label}-tools_${scala.binary.version}:[unknown-version], /Users/0xNacho/dev/github/kudu/java/kudu-spark-tools/pom.xml, line 30, column 17 [WARNING] [WARNING] Some problems were encountered while building the effective model for org.apache.kudu:kudu-spark2_2.11:jar:1.6.0-SNAPSHOT [WARNING] 'artifactId' contains an expression but should be a constant. @ org.apache.kudu:kudu-${spark.version.label}_${scala.binary.version}:[unknown-version], /Users/0xNacho/dev/github/kudu/java/kudu-spark/pom.xml, line 23, column 17 [WARNING] [WARNING] It is highly recommended to fix these problems because they threaten the stability of your build. [WARNING] [WARNING] For this reason, future Maven versions might no longer support building such malformed projects. [WARNING] [INFO] [INFO] Detecting the operating system and CPU architecture [INFO] [INFO] os.detected.name: osx [INFO] os.detected.arch: x86_64 [INFO] os.detected.version: 10.12 [INFO] os.detected.version.major: 10 [INFO] os.detected.version.minor: 12 [INFO] os.detected.classifier: osx-x86_64 [INFO] [INFO] Reactor Build Order: [INFO] [INFO] Kudu [INFO] Kudu Java Client [INFO] Kudu MapReduce bindings [INFO] Kudu Client Tools [INFO] Kudu Flume NG Sink [INFO] Kudu Hive [INFO] Kudu Spark Bindings [INFO] Kudu Spark Tools [INFO] Kudu Flink bindings [INFO] [INFO] [INFO] Building Kudu 1.6.0-SNAPSHOT [INFO] [INFO] [INFO] --- maven-enforcer-plugin:1.4.1:enforce (default) @ kudu-parent --- [INFO] [INFO] --- maven-checkstyle-plugin:2.17:check (validate) @ kudu-parent --- [INFO] Starting audit... Audit done. [INFO] [INFO] --- maven-remote-resources-plugin:1.5:process (process-resource-bundles) @ kudu-parent --- [INFO] [INFO] --- animal-sniffer-maven-plugin:1.16:check (check-java7-apis) @ kudu-parent --- [INFO] Checking unresolved references to org.codehaus.mojo.signature:java17:1.0 [INFO] [INFO] --- maven-antrun-plugin:1.8:run (default) @ kudu-parent --- [INFO] Executing tasks main: [mkdir] Created dir: /Users/0xNacho/dev/github/kudu/java/target/testdata [INFO] Executed tasks [INFO] [INFO] [INFO] Building Kudu Java Client 1.6.0-SNAPSHOT [INFO] [INFO] [INFO] --- maven-checkstyle-plugin:2.17:check (validate) @ kudu-client --- [INFO] Starting audit... [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:490: 'method def modifier' have incorrect indentation level 6, expected level should be one of the following: 8, 10, 12. [Indentation] [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:492: 'method def' child have incorrect indentation level 8, expected level should be one of the following: 10, 12, 14. [Indentation] [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:494: 'method def' child have incorrect indentation level 8, expected level should be one of the following: 10, 12, 14. [Indentation] [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:495: 'method def rcurly' have incorrect indentation level 6, expected level should be one of the following: 8, 10, 12. [Indentation] [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:496: 'object def rcurly' have incorrect indentation level 4, expected level should be one of the following: 6, 8, 10. [Indentation] [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:1103: Line is longer than 100 characters (f
[jira] [Comment Edited] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0
[ https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16262426#comment-16262426 ] Nacho García Fernández edited comment on KUDU-2223 at 11/22/17 12:23 PM: - I've just run the tests one more time after removing the /tmp/master directory. However, some of them are still failing with the same exception: {code:java} 0xNacho ~/dev/github/kudu/java master ● mvn test [INFO] Scanning for projects... [WARNING] [WARNING] Some problems were encountered while building the effective model for org.apache.kudu:kudu-spark2-tools_2.11:jar:1.6.0-SNAPSHOT [WARNING] 'artifactId' contains an expression but should be a constant. @ org.apache.kudu:kudu-${spark.version.label}-tools_${scala.binary.version}:[unknown-version], /Users/0xNacho/dev/github/kudu/java/kudu-spark-tools/pom.xml, line 30, column 17 [WARNING] [WARNING] Some problems were encountered while building the effective model for org.apache.kudu:kudu-spark2_2.11:jar:1.6.0-SNAPSHOT [WARNING] 'artifactId' contains an expression but should be a constant. @ org.apache.kudu:kudu-${spark.version.label}_${scala.binary.version}:[unknown-version], /Users/0xNacho/dev/github/kudu/java/kudu-spark/pom.xml, line 23, column 17 [WARNING] [WARNING] It is highly recommended to fix these problems because they threaten the stability of your build. [WARNING] [WARNING] For this reason, future Maven versions might no longer support building such malformed projects. [WARNING] [INFO] [INFO] Detecting the operating system and CPU architecture [INFO] [INFO] os.detected.name: osx [INFO] os.detected.arch: x86_64 [INFO] os.detected.version: 10.12 [INFO] os.detected.version.major: 10 [INFO] os.detected.version.minor: 12 [INFO] os.detected.classifier: osx-x86_64 [INFO] [INFO] Reactor Build Order: [INFO] [INFO] Kudu [INFO] Kudu Java Client [INFO] Kudu MapReduce bindings [INFO] Kudu Client Tools [INFO] Kudu Flume NG Sink [INFO] Kudu Hive [INFO] Kudu Spark Bindings [INFO] Kudu Spark Tools [INFO] Kudu Flink bindings [INFO] [INFO] [INFO] Building Kudu 1.6.0-SNAPSHOT [INFO] [INFO] [INFO] --- maven-enforcer-plugin:1.4.1:enforce (default) @ kudu-parent --- [INFO] [INFO] --- maven-checkstyle-plugin:2.17:check (validate) @ kudu-parent --- [INFO] Starting audit... Audit done. [INFO] [INFO] --- maven-remote-resources-plugin:1.5:process (process-resource-bundles) @ kudu-parent --- [INFO] [INFO] --- animal-sniffer-maven-plugin:1.16:check (check-java7-apis) @ kudu-parent --- [INFO] Checking unresolved references to org.codehaus.mojo.signature:java17:1.0 [INFO] [INFO] --- maven-antrun-plugin:1.8:run (default) @ kudu-parent --- [INFO] Executing tasks main: [mkdir] Created dir: /Users/0xNacho/dev/github/kudu/java/target/testdata [INFO] Executed tasks [INFO] [INFO] [INFO] Building Kudu Java Client 1.6.0-SNAPSHOT [INFO] [INFO] [INFO] --- maven-checkstyle-plugin:2.17:check (validate) @ kudu-client --- [INFO] Starting audit... [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:490: 'method def modifier' have incorrect indentation level 6, expected level should be one of the following: 8, 10, 12. [Indentation] [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:492: 'method def' child have incorrect indentation level 8, expected level should be one of the following: 10, 12, 14. [Indentation] [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:494: 'method def' child have incorrect indentation level 8, expected level should be one of the following: 10, 12, 14. [Indentation] [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:495: 'method def rcurly' have incorrect indentation level 6, expected level should be one of the following: 8, 10, 12. [Indentation] [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:496: 'object def rcurly' have incorrect indentation level 4, expected level should be one of the following: 6, 8, 10. [Indentation] [WARN] /Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKud
[jira] [Created] (KUDU-2329) Random RPC timeout errors when inserting rows in a Kudu table
Héctor Gutiérrez created KUDU-2329: -- Summary: Random RPC timeout errors when inserting rows in a Kudu table Key: KUDU-2329 URL: https://issues.apache.org/jira/browse/KUDU-2329 Project: Kudu Issue Type: Bug Components: rpc, server Affects Versions: 1.5.0 Reporter: Héctor Gutiérrez When executing inserts into a Kudu table, we are experiencing errors at random times. The first time we found one of these errors was during a bulk update of a Kudu table via Spark (in Scala): {{kuduContext.updateRows(dataFrame, "table_name")}} The error message in Spark was the following: {{java.lang.RuntimeException: failed to write 579 rows from DataFrame to Kudu; sample errors: Timed out: can not complete before timeout: Batch{operations=6, tablet="cd1e33fce0114c9bbd9c14e2559e7591" [0x000F, 0x0010), ignoreAllDuplicateRows=false, rpc=KuduRpc(method=Write, tablet=cd1e33fce0114c9bbd9c14e2559e7591, attempt=3, DeadlineTracker(timeout=3, elapsed=30090), Traces: [0ms] sending RPC to server 6f273933b4d5498e87aadfb99b054a21, [10011ms] received from server 6f273933b4d5498e87aadfb99b054a21 response Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel, [10011ms] delaying RPC due to Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel, [10033ms] sending RPC to server 6f273933b4d5498e87aadfb99b054a21, [20050ms] received from server 6f273933b4d5498e87aadfb99b054a21 response Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel, [20050ms] delaying RPC due to Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel, [20072ms] sending RPC to server 6f273933b4d5498e87aadfb99b054a21, [30090ms] received from server 6f273933b4d5498e87aadfb99b054a21 response Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel, [30090ms] delaying RPC due to Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel)}}} (+ 4 more errors similar to this one in the error message) We first thought it was actually a problem with our Spark code, but when we tried to execute a simple "INSERT INTO" query from the impala shell into a Kudu table, we got the following error: {{[.] > insert into test_kudu values (282, 'hola');}} {{ Query: insert into test_kudu values (282, 'hola')}} {{ Query submitted at: ..}} {{ Query progress can be monitored at: }} {{ WARNINGS: Kudu error(s) reported, first error: Timed out: Failed to write batch of 1 ops to tablet 9c295e90811e483a9550bfd75abcf666 after 1 attempt(s): Failed to write to server: 071bcafbb1644678a697c474662047b7 (.:7050): Write RPC to :7050 timed out after 179.949s (SENT)}} {{Error in Kudu table 'impala:kudu_db.test_kudu': Timed out: Failed to write batch of 1 ops to tablet 9c295e90811e483a9550bfd75abcf666 after 1 attempt(s): Failed to write to server: 071bcafbb1644678a697c474662047b7 (...:7050): Write RPC to ..:7050 timed out after 179.949s (SENT)}} To make things even more confusing, despite getting this error in the impala shell, after a while (and not immediately), the inserted rows ended up in the table, so somehow they were actually inserted. We also tried tweaking the Kudu timeout configuration values that we had previously set, but it didn't solve anything and the problem kept appearing. Furthermore, we don't always get these errors, they only appear at random times. For example, right now we're just getting errors in that update we have in the Spark code, but we are not experiencing issues when working from the impala shell. After all that we have tried, we are pretty certain that this is a bug in Kudu, although we think it is a bit strange that it is undocumented and certainly it's hard to reproduce. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (KUDU-2329) Random RPC timeout errors when inserting rows in a Kudu table
[ https://issues.apache.org/jira/browse/KUDU-2329?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Héctor Gutiérrez updated KUDU-2329: --- Description: When executing inserts into a Kudu table, we are experiencing errors at random times. The first time we found one of these errors was during a bulk update of a Kudu table via Spark (in Scala): {{kuduContext.updateRows(dataFrame, "table_name")}} The error message in Spark was the following: {{java.lang.RuntimeException: failed to write 579 rows from DataFrame to Kudu; sample errors: Timed out: can not complete before timeout: Batch {operations=6, tablet="cd1e33fce0114c9bbd9c14e2559e7591" [0x000F, 0x0010), ignoreAllDuplicateRows=false, rpc=KuduRpc(method=Write, tablet=cd1e33fce0114c9bbd9c14e2559e7591, attempt=3, DeadlineTracker(timeout=3, elapsed=30090), Traces: [0ms] sending RPC to server 6f273933b4d5498e87aadfb99b054a21, [10011ms] received from server 6f273933b4d5498e87aadfb99b054a21 response Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel, [10011ms] delaying RPC due to Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel, [10033ms] sending RPC to server 6f273933b4d5498e87aadfb99b054a21, [20050ms] received from server 6f273933b4d5498e87aadfb99b054a21 response Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel, [20050ms] delaying RPC due to Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel, [20072ms] sending RPC to server 6f273933b4d5498e87aadfb99b054a21, [30090ms] received from server 6f273933b4d5498e87aadfb99b054a21 response Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel, [30090ms] delaying RPC due to Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the channel)} }} (+ 4 more errors similar to this one in the error message) We first thought it was actually a problem with our Spark code, but when we tried to execute a simple "INSERT INTO" query from the impala shell into a Kudu table, we got the following error: {{[.] > insert into test_kudu values (282, 'hola');}} {{ \{{ Query: insert into test_kudu values (282, 'hola') {{ \{{ Query submitted at: .. {{ \{{ Query progress can be monitored at: {{ \{{ WARNINGS: Kudu error(s) reported, first error: Timed out: Failed to write batch of 1 ops to tablet 9c295e90811e483a9550bfd75abcf666 after 1 attempt(s): Failed to write to server: 071bcafbb1644678a697c474662047b7 (.:7050): Write RPC to :7050 timed out after 179.949s (SENT) {{Error in Kudu table 'impala:kudu_db.test_kudu': Timed out: Failed to write batch of 1 ops to tablet 9c295e90811e483a9550bfd75abcf666 after 1 attempt(s): Failed to write to server: 071bcafbb1644678a697c474662047b7 (...:7050): Write RPC to ..:7050 timed out after 179.949s (SENT)}} To make things even more confusing, despite getting this error in the impala shell, after a while (and not immediately), the inserted rows ended up in the table, so somehow they were actually inserted. We also tried tweaking the Kudu timeout configuration values that we had previously set, but it didn't solve anything and the problem kept appearing. Furthermore, we don't always get these errors, they only appear at random times. For example, right now we're just getting errors in that update we have in the Spark code, but we are not experiencing issues when working from the impala shell. After all that we have tried, we are pretty certain that this is a bug in Kudu, although we think it is a bit strange that it is undocumented and certainly it's hard to reproduce. was: When executing inserts into a Kudu table, we are experiencing errors at random times. The first time we found one of these errors was during a bulk update of a Kudu table via Spark (in Scala): {{kuduContext.updateRows(dataFrame, "table_name")}} The error message in Spark was the following: {{java.lang.RuntimeException: failed to write 579 rows from DataFrame to Kudu; sample errors: Timed out: can not complete before timeout: Batch{operations=6, tablet="cd1e33fce0114c9bbd9c14e2559e7591" [0x000F, 0x0010), ignoreAllDuplicateRows=false, rpc=KuduRpc(method=Write, tablet=cd1e33fce0114c9bbd9c14e2559e7591, attempt=3, DeadlineTracker(timeout=3, elapsed=30090), Traces: [0ms] sending RPC to server 6f273933b4d5498e87aadfb99b054a21, [10011ms] received from server 6f273933b4d5498e87aadfb99b054a21 response Network error: [peer 6f27393
[jira] [Created] (KUDU-2332) Error when importing kudu package in Python Anaconda distribution
Michał Sznajder created KUDU-2332: - Summary: Error when importing kudu package in Python Anaconda distribution Key: KUDU-2332 URL: https://issues.apache.org/jira/browse/KUDU-2332 Project: Kudu Issue Type: Bug Components: python Affects Versions: 1.2.0 Environment: ProductName:Mac OS X ProductVersion:10.13.3 BuildVersion:17D102 Anaconda Python distribution Reporter: Michał Sznajder Attachments: error.png I tried to install Kudu on my local machine: ProductName: Mac OS X ProductVersion: 10.13.3 BuildVersion: 17D102 I followed all the steps to build Kudu 1.6 from source: 1. brew install autoconf automake cmake libtool pkg-config pstree 2. git clone https://github.com/apache/incubator-kudu kudu 2. cd kudu 3. PKG_CONFIG_PATH=/usr/local/Cellar/openssl/1.0.2n/lib/pkgconfig thirdparty/build-if-necessary.sh 4. mkdir -p build/release 5. cd build/release 6. PKG_CONFIG_PATH=/usr/local/Cellar/openssl/1.0.2n/lib/pkgconfig ../../thirdparty/installed/common/bin/cmake \ -DCMAKE_BUILD_TYPE=release \ -DOPENSSL_ROOT_DIR=/usr/local/opt/openssl \ ../.. 7. make -j4 8. sudo make install This resulted with following libraries installed: /usr/local/include/kudu /usr/local/include/kudu/util/kudu_export.h /usr/local/lib/libkudu_client.0.1.0.dylib /usr/local/lib/libkudu_client.dylib /usr/local/lib/libkudu_client.0.dylib /usr/local/share/kuduClient /usr/local/share/kuduClient/cmake/kuduClientTargets.cmake /usr/local/share/kuduClient/cmake/kuduClientTargets-release.cmake /usr/local/share/kuduClient/cmake/kuduClientConfig.cmake /usr/local/share/doc/kuduClient Then I followed steps to instal kudu-python package using pip: 1. clean pip cache to make sure it is clean 2. pip install -v kudu-python Then after calling: import kudu I got error like in the attachment "error.png". As first line of this screen states it was Anaconda Python distribution. After removing Anaconda Python and installing Python using Homebrew and again following above steps - all worked. My conclusion: there is some kind of issue happening between Anaconda Python and kudu-python package. Some more details are also in [https://getkudu.slack.com|https://getkudu.slack.com/] slack channel on #kudu-general channel. I am reachable there under msznajder nickname. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (KUDU-2487) kudu quote escape issues in select query
[ https://issues.apache.org/jira/browse/KUDU-2487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16733145#comment-16733145 ] Tiago Simões commented on KUDU-2487: Isn't this an Impala issue? > kudu quote escape issues in select query > > > Key: KUDU-2487 > URL: https://issues.apache.org/jira/browse/KUDU-2487 > Project: Kudu > Issue Type: Bug > Components: impala >Affects Versions: 1.7.0 > Environment: kudu-1.7.0 > cdh5.15.0 > impala-2.12.0 >Reporter: fangwan >Priority: Critical > > From impala shell, I am getting wrong results by using escape single quote > for query, details explained below: > 1. create a kudu table like: > {code:java} > CREATE TABLE test_cf_escape( > x INT NOT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION, > y STRING NOT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION, > z INT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION, > PRIMARY KEY(x, y) > ) > STORED AS KUDU > TBLPROPERTIES ('kudu.master_addresses'='hadoop-nn.xxx.com'){code} > 2. insert data to the table: > {code:java} > insert into test_cf_escape (x,y,z) values (1, "chili\\\'s", 0); > insert into test_cf_escape (x,y,z) values (1, "chili\'s", 2); > select * from test_cf_escape; > => > +---+--+---+ > | x | y| z | > +---+--+---+ > | 1 | chili's | 2 | > | 1 | chili\'s | 0 | > +---+--+---+ > {code} > 3. we should get the same row by using below two select query, but failed: > {code:java} > select * from test_cf_escape where y = 'chili\'s'; => (1, 'chili\'s', 0) > (supposed to read data as (1, "chili's", 2)) > select * from test_cf_escape where y = "chili's"; => (1, "chili's", 2) > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Comment Edited] (KUDU-2487) kudu quote escape issues in select query
[ https://issues.apache.org/jira/browse/KUDU-2487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16733145#comment-16733145 ] Tiago Simões edited comment on KUDU-2487 at 1/3/19 3:18 PM: Isn't this an Impala issue? Have you tried to do the same using a parquet file? was (Author: tiagorodriguessimoes): Isn't this an Impala issue? > kudu quote escape issues in select query > > > Key: KUDU-2487 > URL: https://issues.apache.org/jira/browse/KUDU-2487 > Project: Kudu > Issue Type: Bug > Components: impala >Affects Versions: 1.7.0 > Environment: kudu-1.7.0 > cdh5.15.0 > impala-2.12.0 >Reporter: fangwan >Priority: Critical > > From impala shell, I am getting wrong results by using escape single quote > for query, details explained below: > 1. create a kudu table like: > {code:java} > CREATE TABLE test_cf_escape( > x INT NOT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION, > y STRING NOT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION, > z INT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION, > PRIMARY KEY(x, y) > ) > STORED AS KUDU > TBLPROPERTIES ('kudu.master_addresses'='hadoop-nn.xxx.com'){code} > 2. insert data to the table: > {code:java} > insert into test_cf_escape (x,y,z) values (1, "chili\\\'s", 0); > insert into test_cf_escape (x,y,z) values (1, "chili\'s", 2); > select * from test_cf_escape; > => > +---+--+---+ > | x | y| z | > +---+--+---+ > | 1 | chili's | 2 | > | 1 | chili\'s | 0 | > +---+--+---+ > {code} > 3. we should get the same row by using below two select query, but failed: > {code:java} > select * from test_cf_escape where y = 'chili\'s'; => (1, 'chili\'s', 0) > (supposed to read data as (1, "chili's", 2)) > select * from test_cf_escape where y = "chili's"; => (1, "chili's", 2) > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (KUDU-3588) SecurityITest.TestNonDefaultPrincipalMultipleMaster flakiness
Bakai Ádám created KUDU-3588: Summary: SecurityITest.TestNonDefaultPrincipalMultipleMaster flakiness Key: KUDU-3588 URL: https://issues.apache.org/jira/browse/KUDU-3588 Project: Kudu Issue Type: Bug Reporter: Bakai Ádám -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Closed] (KUDU-3588) SecurityITest.TestNonDefaultPrincipalMultipleMaster flakiness
[ https://issues.apache.org/jira/browse/KUDU-3588?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám closed KUDU-3588. Resolution: Duplicate > SecurityITest.TestNonDefaultPrincipalMultipleMaster flakiness > - > > Key: KUDU-3588 > URL: https://issues.apache.org/jira/browse/KUDU-3588 > Project: Kudu > Issue Type: Bug >Reporter: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3589) SecurityITest.TestNonDefaultPrincipalMultipleMaster is flaky
Bakai Ádám created KUDU-3589: Summary: SecurityITest.TestNonDefaultPrincipalMultipleMaster is flaky Key: KUDU-3589 URL: https://issues.apache.org/jira/browse/KUDU-3589 Project: Kudu Issue Type: Sub-task Reporter: Bakai Ádám -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3590) update certs in test_certs.cc
Bakai Ádám created KUDU-3590: Summary: update certs in test_certs.cc Key: KUDU-3590 URL: https://issues.apache.org/jira/browse/KUDU-3590 Project: Kudu Issue Type: Sub-task Reporter: Bakai Ádám -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (KUDU-3590) update certs in test_certs.cc
[ https://issues.apache.org/jira/browse/KUDU-3590?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3590: Assignee: Bakai Ádám > update certs in test_certs.cc > - > > Key: KUDU-3590 > URL: https://issues.apache.org/jira/browse/KUDU-3590 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (KUDU-3589) SecurityITest.TestNonDefaultPrincipalMultipleMaster is flaky
[ https://issues.apache.org/jira/browse/KUDU-3589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3589: Assignee: Bakai Ádám > SecurityITest.TestNonDefaultPrincipalMultipleMaster is flaky > > > Key: KUDU-3589 > URL: https://issues.apache.org/jira/browse/KUDU-3589 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3605) org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally is flaky
Bakai Ádám created KUDU-3605: Summary: org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally is flaky Key: KUDU-3605 URL: https://issues.apache.org/jira/browse/KUDU-3605 Project: Kudu Issue Type: Sub-task Reporter: Bakai Ádám -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (KUDU-3605) org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally is flaky
[ https://issues.apache.org/jira/browse/KUDU-3605?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3605: Assignee: Bakai Ádám > org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally > is flaky > - > > Key: KUDU-3605 > URL: https://issues.apache.org/jira/browse/KUDU-3605 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (KUDU-3605) org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally is flaky
[ https://issues.apache.org/jira/browse/KUDU-3605?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17872068#comment-17872068 ] Bakai Ádám commented on KUDU-3605: -- Found the probable cause of issue, which is missing data on the disk for syscatalog: {code:java} 13:14:38.568 [INFO - cluster stderr printer] (MiniKuduCluster.java:751) E20240808 13:14:38.568001 1477 master.cc:392] Unable to init master catalog manager: Not found: Unable to initialize catalog manager: Failed to initialize sys tables async: Unable to load consensus metadata for tablet : Unable to load consensus metadata for tablet : /tmp/dist-test-task6KabZO/test-tmp/mini-kudu-cluster692925529015678/master-1/wal/consensus-meta/: No such file or directory (error 2) 13:14:38.568 [INFO - cluster stderr printer] (MiniKuduCluster.java:751) I20240808 13:14:38.568156 1409 master.cc:537] Master@127.1.11.125:39515 shutting down... 13:14:38.585 [INFO - cluster stderr printer] (MiniKuduCluster.java:751) I20240808 13:14:38.585610 1409 master.cc:552] Master@127.1.11.125:39515 shutdown complete. {code} > org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally > is flaky > - > > Key: KUDU-3605 > URL: https://issues.apache.org/jira/browse/KUDU-3605 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (KUDU-3589) SecurityITest.TestNonDefaultPrincipalMultipleMaster is flaky
[ https://issues.apache.org/jira/browse/KUDU-3589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám resolved KUDU-3589. -- Fix Version/s: n/a Resolution: Fixed > SecurityITest.TestNonDefaultPrincipalMultipleMaster is flaky > > > Key: KUDU-3589 > URL: https://issues.apache.org/jira/browse/KUDU-3589 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > Fix For: n/a > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (KUDU-3605) org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally is flaky
[ https://issues.apache.org/jira/browse/KUDU-3605?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám resolved KUDU-3605. -- Fix Version/s: n/a Resolution: Fixed > org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally > is flaky > - > > Key: KUDU-3605 > URL: https://issues.apache.org/jira/browse/KUDU-3605 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > Fix For: n/a > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (KUDU-3624) EchoSubprocessTest.TestSubprocessMetricsOnError is flaky
[ https://issues.apache.org/jira/browse/KUDU-3624?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3624: Assignee: Bakai Ádám > EchoSubprocessTest.TestSubprocessMetricsOnError is flaky > > > Key: KUDU-3624 > URL: https://issues.apache.org/jira/browse/KUDU-3624 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > > {code:java} > I20241104 14:09:34.411099 543202 server.cc:273] Received an EOF from the > subprocess > W20241104 14:09:34.411275 543165 server.cc:408] The subprocess has exited > with status 9 > I20241104 14:09:34.417060 543203 server.cc:440] outbound queue shut down: > Aborted: > I20241104 14:09:34.417075 543200 server.cc:366] get failed, inbound queue > shut down: Aborted: > I20241104 14:09:34.417109 543201 server.cc:366] get failed, inbound queue > shut down: Aborted: > I20241104 14:09:34.417068 543199 server.cc:366] get failed, inbound queue > shut down: Aborted: > I20241104 14:09:37.790342 543244 server.cc:273] Received an EOF from the > subprocess > F20241104 14:09:37.790630 543207 server.cc:401] Check failed: _s.ok() Bad > status: Runtime error: Unable to wait on child: No child processes (error 10) > *** Check failure stack trace: *** > I20241104 14:09:37.790678 543242 server.cc:366] get failed, inbound queue > shut down: Aborted: > I20241104 14:09:37.790673 543243 server.cc:366] get failed, inbound queue > shut down: Aborted: > *** Aborted at 1730729377 (unix time) try "date -d @1730729377" if you are > using GNU date *** > I20241104 14:09:37.790684 543241 server.cc:366] get failed, inbound queue > shut down: Aborted: > I20241104 14:09:37.790699 543245 server.cc:440] outbound queue shut down: > Aborted: > PC: @ 0x0 (unknown) > *** SIGABRT (@0x848a9) received by PID 542889 (TID 0x704a47600700) from PID > 542889; stack trace: *** > @ 0x704a4d478980 (unknown) > @ 0x704a4d0b3e87 gsignal > @ 0x704a4d0b57f1 abort > @ 0x704a4e171d8d google::LogMessage::Fail() > @ 0x704a4e175b53 google::LogMessage::SendToLog() > @ 0x704a4e17178c google::LogMessage::Flush() > @ 0x704a4e172f19 google::LogMessageFatal::~LogMessageFatal() > @ 0x704a4f855a7e > kudu::subprocess::SubprocessServer::ExitCheckerThread() > @ 0x704a4f8529bd > _ZZN4kudu10subprocess16SubprocessServer4InitEvENKUlvE0_clEv > @ 0x704a4f8568ab > _ZNSt17_Function_handlerIFvvEZN4kudu10subprocess16SubprocessServer4InitEvEUlvE0_E9_M_invokeERKSt9_Any_data > @ 0x704a4f8cb06a std::function<>::operator()() > @ 0x704a4eff682b kudu::Thread::SuperviseThread() > @ 0x704a4d46d6db start_thread > @ 0x704a4d19661f clone > zsh: abort (core dumped) ./bin/subprocess_proxy-test --gtest_repeat=999 > --gtest_break_on_failure {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3624) EchoSubprocessTest.TestSubprocessMetricsOnError is flaky
Bakai Ádám created KUDU-3624: Summary: EchoSubprocessTest.TestSubprocessMetricsOnError is flaky Key: KUDU-3624 URL: https://issues.apache.org/jira/browse/KUDU-3624 Project: Kudu Issue Type: Sub-task Reporter: Bakai Ádám {code:java} I20241104 14:09:34.411099 543202 server.cc:273] Received an EOF from the subprocess W20241104 14:09:34.411275 543165 server.cc:408] The subprocess has exited with status 9 I20241104 14:09:34.417060 543203 server.cc:440] outbound queue shut down: Aborted: I20241104 14:09:34.417075 543200 server.cc:366] get failed, inbound queue shut down: Aborted: I20241104 14:09:34.417109 543201 server.cc:366] get failed, inbound queue shut down: Aborted: I20241104 14:09:34.417068 543199 server.cc:366] get failed, inbound queue shut down: Aborted: I20241104 14:09:37.790342 543244 server.cc:273] Received an EOF from the subprocess F20241104 14:09:37.790630 543207 server.cc:401] Check failed: _s.ok() Bad status: Runtime error: Unable to wait on child: No child processes (error 10) *** Check failure stack trace: *** I20241104 14:09:37.790678 543242 server.cc:366] get failed, inbound queue shut down: Aborted: I20241104 14:09:37.790673 543243 server.cc:366] get failed, inbound queue shut down: Aborted: *** Aborted at 1730729377 (unix time) try "date -d @1730729377" if you are using GNU date *** I20241104 14:09:37.790684 543241 server.cc:366] get failed, inbound queue shut down: Aborted: I20241104 14:09:37.790699 543245 server.cc:440] outbound queue shut down: Aborted: PC: @ 0x0 (unknown) *** SIGABRT (@0x848a9) received by PID 542889 (TID 0x704a47600700) from PID 542889; stack trace: *** @ 0x704a4d478980 (unknown) @ 0x704a4d0b3e87 gsignal @ 0x704a4d0b57f1 abort @ 0x704a4e171d8d google::LogMessage::Fail() @ 0x704a4e175b53 google::LogMessage::SendToLog() @ 0x704a4e17178c google::LogMessage::Flush() @ 0x704a4e172f19 google::LogMessageFatal::~LogMessageFatal() @ 0x704a4f855a7e kudu::subprocess::SubprocessServer::ExitCheckerThread() @ 0x704a4f8529bd _ZZN4kudu10subprocess16SubprocessServer4InitEvENKUlvE0_clEv @ 0x704a4f8568ab _ZNSt17_Function_handlerIFvvEZN4kudu10subprocess16SubprocessServer4InitEvEUlvE0_E9_M_invokeERKSt9_Any_data @ 0x704a4f8cb06a std::function<>::operator()() @ 0x704a4eff682b kudu::Thread::SuperviseThread() @ 0x704a4d46d6db start_thread @ 0x704a4d19661f clone zsh: abort (core dumped) ./bin/subprocess_proxy-test --gtest_repeat=999 --gtest_break_on_failure {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (KUDU-3605) org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally is flaky
[ https://issues.apache.org/jira/browse/KUDU-3605?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17884277#comment-17884277 ] Bakai Ádám commented on KUDU-3605: -- I had to recreate my dev environment, and the same test now fails with a different error: {code:java} There was 1 failure: 1) testExternallyProvidedSubjectRefreshedExternally(org.apache.kudu.client.TestSecurity) org.apache.kudu.client.NonRecoverableException: cannot complete before timeout: KuduRpc(method=ListTabletServers, tablet=null, attempt=24, TimeoutTracker(timeout=3, elapsed=28366), Traces: [0ms] refreshing cache from master, [30ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.62:46867, [105ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.60:40051, [107ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.61:46759, [144ms] Sub RPC ConnectToMaster: received response from server master-127.1.11.62:46867: Network error: Failed to connect to peer master-127.1.11.62:46867(127.1.11.62:46867): Connection refused: /127.1.11.62:46867, [148ms] Sub RPC ConnectToMaster: received response from server master-127.1.11.61:46759: Network error: Failed to connect to peer master-127.1.11.61:46759(127.1.11.61:46759): Connection refused: /127.1.11.61:46759, [431ms] Sub RPC ConnectToMaster: received response from server master-127.1.11.60:40051: OK, [436ms] delaying RPC due to: Service unavailable: Master config (127.1.11.62:46867,127.1.11.60:40051,127.1.11.61:46759) has no leader. Exceptions received: org.apache.kudu.client.RecoverableException: Failed to connect to peer master-127.1.11.62:46867(127.1.11.62:46867): Connection refused: /127.1.11.62:46867,org.apache.kudu.client.RecoverableException: Failed to connect to peer master-127.1.11.61:46759(127.1.11.61:46759): Connection refused: /127.1.11.61:46759, [455ms] refreshing cache from master, [456ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.62:46867, [458ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.60:40051, [460ms] Sub RPC ConnectToMaster: received response from server master-127.1.11.62:46867: Network error: Failed to connect to peer master-127.1.11.62:46867(127.1.11.62:46867): Connection refused: /127.1.11.62:46867, [460ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.61:46759, [463ms] Sub RPC ConnectToMaster: received response from server master-127.1.11.61:46759: Network error: Failed to connect to peer master-127.1.11.61:46759(127.1.11.61:46759): Connection refused: /127.1.11.61:46759, [465ms] Sub RPC ConnectToMaster: received response from server master-127.1.11.60:40051: OK, [466ms] delaying RPC due to: Service unavailable: Master config (127.1.11.62:46867,127.1.11.60:40051,127.1.11.61:46759) has no leader. Exceptions received: org.apache.kudu.client.RecoverableException: Failed to connect to peer master-127.1.11.62:46867(127.1.11.62:46867): Connection refused: /127.1.11.62:46867,org.apache.kudu.client.RecoverableException: Failed to connect to peer master-127.1.11.61:46759(127.1.11.61:46759): Connection refused: /127.1.11.61:46759, [475ms] refreshing cache from master, [476ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.62:46867, [477ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.60:40051, [477ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.61:46759, [480ms] Sub RPC ConnectToMaster: received response from server master-127.1.11.61:46759: Network error: Failed to connect to peer master-127.1.11.61:46759(127.1.11.61:46759): Connection refused: /127.1.11.61:46759, [481ms] Sub RPC ConnectToMaster: received response from server master-127.1.11.60:40051: OK, [482ms] Sub RPC ConnectToMaster: received response from server master-127.1.11.62:46867: Network error: Failed to connect to peer master-127.1.11.62:46867(127.1.11.62:46867): Connection refused: /127.1.11.62:46867, [482ms] delaying RPC due to: Service unavailable: Master config (127.1.11.62:46867,127.1.11.60:40051,127.1.11.61:46759) has no leader. Exceptions received: org.apache.kudu.client.RecoverableException: Failed to connect to peer master-127.1.11.61:46759(127.1.11.61:46759): Connection refused: /127.1.11.61:46759,org.apache.kudu.client.RecoverableException: Failed to connect to peer master-127.1.11.62:46867(127.1.11.62:46867): Connection refused: /127.1.11.62:46867, [495ms] refreshing cache from master, [496ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.62:46867, [497ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.60:40051, [497ms] Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.61:46759, [500ms] Sub RPC ConnectToMaster: received response from server master-127.1.11.61:46759: Network error: Failed to connect to peer master-127.1.11.61:46759(127.1.11.61:46759): Connection ref
[jira] [Assigned] (KUDU-3628) ksck_->CheckMasterConsensus() is flaky in TSAN build
[ https://issues.apache.org/jira/browse/KUDU-3628?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3628: Assignee: Bakai Ádám > ksck_->CheckMasterConsensus() is flaky in TSAN build > > > Key: KUDU-3628 > URL: https://issues.apache.org/jira/browse/KUDU-3628 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3628) ksck_->CheckMasterConsensus() is flaky in TSAN build
Bakai Ádám created KUDU-3628: Summary: ksck_->CheckMasterConsensus() is flaky in TSAN build Key: KUDU-3628 URL: https://issues.apache.org/jira/browse/KUDU-3628 Project: Kudu Issue Type: Sub-task Reporter: Bakai Ádám -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (KUDU-3628) ksck_->CheckMasterConsensus() is flaky in TSAN build
[ https://issues.apache.org/jira/browse/KUDU-3628?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17902633#comment-17902633 ] Bakai Ádám commented on KUDU-3628: -- This issue is a manifestation of KUDU-2709. > ksck_->CheckMasterConsensus() is flaky in TSAN build > > > Key: KUDU-3628 > URL: https://issues.apache.org/jira/browse/KUDU-3628 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (KUDU-3631) TSAN errors in tests when using InternalMiniCluster
[ https://issues.apache.org/jira/browse/KUDU-3631?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17902678#comment-17902678 ] Bakai Ádám commented on KUDU-3631: -- One example: {code:java} WARNING: ThreadSanitizer: destroy of a locked mutex (pid=411) #0 pthread_mutex_destroy /root/tmp/test123/kudu/thirdparty/src/llvm-11.0.0.src/projects/compiler-rt/lib/tsan/rtl/tsan_interceptors_posix.cpp:1241 (ksck_remote-test+0x308066) #1 kudu::Mutex::~Mutex() /root/tmp/test123/kudu/src/kudu/util/mutex.cc:76:12 (libkudu_util.so+0x3df980) #2 kudu::CountDownLatch::~CountDownLatch() /root/tmp/test123/kudu/src/kudu/util/countdown_latch.h:33:7 (ksck_remote-test+0x372262) #3 kudu::server::ServerBase::~ServerBase() /root/tmp/test123/kudu/src/kudu/server/server_base.cc:772:1 (libserver_process.so+0x1427ef) #4 kudu::kserver::KuduServer::~KuduServer() /root/tmp/test123/kudu/src/kudu/kserver/kserver.h:41:7 (libmaster.so+0x400b3f) #5 kudu::master::Master::~Master() /root/tmp/test123/kudu/src/kudu/master/master.cc:263:1 (libmaster.so+0x3f6dc2) #6 kudu::master::Master::~Master() /root/tmp/test123/kudu/src/kudu/master/master.cc:261:19 (libmaster.so+0x3f70b9) #7 std::__1::default_delete::operator()(kudu::master::Master*) const /root/tmp/test123/kudu/thirdparty/installed/tsan/include/c++/v1/memory:2262:5 (libmaster.so+0x434907) #8 std::__1::unique_ptr >::reset(kudu::master::Master*) /root/tmp/test123/kudu/thirdparty/installed/tsan/include/c++/v1/memory:2517:7 (libmaster.so+0x42d9cd) #9 kudu::master::MiniMaster::Shutdown() /root/tmp/test123/kudu/src/kudu/master/mini_master.cc:120:13 (libmaster.so+0x4c18c2) #10 kudu::tools::RemoteKsckTest_TestClusterWithLocation_Test::TestBody() /root/tmp/test123/kudu/src/kudu/tools/ksck_remote-test.cc:612:13 (ksck_remote-test+0x35a1f8) #11 void testing::internal::HandleSehExceptionsInMethodIfSupported(testing::Test*, void (testing::Test::*)(), char const*) /root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2599:10 (libgtest.so.1.12.1+0x64dbf) #12 void testing::internal::HandleExceptionsInMethodIfSupported(testing::Test*, void (testing::Test::*)(), char const*) /root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2635:14 (libgtest.so.1.12.1+0x64dbf) #13 testing::Test::Run() /root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2674:5 (libgtest.so.1.12.1+0x429c1) #14 testing::TestInfo::Run() /root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2853:11 (libgtest.so.1.12.1+0x43cd8) #15 testing::TestSuite::Run() /root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:3012:30 (libgtest.so.1.12.1+0x44cb4) #16 testing::internal::UnitTestImpl::RunAllTests() /root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:5870:44 (libgtest.so.1.12.1+0x597a4) #17 bool testing::internal::HandleSehExceptionsInMethodIfSupported(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2599:10 (libgtest.so.1.12.1+0x65c7f) #18 bool testing::internal::HandleExceptionsInMethodIfSupported(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2635:14 (libgtest.so.1.12.1+0x65c7f) #19 testing::UnitTest::Run() /root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:5444:10 (libgtest.so.1.12.1+0x58d5c) #20 RUN_ALL_TESTS() /root/tmp/test123/kudu/thirdparty/installed/tsan/include/gtest/gtest.h:2293:73 (ksck_remote-test+0x38d91b) #21 main /root/tmp/test123/kudu/src/kudu/util/test_main.cc:109:10 (ksck_remote-test+0x38c81c) and: #0 pthread_cond_timedwait /root/tmp/test123/kudu/thirdparty/src/llvm-11.0.0.src/projects/compiler-rt/lib/tsan/rtl/tsan_interceptors_posix.cpp:1181 (ksck_remote-test+0x2f5550) #1 kudu::ConditionVariable::WaitUntil(kudu::MonoTime const&) const /root/tmp/test123/kudu/src/kudu/util/condition_variable.cc:87:12 (libkudu_util.so+0x2c400a) #2 kudu::CountDownLatch::WaitUntil(kudu::MonoTime const&) const /root/tmp/test123/kudu/src/kudu/util/countdown_latch.h:89:18 (ksck_remote-test+0x38abe8) #3 kudu::CountDownLatch::WaitFor(kudu::MonoDelta const&) const /root/tmp/test123/kudu/src/kudu/util/countdown_latch.h:99:12 (ksck_remote-test+0x38ab5e) #4 kudu::master::Master::ExpiredReservedTablesDeleterThread() /root/tmp/test123/kudu/src/kudu/master/master.cc:568:42 (libmaster.so+0x3f975e) #5 kudu::master::Master::StartExpiredReservedTablesDeleterThread()::$_2::operator()() const /root/tmp/te
[jira] [Created] (KUDU-3631) TSAN errors in tests related to InternalMiniCluster
Bakai Ádám created KUDU-3631: Summary: TSAN errors in tests related to InternalMiniCluster Key: KUDU-3631 URL: https://issues.apache.org/jira/browse/KUDU-3631 Project: Kudu Issue Type: Sub-task Reporter: Bakai Ádám -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3631) TSAN errors in tests when using InternalMiniCluster
[ https://issues.apache.org/jira/browse/KUDU-3631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3631: - Summary: TSAN errors in tests when using InternalMiniCluster (was: TSAN errors in tests related to InternalMiniCluster) > TSAN errors in tests when using InternalMiniCluster > --- > > Key: KUDU-3631 > URL: https://issues.apache.org/jira/browse/KUDU-3631 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3631) TSAN error in CountDownLatch destructor when using InternalMiniCluster
[ https://issues.apache.org/jira/browse/KUDU-3631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3631: - Summary: TSAN error in CountDownLatch destructor when using InternalMiniCluster (was: TSAN errors in tests when using InternalMiniCluster) > TSAN error in CountDownLatch destructor when using InternalMiniCluster > -- > > Key: KUDU-3631 > URL: https://issues.apache.org/jira/browse/KUDU-3631 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (KUDU-3633) Threadpool check flakiness in ksck_remote-test during MiniMaster shutdown
[ https://issues.apache.org/jira/browse/KUDU-3633?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3633: Assignee: Bakai Ádám > Threadpool check flakiness in ksck_remote-test during MiniMaster shutdown > - > > Key: KUDU-3633 > URL: https://issues.apache.org/jira/browse/KUDU-3633 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > > {code:java} > F20241204 12:57:40.147302 16123 threadpool.cc:391] Check failed: 1 == > tokens_.size() (1 vs. 3) Threadpool raft destroyed with 3 allocated tokens > *** Check failure stack trace: ***\{code} > {code:java} > @ 0x7f6b96b2cd64 google::LogMessage::SendToLog() at ??:0 > @ 0x7f6b96b2d910 google::LogMessage::Flush() at ??:0 > @ 0x7f6b96b32a4b google::LogMessageFatal::~LogMessageFatal() at ??:0 > @ 0x7f6b974a777d kudu::ThreadPool::~ThreadPool() at ??:0 > I20241204 12:57:40.556027 23288 raft_consensus.cc:1270] T > df574f38d0a746d1929d9494d82da991 P c273df5d41694d4da3bc1b5bc5e81b84 [term 2 > FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: > Log matching property violated. Preceding OpId in replica: term: 1 index: 1. > Preceding OpId from leader: term: 2 index: 2. (index mismatch) > I20241204 12:57:40.558073 23666 consensus_queue.cc:1035] T > df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: > Connected to new peer: Peer: permanent_uuid: > "c273df5d41694d4da3bc1b5bc5e81b84" member_type: VOTER last_known_addr { host: > "127.15.190.193" port: 33967 }, Status: LMP_MISMATCH, Last received: 0.0, > Next index: 2, Last known committed idx: 1, Time since last communication: > 0.000s > @ 0x7f6b9ff4f6bf std::__1::default_delete<>::operator()() at ??:0 > I20241204 12:57:40.605798 23460 raft_consensus.cc:1270] T > df574f38d0a746d1929d9494d82da991 P 87f06d0d674a4791871f81a7af62b7be [term 2 > FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: > Log matching property violated. Preceding OpId in replica: term: 1 index: 1. > Preceding OpId from leader: term: 2 index: 2. (index mismatch) > I20241204 12:57:40.611544 23707 consensus_queue.cc:1035] T > df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: > Connected to new peer: Peer: permanent_uuid: > "87f06d0d674a4791871f81a7af62b7be" member_type: VOTER last_known_addr { host: > "127.15.190.195" port: 35365 }, Status: LMP_MISMATCH, Last received: 0.0, > Next index: 2, Last known committed idx: 1, Time since last communication: > 0.000s > @ 0x7f6b9ff4f62e std::__1::unique_ptr<>::reset() at ??:0 > @ 0x7f6b9ff0e2cc std::__1::unique_ptr<>::~unique_ptr() at ??:0 > @ 0x7f6b9ffb65b4 kudu::kserver::KuduServer::~KuduServer() at ??:0 > @ 0x7f6b9ffac863 kudu::master::Master::~Master() at ??:0 > @ 0x7f6b9ffacb5a kudu::master::Master::~Master() at ??:0 > @ 0x7f6b9ffea408 std::__1::default_delete<>::operator()() at ??:0 > @ 0x7f6b9ffe34ce std::__1::unique_ptr<>::reset() at ??:0 > @ 0x7f6ba00773c3 kudu::master::MiniMaster::Shutdown() at ??:0 > @ 0x354ea9 > kudu::tools::RemoteKsckTest_TestClusterWithLocation_Test::TestBody() at > /root/tmp/test123/kudu/src/kudu/tools/ksck_remote-test.cc:607 > @ 0x7f6ba045adc0 > testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0 > @ 0x7f6ba04389c2 testing::Test::Run() at ??:0 > @ 0x7f6ba0439cd9 testing::TestInfo::Run() at ??:0 > @ 0x7f6ba043acb5 testing::TestSuite::Run() at ??:0 > @ 0x7f6ba044f7a5 testing::internal::UnitTestImpl::RunAllTests() at > ??:0 > @ 0x7f6ba045bc80 > testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0 > @ 0x7f6ba044ed5d testing::UnitTest::Run() at ??:0 > @ 0x3801bc RUN_ALL_TESTS() at > /root/tmp/test123/kudu/thirdparty/installed/tsan/include/gtest/gtest.h:? > @ 0x37f0bd main at > /root/tmp/test123/kudu/src/kudu/util/test_main.cc:? > @ 0x7f6b93f58bf7 __libc_start_main at ??:0 > @ 0x298ada _start at ??:? {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3633) Threadpool check flakiness in ksck_remote-test during MiniMaster shutdown
[ https://issues.apache.org/jira/browse/KUDU-3633?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3633: - Summary: Threadpool check flakiness in ksck_remote-test during MiniMaster shutdown (was: Threadpool check flakiness in ksck_remote-test) > Threadpool check flakiness in ksck_remote-test during MiniMaster shutdown > - > > Key: KUDU-3633 > URL: https://issues.apache.org/jira/browse/KUDU-3633 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Priority: Major > > {code:java} > F20241204 12:57:40.147302 16123 threadpool.cc:391] Check failed: 1 == > tokens_.size() (1 vs. 3) Threadpool raft destroyed with 3 allocated tokens > *** Check failure stack trace: ***\{code} > {code:java} > @ 0x7f6b96b2cd64 google::LogMessage::SendToLog() at ??:0 > @ 0x7f6b96b2d910 google::LogMessage::Flush() at ??:0 > @ 0x7f6b96b32a4b google::LogMessageFatal::~LogMessageFatal() at ??:0 > @ 0x7f6b974a777d kudu::ThreadPool::~ThreadPool() at ??:0 > I20241204 12:57:40.556027 23288 raft_consensus.cc:1270] T > df574f38d0a746d1929d9494d82da991 P c273df5d41694d4da3bc1b5bc5e81b84 [term 2 > FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: > Log matching property violated. Preceding OpId in replica: term: 1 index: 1. > Preceding OpId from leader: term: 2 index: 2. (index mismatch) > I20241204 12:57:40.558073 23666 consensus_queue.cc:1035] T > df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: > Connected to new peer: Peer: permanent_uuid: > "c273df5d41694d4da3bc1b5bc5e81b84" member_type: VOTER last_known_addr { host: > "127.15.190.193" port: 33967 }, Status: LMP_MISMATCH, Last received: 0.0, > Next index: 2, Last known committed idx: 1, Time since last communication: > 0.000s > @ 0x7f6b9ff4f6bf std::__1::default_delete<>::operator()() at ??:0 > I20241204 12:57:40.605798 23460 raft_consensus.cc:1270] T > df574f38d0a746d1929d9494d82da991 P 87f06d0d674a4791871f81a7af62b7be [term 2 > FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: > Log matching property violated. Preceding OpId in replica: term: 1 index: 1. > Preceding OpId from leader: term: 2 index: 2. (index mismatch) > I20241204 12:57:40.611544 23707 consensus_queue.cc:1035] T > df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: > Connected to new peer: Peer: permanent_uuid: > "87f06d0d674a4791871f81a7af62b7be" member_type: VOTER last_known_addr { host: > "127.15.190.195" port: 35365 }, Status: LMP_MISMATCH, Last received: 0.0, > Next index: 2, Last known committed idx: 1, Time since last communication: > 0.000s > @ 0x7f6b9ff4f62e std::__1::unique_ptr<>::reset() at ??:0 > @ 0x7f6b9ff0e2cc std::__1::unique_ptr<>::~unique_ptr() at ??:0 > @ 0x7f6b9ffb65b4 kudu::kserver::KuduServer::~KuduServer() at ??:0 > @ 0x7f6b9ffac863 kudu::master::Master::~Master() at ??:0 > @ 0x7f6b9ffacb5a kudu::master::Master::~Master() at ??:0 > @ 0x7f6b9ffea408 std::__1::default_delete<>::operator()() at ??:0 > @ 0x7f6b9ffe34ce std::__1::unique_ptr<>::reset() at ??:0 > @ 0x7f6ba00773c3 kudu::master::MiniMaster::Shutdown() at ??:0 > @ 0x354ea9 > kudu::tools::RemoteKsckTest_TestClusterWithLocation_Test::TestBody() at > /root/tmp/test123/kudu/src/kudu/tools/ksck_remote-test.cc:607 > @ 0x7f6ba045adc0 > testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0 > @ 0x7f6ba04389c2 testing::Test::Run() at ??:0 > @ 0x7f6ba0439cd9 testing::TestInfo::Run() at ??:0 > @ 0x7f6ba043acb5 testing::TestSuite::Run() at ??:0 > @ 0x7f6ba044f7a5 testing::internal::UnitTestImpl::RunAllTests() at > ??:0 > @ 0x7f6ba045bc80 > testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0 > @ 0x7f6ba044ed5d testing::UnitTest::Run() at ??:0 > @ 0x3801bc RUN_ALL_TESTS() at > /root/tmp/test123/kudu/thirdparty/installed/tsan/include/gtest/gtest.h:? > @ 0x37f0bd main at > /root/tmp/test123/kudu/src/kudu/util/test_main.cc:? > @ 0x7f6b93f58bf7 __libc_start_main at ??:0 > @ 0x298ada _start at ??:? {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3633) Threadpool check flakiness in ksck_remote-test
Bakai Ádám created KUDU-3633: Summary: Threadpool check flakiness in ksck_remote-test Key: KUDU-3633 URL: https://issues.apache.org/jira/browse/KUDU-3633 Project: Kudu Issue Type: Sub-task Reporter: Bakai Ádám -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3633) Threadpool check flakiness in ksck_remote-test
[ https://issues.apache.org/jira/browse/KUDU-3633?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3633: - Description: {code:java} F20241204 12:57:40.147302 16123 threadpool.cc:391] Check failed: 1 == tokens_.size() (1 vs. 3) Threadpool raft destroyed with 3 allocated tokens *** Check failure stack trace: ***\{code} {code:java} @ 0x7f6b96b2cd64 google::LogMessage::SendToLog() at ??:0 @ 0x7f6b96b2d910 google::LogMessage::Flush() at ??:0 @ 0x7f6b96b32a4b google::LogMessageFatal::~LogMessageFatal() at ??:0 @ 0x7f6b974a777d kudu::ThreadPool::~ThreadPool() at ??:0 I20241204 12:57:40.556027 23288 raft_consensus.cc:1270] T df574f38d0a746d1929d9494d82da991 P c273df5d41694d4da3bc1b5bc5e81b84 [term 2 FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: Log matching property violated. Preceding OpId in replica: term: 1 index: 1. Preceding OpId from leader: term: 2 index: 2. (index mismatch) I20241204 12:57:40.558073 23666 consensus_queue.cc:1035] T df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: Connected to new peer: Peer: permanent_uuid: "c273df5d41694d4da3bc1b5bc5e81b84" member_type: VOTER last_known_addr { host: "127.15.190.193" port: 33967 }, Status: LMP_MISMATCH, Last received: 0.0, Next index: 2, Last known committed idx: 1, Time since last communication: 0.000s @ 0x7f6b9ff4f6bf std::__1::default_delete<>::operator()() at ??:0 I20241204 12:57:40.605798 23460 raft_consensus.cc:1270] T df574f38d0a746d1929d9494d82da991 P 87f06d0d674a4791871f81a7af62b7be [term 2 FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: Log matching property violated. Preceding OpId in replica: term: 1 index: 1. Preceding OpId from leader: term: 2 index: 2. (index mismatch) I20241204 12:57:40.611544 23707 consensus_queue.cc:1035] T df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: Connected to new peer: Peer: permanent_uuid: "87f06d0d674a4791871f81a7af62b7be" member_type: VOTER last_known_addr { host: "127.15.190.195" port: 35365 }, Status: LMP_MISMATCH, Last received: 0.0, Next index: 2, Last known committed idx: 1, Time since last communication: 0.000s @ 0x7f6b9ff4f62e std::__1::unique_ptr<>::reset() at ??:0 @ 0x7f6b9ff0e2cc std::__1::unique_ptr<>::~unique_ptr() at ??:0 @ 0x7f6b9ffb65b4 kudu::kserver::KuduServer::~KuduServer() at ??:0 @ 0x7f6b9ffac863 kudu::master::Master::~Master() at ??:0 @ 0x7f6b9ffacb5a kudu::master::Master::~Master() at ??:0 @ 0x7f6b9ffea408 std::__1::default_delete<>::operator()() at ??:0 @ 0x7f6b9ffe34ce std::__1::unique_ptr<>::reset() at ??:0 @ 0x7f6ba00773c3 kudu::master::MiniMaster::Shutdown() at ??:0 @ 0x354ea9 kudu::tools::RemoteKsckTest_TestClusterWithLocation_Test::TestBody() at /root/tmp/test123/kudu/src/kudu/tools/ksck_remote-test.cc:607 @ 0x7f6ba045adc0 testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0 @ 0x7f6ba04389c2 testing::Test::Run() at ??:0 @ 0x7f6ba0439cd9 testing::TestInfo::Run() at ??:0 @ 0x7f6ba043acb5 testing::TestSuite::Run() at ??:0 @ 0x7f6ba044f7a5 testing::internal::UnitTestImpl::RunAllTests() at ??:0 @ 0x7f6ba045bc80 testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0 @ 0x7f6ba044ed5d testing::UnitTest::Run() at ??:0 @ 0x3801bc RUN_ALL_TESTS() at /root/tmp/test123/kudu/thirdparty/installed/tsan/include/gtest/gtest.h:? @ 0x37f0bd main at /root/tmp/test123/kudu/src/kudu/util/test_main.cc:? @ 0x7f6b93f58bf7 __libc_start_main at ??:0 @ 0x298ada _start at ??:? {code} > Threadpool check flakiness in ksck_remote-test > ------ > > Key: KUDU-3633 > URL: https://issues.apache.org/jira/browse/KUDU-3633 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Priority: Major > > {code:java} > F20241204 12:57:40.147302 16123 threadpool.cc:391] Check failed: 1 == > tokens_.size() (1 vs. 3) Threadpool raft destroyed with 3 allocated tokens > *** Check failure stack trace: ***\{code} > {code:java} > @ 0x7f6b96b2cd64 google::LogMessage::SendToLog() at ??:0 > @ 0x7f6b96b2d910 google::LogMessage::Flush() at ??:0 > @ 0x7f6b96b32a4b google::LogMessageFatal::~LogMessageFatal() at ??:0 > @ 0x7f6b974a777d kudu::ThreadPool::~ThreadPool() at ??:0 > I20241204 12:57:40.556027 23288 raft_consensus.cc:1270] T > df574f38d0a746d1929d9494d82da991 P c273df5d41694d4da3bc1b5bc5e81b84 [term 2 > FOLLOWER]: Refusing update from remote peer 2e54eeefd5f94
[jira] [Resolved] (KUDU-3563) Output tablet-level metrics in Prometheus format
[ https://issues.apache.org/jira/browse/KUDU-3563?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám resolved KUDU-3563. -- Fix Version/s: 1.18.0 Resolution: Fixed > Output tablet-level metrics in Prometheus format > > > Key: KUDU-3563 > URL: https://issues.apache.org/jira/browse/KUDU-3563 > Project: Kudu > Issue Type: Improvement > Components: master, server, tserver >Reporter: Alexey Serbin >Assignee: Bakai Ádám >Priority: Major > Labels: metrics, prometheus, supportability, visibility > Fix For: 1.18.0 > > > The request to support outputting Kudu metrics in Prometheus format is > tracked [KUDU-3375|https://issues.apache.org/jira/browse/KUDU-3375]. The > [first take on > this|https://github.com/apache/kudu/commit/00efc6826ac9a1f5d10750296c7357790a041fec] > has taken care of the server-level metrics, ignoring all the tablet-level > metrics. > In the scope of this JIRA item, it's necessary to output all the tablet-level > metrics in Prometheus format as well. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3634) check error in ksck_results.cc in TSAN build
Bakai Ádám created KUDU-3634: Summary: check error in ksck_results.cc in TSAN build Key: KUDU-3634 URL: https://issues.apache.org/jira/browse/KUDU-3634 Project: Kudu Issue Type: Sub-task Reporter: Bakai Ádám -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3634) check error in ksck_results.cc in TSAN build
[ https://issues.apache.org/jira/browse/KUDU-3634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3634: - Description: {code:java} F20241218 13:05:04.229065 4953 ksck_results.cc:160] Check failed: servers.size() <= server_count (6 vs. 3) *** Check failure stack trace: *** *** Aborted at 1734527104 (unix time) try "date -d @1734527104" if you are using GNU date *** PC: @ 0x0 (unknown) *** SIGABRT (@0x3e81359) received by PID 4953 (TID 0x7f5270649880) from PID 4953; stack trace: *** @ 0x2d8140 __tsan::CallUserSignalHandler() at /root/tmp/test123/kudu/thirdparty/src/llvm-11.0.0.src/projects/compiler-rt/lib/tsan/rtl/tsan_interceptors_posix.cpp:1915 @ 0x2da623 rtl_sigaction() at /root/tmp/test123/kudu/thirdparty/src/llvm-11.0.0.src/projects/compiler-rt/lib/tsan/rtl/tsan_interceptors_posix.cpp:2005 I20241218 13:05:04.427536 14086 raft_consensus.cc:1270] T 5d3037cff4d843b0a7dbc9ff78a339ba P 3e6bbc8984eb4fbfbefd27d5036db51a [term 5 FOLLOWER]: Refusing update from remote peer 1a37ef9543954c8e823eeae2d8592107: Log matching property violated. Preceding OpId in replica: term: 4 index: 3. Preceding OpId from leader: term: 5 index: 4. (index mismatch) I20241218 13:05:04.429109 14549 consensus_queue.cc:1035] T 5d3037cff4d843b0a7dbc9ff78a339ba P 1a37ef9543954c8e823eeae2d8592107 [LEADER]: Connected to new peer: Peer: permanent_uuid: "3e6bbc8984eb4fbfbefd27d5036db51a" member_type: VOTER last_known_addr { host: "127.4.214.67" port: 37251 }, Status: LMP_MISMATCH, Last received: 0.0, Next index: 4, Last known committed idx: 3, Time since last communication: 0.000s @ 0x7f527808f980 (unknown) at ??:0 I20241218 13:05:04.509820 14545 raft_consensus.cc:491] T 5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a [term 5 FOLLOWER]: Starting pre-election (no leader contacted us within the election timeout) I20241218 13:05:04.510313 14545 raft_consensus.cc:513] T 5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a [term 5 FOLLOWER]: Starting pre-election with config: opid_index: -1 OBSOLETE_local: false peers { permanent_uuid: "0bfce15bbb5f4f0aa5a9e60872491f3a" member_type: VOTER last_known_addr { host: "127.4.214.65" port: 36387 } } peers { permanent_uuid: "1a37ef9543954c8e823eeae2d8592107" member_type: VOTER last_known_addr { host: "127.4.214.66" port: 36001 } } peers { permanent_uuid: "3e6bbc8984eb4fbfbefd27d5036db51a" member_type: VOTER last_known_addr { host: "127.4.214.67" port: 37251 } } I20241218 13:05:04.513247 14001 tablet_service.cc:1812] Received RequestConsensusVote() RPC: tablet_id: "5d3037cff4d843b0a7dbc9ff78a339ba" candidate_uuid: "0bfce15bbb5f4f0aa5a9e60872491f3a" candidate_term: 6 candidate_status { last_received { term: 3 index: 2 } } ignore_live_leader: false dest_uuid: "1a37ef9543954c8e823eeae2d8592107" is_pre_election: true I20241218 13:05:04.516223 14089 tablet_service.cc:1812] Received RequestConsensusVote() RPC: tablet_id: "5d3037cff4d843b0a7dbc9ff78a339ba" candidate_uuid: "0bfce15bbb5f4f0aa5a9e60872491f3a" candidate_term: 6 candidate_status { last_received { term: 3 index: 2 } } ignore_live_leader: false dest_uuid: "3e6bbc8984eb4fbfbefd27d5036db51a" is_pre_election: true I20241218 13:05:04.517762 13891 leader_election.cc:304] T 5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a [CANDIDATE]: Term 6 pre-election: Election decided. Result: candidate lost. Election summary: received 3 responses out of 3 voters: 1 yes votes; 2 no votes. yes voters: 0bfce15bbb5f4f0aa5a9e60872491f3a; no voters: 1a37ef9543954c8e823eeae2d8592107, 3e6bbc8984eb4fbfbefd27d5036db51a I20241218 13:05:04.518554 14506 raft_consensus.cc:2743] T 5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a [term 5 FOLLOWER]: Leader pre-election lost for term 6. Reason: could not achieve majority I20241218 13:05:04.518522 14545 leader_election.cc:290] T 5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a [CANDIDATE]: Term 6 pre-election: Requested pre-vote from peers 1a37ef9543954c8e823eeae2d8592107 (127.4.214.66:36001), 3e6bbc8984eb4fbfbefd27d5036db51a (127.4.214.67:37251) @ 0x7f5271309fb7 gsignal at ??:0 I20241218 13:05:04.633380 13919 raft_consensus.cc:1270] T 5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a [term 5 FOLLOWER]: Refusing update from remote peer 1a37ef9543954c8e823eeae2d8592107: Log matching property violated. Preceding OpId in replica: term: 3 index: 2. Preceding OpId from leader: term: 5 index: 4. (index mismatch) I20241218 13:05:04.634791 14549 consensus_queue.cc:1035] T 5d3037cff4d843b0a7dbc9ff78a339ba P 1a37ef9543954c8e823eeae2d8592107 [LEADER]: Connected to new peer: Peer: permanent_uui
[jira] [Assigned] (KUDU-3631) TSAN error in CountDownLatch destructor when using InternalMiniCluster
[ https://issues.apache.org/jira/browse/KUDU-3631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3631: Assignee: Bakai Ádám > TSAN error in CountDownLatch destructor when using InternalMiniCluster > -- > > Key: KUDU-3631 > URL: https://issues.apache.org/jira/browse/KUDU-3631 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (KUDU-3637) TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky
Bakai Ádám created KUDU-3637: Summary: TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky Key: KUDU-3637 URL: https://issues.apache.org/jira/browse/KUDU-3637 Project: Kudu Issue Type: Sub-task Reporter: Bakai Ádám -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (KUDU-3637) TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky
[ https://issues.apache.org/jira/browse/KUDU-3637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám reassigned KUDU-3637: Assignee: Bakai Ádám > TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky > --- > > Key: KUDU-3637 > URL: https://issues.apache.org/jira/browse/KUDU-3637 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (KUDU-3637) TestKuduTransaction is flaky
[ https://issues.apache.org/jira/browse/KUDU-3637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bakai Ádám updated KUDU-3637: - Summary: TestKuduTransaction is flaky (was: TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky) > TestKuduTransaction is flaky > > > Key: KUDU-3637 > URL: https://issues.apache.org/jira/browse/KUDU-3637 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (KUDU-3637) TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky
[ https://issues.apache.org/jira/browse/KUDU-3637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17912931#comment-17912931 ] Bakai Ádám commented on KUDU-3637: -- In testTxnKeepaliveSwitchesToOtherTxnManager (keep in mind, this is the not rolling version) According to the logs, for some reason the operation was aborted: {code:java} 13:39:39.728 [INFO - cluster stderr printer] (MiniKuduCluster.java:751) W20250113 13:39:39.728489 26745 tablet_replica.cc:1307] Aborted: operation has been aborted: cancelling pending write operations {code} I have not found yet the reason why. > TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky > --- > > Key: KUDU-3637 > URL: https://issues.apache.org/jira/browse/KUDU-3637 > Project: Kudu > Issue Type: Sub-task >Reporter: Bakai Ádám >Assignee: Bakai Ádám >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)