[jira] [Created] (KUDU-3019) Kudu client hangs in Deffered.join()

2019-12-12 Thread Jira
Zoltán Borók-Nagy created KUDU-3019:
---

 Summary: Kudu client hangs in Deffered.join()
 Key: KUDU-3019
 URL: https://issues.apache.org/jira/browse/KUDU-3019
 Project: Kudu
  Issue Type: Bug
Reporter: Zoltán Borók-Nagy


In Impala we've seen the Kudu client hanging with the following stack trace:
{noformat}
Thread 53015: (state = BLOCKED) - java.lang.Object.wait(long) @bci=0 (Compiled 
frame; information may be imprecise) - java.lang.Object.wait() @bci=2, line=502 
(Compiled frame) - com.stumbleupon.async.Deferred.doJoin(boolean, long) 
@bci=77, line=1122 (Compiled frame) - com.stumbleupon.async.Deferred.join() 
@bci=3, line=1006 (Compiled frame) - 
org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred)
 @bci=1, line=340 (Compiled frame) - 
org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, line=212 
(Compiled frame) - 
org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer)
 @bci=32, line=115 (Compiled frame)
...{noformat}
The client hangs in Deferred.join():

[https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423]

To at least mitigate the problem, maybe Deferred.join(long timeout) could be 
used instead.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (KUDU-3019) Kudu client hangs in Deffered.join()

2019-12-12 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3019?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltán Borók-Nagy updated KUDU-3019:

Description: 
In Impala we've seen the Kudu client hanging with the following stack trace:
{noformat}
Thread 53015: (state = BLOCKED)
 - java.lang.Object.wait(long) @bci=0 (Compiled frame; information may be 
imprecise)
 - java.lang.Object.wait() @bci=2, line=502 (Compiled frame)
 - com.stumbleupon.async.Deferred.doJoin(boolean, long) @bci=77, line=1122 
(Compiled frame)
 - com.stumbleupon.async.Deferred.join() @bci=3, line=1006 (Compiled frame)
 - 
org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred)
 @bci=1, line=340 (Compiled frame)
 - org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, 
line=212 (Compiled frame)
 - 
org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer)
 @bci=32, line=115 (Compiled frame)
 - 
org.apache.impala.planner.SingleNodePlanner.createScanNode(org.apache.impala.analysis.TableRef,
 org.apache.impala.analysis.AggregateInfo, org.apache.impala.analysis.Analyzer) 
@bci=252, line=1312 (Compiled frame)
...{noformat}
The client hangs in Deferred.join():

[https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423]

To at least mitigate the problem, maybe Deferred.join(long timeout) could be 
used instead.

  was:
In Impala we've seen the Kudu client hanging with the following stack trace:
{noformat}
Thread 53015: (state = BLOCKED) - java.lang.Object.wait(long) @bci=0 (Compiled 
frame; information may be imprecise) - java.lang.Object.wait() @bci=2, line=502 
(Compiled frame) - com.stumbleupon.async.Deferred.doJoin(boolean, long) 
@bci=77, line=1122 (Compiled frame) - com.stumbleupon.async.Deferred.join() 
@bci=3, line=1006 (Compiled frame) - 
org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred)
 @bci=1, line=340 (Compiled frame) - 
org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, line=212 
(Compiled frame) - 
org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer)
 @bci=32, line=115 (Compiled frame)
...{noformat}
The client hangs in Deferred.join():

[https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423]

To at least mitigate the problem, maybe Deferred.join(long timeout) could be 
used instead.


> Kudu client hangs in Deffered.join()
> 
>
> Key: KUDU-3019
> URL: https://issues.apache.org/jira/browse/KUDU-3019
> Project: Kudu
>  Issue Type: Bug
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>
> In Impala we've seen the Kudu client hanging with the following stack trace:
> {noformat}
> Thread 53015: (state = BLOCKED)
>  - java.lang.Object.wait(long) @bci=0 (Compiled frame; information may be 
> imprecise)
>  - java.lang.Object.wait() @bci=2, line=502 (Compiled frame)
>  - com.stumbleupon.async.Deferred.doJoin(boolean, long) @bci=77, line=1122 
> (Compiled frame)
>  - com.stumbleupon.async.Deferred.join() @bci=3, line=1006 (Compiled frame)
>  - 
> org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred)
>  @bci=1, line=340 (Compiled frame)
>  - org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, 
> line=212 (Compiled frame)
>  - 
> org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer)
>  @bci=32, line=115 (Compiled frame)
>  - 
> org.apache.impala.planner.SingleNodePlanner.createScanNode(org.apache.impala.analysis.TableRef,
>  org.apache.impala.analysis.AggregateInfo, 
> org.apache.impala.analysis.Analyzer) @bci=252, line=1312 (Compiled frame)
> ...{noformat}
> The client hangs in Deferred.join():
> [https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423]
> To at least mitigate the problem, maybe Deferred.join(long timeout) could be 
> used instead.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (KUDU-3019) Kudu client hangs in Deffered.join()

2019-12-13 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3019?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltán Borók-Nagy updated KUDU-3019:

Affects Version/s: 1.6.0

> Kudu client hangs in Deffered.join()
> 
>
> Key: KUDU-3019
> URL: https://issues.apache.org/jira/browse/KUDU-3019
> Project: Kudu
>  Issue Type: Bug
>Affects Versions: 1.6.0
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>
> In Impala we've seen the Kudu client hanging with the following stack trace:
> {noformat}
> Thread 53015: (state = BLOCKED)
>  - java.lang.Object.wait(long) @bci=0 (Compiled frame; information may be 
> imprecise)
>  - java.lang.Object.wait() @bci=2, line=502 (Compiled frame)
>  - com.stumbleupon.async.Deferred.doJoin(boolean, long) @bci=77, line=1122 
> (Compiled frame)
>  - com.stumbleupon.async.Deferred.join() @bci=3, line=1006 (Compiled frame)
>  - 
> org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred)
>  @bci=1, line=340 (Compiled frame)
>  - org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, 
> line=212 (Compiled frame)
>  - 
> org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer)
>  @bci=32, line=115 (Compiled frame)
>  - 
> org.apache.impala.planner.SingleNodePlanner.createScanNode(org.apache.impala.analysis.TableRef,
>  org.apache.impala.analysis.AggregateInfo, 
> org.apache.impala.analysis.Analyzer) @bci=252, line=1312 (Compiled frame)
> ...{noformat}
> The client hangs in Deferred.join():
> [https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423]
> To at least mitigate the problem, maybe Deferred.join(long timeout) could be 
> used instead.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KUDU-3019) Kudu client hangs in Deffered.join()

2019-12-13 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3019?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16995540#comment-16995540
 ] 

Zoltán Borók-Nagy commented on KUDU-3019:
-

Thanks for looking at it, [~tlipcon].

It's version 1.6.0. The timeouts are the Impala defaults, which means 18ms 
for "defaultAdminOperationTimeoutMs" and "defaultOperationTimeoutMs" for the 
Kudu client. The query hangs for hours in that RPC call.

I'm seeing it at a customer, not able to repro.

> Kudu client hangs in Deffered.join()
> 
>
> Key: KUDU-3019
>     URL: https://issues.apache.org/jira/browse/KUDU-3019
> Project: Kudu
>  Issue Type: Bug
>Affects Versions: 1.6.0
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>
> In Impala we've seen the Kudu client hanging with the following stack trace:
> {noformat}
> Thread 53015: (state = BLOCKED)
>  - java.lang.Object.wait(long) @bci=0 (Compiled frame; information may be 
> imprecise)
>  - java.lang.Object.wait() @bci=2, line=502 (Compiled frame)
>  - com.stumbleupon.async.Deferred.doJoin(boolean, long) @bci=77, line=1122 
> (Compiled frame)
>  - com.stumbleupon.async.Deferred.join() @bci=3, line=1006 (Compiled frame)
>  - 
> org.apache.kudu.client.KuduClient.joinAndHandleException(com.stumbleupon.async.Deferred)
>  @bci=1, line=340 (Compiled frame)
>  - org.apache.kudu.client.KuduClient.openTable(java.lang.String) @bci=10, 
> line=212 (Compiled frame)
>  - 
> org.apache.impala.planner.KuduScanNode.init(org.apache.impala.analysis.Analyzer)
>  @bci=32, line=115 (Compiled frame)
>  - 
> org.apache.impala.planner.SingleNodePlanner.createScanNode(org.apache.impala.analysis.TableRef,
>  org.apache.impala.analysis.AggregateInfo, 
> org.apache.impala.analysis.Analyzer) @bci=252, line=1312 (Compiled frame)
> ...{noformat}
> The client hangs in Deferred.join():
> [https://github.com/apache/kudu/blob/a8c6ea258c06407c1a4fef260da3a1cb70529bd9/java/kudu-client/src/main/java/org/apache/kudu/client/KuduClient.java#L423]
> To at least mitigate the problem, maybe Deferred.join(long timeout) could be 
> used instead.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (KUDU-3041) Kudu Java client shade is incomplete

2020-01-14 Thread Jira
Ismaël Mejía created KUDU-3041:
--

 Summary: Kudu Java client shade is incomplete
 Key: KUDU-3041
 URL: https://issues.apache.org/jira/browse/KUDU-3041
 Project: Kudu
  Issue Type: Bug
  Components: client
Affects Versions: 1.11.1
Reporter: Ismaël Mejía


While working on an update of the Kudu integration on Apache Beam BEAM-5086 We 
found this issue. We use [tool to test for linkage 
errors|https://github.com/GoogleCloudPlatform/cloud-opensource-java] and it 
reports the classes that are missing but required by other classes.

This is the result for the kudu-client case:
{code:java}
Class javax.servlet.ServletOutputStream is not found;
 referenced by 1 class file
 
org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet 
(kudu-client-1.11.1.jar)
Class javax.servlet.http.HttpServlet is not found;
 referenced by 1 class file
 
org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet 
(kudu-client-1.11.1.jar)
Class javax.servlet.ServletException is not found;
 referenced by 1 class file
 
org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet 
(kudu-client-1.11.1.jar)
Class javax.servlet.ServletConfig is not found;
 referenced by 1 class file
 
org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet 
(kudu-client-1.11.1.jar)
Class javax.servlet.http.HttpServletRequest is not found;
 referenced by 1 class file
 
org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet 
(kudu-client-1.11.1.jar)
Class javax.servlet.http.HttpServletResponse is not found;
 referenced by 1 class file
 
org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet 
(kudu-client-1.11.1.jar)
Class org.jboss.marshalling.ByteInput is not found;
 referenced by 4 class files
 
org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ChannelBufferByteInput
 (kudu-client-1.11.1.jar)
 
org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.LimitingByteInput
 (kudu-client-1.11.1.jar)
 
org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ChannelBufferByteInput
 (beam-vendor-grpc-1_21_0-0.1.jar)
 
org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.LimitingByteInput
 (beam-vendor-grpc-1_21_0-0.1.jar)
Class org.jboss.marshalling.ByteOutput is not found;
 referenced by 2 class files
 
org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ChannelBufferByteOutput
 (kudu-client-1.11.1.jar)
 
org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ChannelBufferByteOutput
 (beam-vendor-grpc-1_21_0-0.1.jar)
Class org.jboss.marshalling.Unmarshaller is not found;
 referenced by 8 class files
 
org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.CompatibleMarshallingDecoder
 (kudu-client-1.11.1.jar)
 
org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ContextBoundUnmarshallerProvider
 (kudu-client-1.11.1.jar)
 
org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.MarshallingDecoder
 (kudu-client-1.11.1.jar)
 
org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ThreadLocalUnmarshallerProvider
 (kudu-client-1.11.1.jar)
 
org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.MarshallingDecoder
 (beam-vendor-grpc-1_21_0-0.1.jar)
 
org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.CompatibleMarshallingDecoder
 (beam-vendor-grpc-1_21_0-0.1.jar)
 
org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ThreadLocalUnmarshallerProvider
 (beam-vendor-grpc-1_21_0-0.1.jar)
 
org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ContextBoundUnmarshallerProvider
 (beam-vendor-grpc-1_21_0-0.1.jar)
Class org.jboss.marshalling.Marshaller is not found;
 referenced by 6 class files
 
org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.CompatibleMarshallingEncoder
 (kudu-client-1.11.1.jar)
 
org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.MarshallingEncoder
 (kudu-client-1.11.1.jar)
 
org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ThreadLocalMarshallerProvider
 (kudu-client-1.11.1.jar)
 
org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.CompatibleMarshallingEncoder
 (beam-vendor-grpc-1_21_0-0.1.jar)
 
org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.MarshallingEncoder
 (beam-vendor-grpc-1_21_0-0.1.jar)
 
org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ThreadLocalMarshallerProvider
 (beam-vendor-grpc-1_21_0-0.1.jar)
Class org.jboss.marshalling.MarshallerFactory is not found;
 referenced by 8 class files
 
org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.DefaultMarshallerProvider
 (kudu-client-1.11.1.jar

[jira] [Commented] (KUDU-3041) Kudu Java client shade is incomplete

2020-01-14 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3041?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014973#comment-17014973
 ] 

Ismaël Mejía commented on KUDU-3041:


Please ignore the `beam-vendor-grpc-1_21_0-0.1.jar` those are issues on our own 
shading.

> Kudu Java client shade is incomplete
> 
>
> Key: KUDU-3041
> URL: https://issues.apache.org/jira/browse/KUDU-3041
> Project: Kudu
>  Issue Type: Bug
>  Components: client
>Affects Versions: 1.11.1
>Reporter: Ismaël Mejía
>Priority: Major
>
> While working on an update of the Kudu integration on Apache Beam BEAM-5086 
> We found this issue. We use [tool to test for linkage 
> errors|https://github.com/GoogleCloudPlatform/cloud-opensource-java] and it 
> reports the classes that are missing but required by other classes.
> This is the result for the kudu-client case:
> {code:java}
> Class javax.servlet.ServletOutputStream is not found;
>  referenced by 1 class file
>  
> org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet
>  (kudu-client-1.11.1.jar)
> Class javax.servlet.http.HttpServlet is not found;
>  referenced by 1 class file
>  
> org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet
>  (kudu-client-1.11.1.jar)
> Class javax.servlet.ServletException is not found;
>  referenced by 1 class file
>  
> org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet
>  (kudu-client-1.11.1.jar)
> Class javax.servlet.ServletConfig is not found;
>  referenced by 1 class file
>  
> org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet
>  (kudu-client-1.11.1.jar)
> Class javax.servlet.http.HttpServletRequest is not found;
>  referenced by 1 class file
>  
> org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet
>  (kudu-client-1.11.1.jar)
> Class javax.servlet.http.HttpServletResponse is not found;
>  referenced by 1 class file
>  
> org.apache.kudu.shaded.org.jboss.netty.channel.socket.http.HttpTunnelingServlet
>  (kudu-client-1.11.1.jar)
> Class org.jboss.marshalling.ByteInput is not found;
>  referenced by 4 class files
>  
> org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ChannelBufferByteInput
>  (kudu-client-1.11.1.jar)
>  
> org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.LimitingByteInput
>  (kudu-client-1.11.1.jar)
>  
> org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ChannelBufferByteInput
>  (beam-vendor-grpc-1_21_0-0.1.jar)
>  
> org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.LimitingByteInput
>  (beam-vendor-grpc-1_21_0-0.1.jar)
> Class org.jboss.marshalling.ByteOutput is not found;
>  referenced by 2 class files
>  
> org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ChannelBufferByteOutput
>  (kudu-client-1.11.1.jar)
>  
> org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ChannelBufferByteOutput
>  (beam-vendor-grpc-1_21_0-0.1.jar)
> Class org.jboss.marshalling.Unmarshaller is not found;
>  referenced by 8 class files
>  
> org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.CompatibleMarshallingDecoder
>  (kudu-client-1.11.1.jar)
>  
> org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ContextBoundUnmarshallerProvider
>  (kudu-client-1.11.1.jar)
>  
> org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.MarshallingDecoder
>  (kudu-client-1.11.1.jar)
>  
> org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.ThreadLocalUnmarshallerProvider
>  (kudu-client-1.11.1.jar)
>  
> org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.MarshallingDecoder
>  (beam-vendor-grpc-1_21_0-0.1.jar)
>  
> org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.CompatibleMarshallingDecoder
>  (beam-vendor-grpc-1_21_0-0.1.jar)
>  
> org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ThreadLocalUnmarshallerProvider
>  (beam-vendor-grpc-1_21_0-0.1.jar)
>  
> org.apache.beam.vendor.grpc.v1p21p0.io.netty.handler.codec.marshalling.ContextBoundUnmarshallerProvider
>  (beam-vendor-grpc-1_21_0-0.1.jar)
> Class org.jboss.marshalling.Marshaller is not found;
>  referenced by 6 class files
>  
> org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.CompatibleMarshallingEncoder
>  (kudu-client-1.11.1.jar)
>  
> org.apache.kudu.shaded.org.jboss.netty.handler.codec.marshalling.Mars

[jira] [Created] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20)

2020-07-26 Thread Jira
João Boto created KUDU-3174:
---

 Summary: KuduTestHarness not working on focal (ubuntu 20)
 Key: KUDU-3174
 URL: https://issues.apache.org/jira/browse/KUDU-3174
 Project: Kudu
  Issue Type: Bug
Affects Versions: 1.11.1, 1.12.0
Reporter: João Boto


when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting 
this error:
{code:java}
// code placeholder

java.io.EOFException
at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
at 
org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
at 
org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
tested with 1.11 and 1.12 on local and on travis (with focal dist)

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20)

2020-07-26 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

João Boto updated KUDU-3174:

Description: 
when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting 
this error:
{code:java}
// code placeholder

java.io.EOFException
at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
at 
org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
at 
org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
tested with 1.11 and 1.12 on local and on travis (with focal dist)

both with jdk8 and jdk11

 

  was:
when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting 
this error:
{code:java}
// code placeholder

java.io.EOFException
at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
at 
org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
at 
org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
tested with 1.11 and 1.12 on local and on travis (with focal dist)

 


> KuduTestHarness not working on focal (ubuntu 20)
> 
>
> Key: KUDU-3174
> URL: https://issues.apache.org/jira/browse/KUDU-3174
> Project: Kudu
>  Issue Type: Bug
>Affects Versions: 1.12.0, 1.11.1
>Reporter: João Boto
>Priority: Major
>
> when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im 
> getting this error:
> {code:java}
> // code placeholder
> java.io.EOFException
>   at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
>   at 
> org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
> tested with 1.11 and 1.12 on local and on travis (with focal dist)
> both with jdk8 and jdk11
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (KUDU-3175) .

2020-07-26 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3175?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

João Boto closed KUDU-3175.
---
Resolution: Invalid

> .
> -
>
> Key: KUDU-3175
> URL: https://issues.apache.org/jira/browse/KUDU-3175
> Project: Kudu
>  Issue Type: Improvement
>Reporter: João Boto
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (KUDU-3175) upgrade flink to 1.11.1

2020-07-26 Thread Jira
João Boto created KUDU-3175:
---

 Summary: upgrade flink to 1.11.1
 Key: KUDU-3175
 URL: https://issues.apache.org/jira/browse/KUDU-3175
 Project: Kudu
  Issue Type: Improvement
Reporter: João Boto






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (KUDU-3175) .

2020-07-26 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3175?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

João Boto updated KUDU-3175:

Summary: .  (was: upgrade flink to 1.11.1)

> .
> -
>
> Key: KUDU-3175
> URL: https://issues.apache.org/jira/browse/KUDU-3175
> Project: Kudu
>  Issue Type: Improvement
>Reporter: João Boto
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20)

2020-07-26 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

João Boto updated KUDU-3174:

Description: 
when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting 
this error:
{code:java}
java.io.EOFException
at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
at 
org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
at 
org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
tested with 1.11 and 1.12 on local and on travis (with focal dist)

both with jdk8 and jdk11

 

  was:
when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting 
this error:
{code:java}
// code placeholder

java.io.EOFException
at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
at 
org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
at 
org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
tested with 1.11 and 1.12 on local and on travis (with focal dist)

both with jdk8 and jdk11

 


> KuduTestHarness not working on focal (ubuntu 20)
> 
>
> Key: KUDU-3174
> URL: https://issues.apache.org/jira/browse/KUDU-3174
> Project: Kudu
>  Issue Type: Bug
>Affects Versions: 1.12.0, 1.11.1
>Reporter: João Boto
>Priority: Major
>
> when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im 
> getting this error:
> {code:java}
> java.io.EOFException
>   at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
>   at 
> org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
> tested with 1.11 and 1.12 on local and on travis (with focal dist)
> both with jdk8 and jdk11
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20 LTS)

2020-08-27 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

João Boto updated KUDU-3174:

Summary: KuduTestHarness not working on focal (ubuntu 20 LTS)  (was: 
KuduTestHarness not working on focal (ubuntu 20))

> KuduTestHarness not working on focal (ubuntu 20 LTS)
> 
>
> Key: KUDU-3174
> URL: https://issues.apache.org/jira/browse/KUDU-3174
> Project: Kudu
>  Issue Type: Bug
>Affects Versions: 1.12.0, 1.11.1
>Reporter: João Boto
>Priority: Major
>
> when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im 
> getting this error:
> {code:java}
> java.io.EOFException
>   at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
>   at 
> org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
> tested with 1.11 and 1.12 on local and on travis (with focal dist)
> both with jdk8 and jdk11
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20 LTS)

2020-08-27 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17185755#comment-17185755
 ] 

João Boto commented on KUDU-3174:
-

[~adar] maybe could you check

> KuduTestHarness not working on focal (ubuntu 20 LTS)
> 
>
> Key: KUDU-3174
> URL: https://issues.apache.org/jira/browse/KUDU-3174
> Project: Kudu
>  Issue Type: Bug
>Affects Versions: 1.12.0, 1.11.1
>Reporter: João Boto
>Priority: Major
>
> when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im 
> getting this error:
> {code:java}
> java.io.EOFException
>   at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
>   at 
> org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
> tested with 1.11 and 1.12 on local and on travis (with focal dist)
> both with jdk8 and jdk11
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (KUDU-3174) KuduTestHarness not working on focal (ubuntu 20 LTS)

2020-08-27 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

João Boto updated KUDU-3174:

Description: 
when executing tests with KuduTestHarness on ubuntu 20 LTS (focal fossa) Im 
getting this error:
{code:java}
java.io.EOFException
at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
at 
org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
at 
org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
tested with 1.11 and 1.12 on local and on travis (with focal dist)

both with jdk8 and jdk11

 

  was:
when executing tests with KuduTestHarness on ubuntu 20 (focal fossa) Im getting 
this error:
{code:java}
java.io.EOFException
at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
at 
org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
at 
org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
at 
org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
tested with 1.11 and 1.12 on local and on travis (with focal dist)

both with jdk8 and jdk11

 


> KuduTestHarness not working on focal (ubuntu 20 LTS)
> 
>
> Key: KUDU-3174
> URL: https://issues.apache.org/jira/browse/KUDU-3174
> Project: Kudu
>  Issue Type: Bug
>Affects Versions: 1.12.0, 1.11.1
>Reporter: João Boto
>Priority: Major
>
> when executing tests with KuduTestHarness on ubuntu 20 LTS (focal fossa) Im 
> getting this error:
> {code:java}
> java.io.EOFException
>   at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.sendRequestToCluster(MiniKuduCluster.java:162)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.start(MiniKuduCluster.java:235)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster.access$300(MiniKuduCluster.java:72)
>   at 
> org.apache.kudu.test.cluster.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:697)
>   at 
> org.apache.kudu.test.KuduTestHarness.before(KuduTestHarness.java:140){code}
> tested with 1.11 and 1.12 on local and on travis (with focal dist)
> both with jdk8 and jdk11
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (KUDU-1698) Kudu C++ client: add a new unit test to make sure default_rpc_timeout and session timeout are separate entities

2022-10-18 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-1698?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-1698:


Assignee: Bakai Ádám

> Kudu C++ client: add a new unit test to make sure default_rpc_timeout and 
> session timeout are separate entities
> ---
>
> Key: KUDU-1698
> URL: https://issues.apache.org/jira/browse/KUDU-1698
> Project: Kudu
>  Issue Type: Task
>  Components: client, test
>Reporter: Alexey Serbin
>Assignee: Bakai Ádám
>Priority: Minor
>  Labels: newbie
>
> We need a new unit test that makes sure there is a difference between 
> top-level operation timeout and per-call RPC timeout in Kudu C++ client 
> library.  Prior to change introduced in 
> 5195ce573850653e0e53094cdd35a1da93d33444 it was the same (which was a bug).
> The test should:
> * set  per-call RPC timeout when creating KuduClient object
> * set KuduSession::SetTimeoutMillis() for the target session: the value 
> should be 2 times of per-call RPC timeout or such.
> * create a tablet with replication factor of 2 at least.
> * find current tablet replica leader and pause it (send SIGSTOP)
> * make a write into the table
> * make sure the write operation was successful
> Prior to change introduced in 5195ce573850653e0e53094cdd35a1da93d33444 such a 
> test would fail because the C++ client used the full operation deadline on 
> every RPC call.
> I.e., it would wait till the call to current leader times out, and that would 
> consume time budget of the whole operation.  Once RPC timeout is less thatn 
> the timeout for the whole write operation, the call to the frozen tablet 
> server should timeout, and the client should re-discover a new tablet 
> replicate leader and complete the write operation successfully.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3414) Client tries with the same address over and over in case of negotiation timeout

2022-10-26 Thread Jira
Bakai Ádám created KUDU-3414:


 Summary: Client tries with the same address over and over in case 
of negotiation timeout
 Key: KUDU-3414
 URL: https://issues.apache.org/jira/browse/KUDU-3414
 Project: Kudu
  Issue Type: Bug
  Components: client
Reporter: Bakai Ádám


Reproduction:
1. Start new master and tablet servers.
2. Create a client and create a table and do a succesfull write operation with 
it.

3. Pause the leader tablet server.
4. Now if you attempt to write with the same client, it will timeout, and it 
will try negotiation with the same server until the session times out. (If a 
new client is created, then the write operation is succesful.)

KUDU-1580  works on a similar issue, the error is retriable for now, but 
[https://gerrit.cloudera.org/#/c/6926/] states that "the Kudu C++ client 
retries an RPC with other tablet replica if the connection negotiation with 
current replica timed out." which doesn't happen right now. (Didn't check if 
the current behaviour is a regression or not.)

Here is a reproduction test code into client-test.cc:
{code:java}
TEST_F(ClientTest, TestDefaultRPCTimeoutSessionTimeoutDifferent) {
cluster_->Shutdown();
FLAGS_raft_heartbeat_interval_ms = 50;
FLAGS_leader_failure_exp_backoff_max_delta_ms = 1000;

const string table_name = "TestTimeout";

unique_ptr ext_cluster_;
ext_cluster_.reset(new ExternalMiniCluster( ExternalMiniClusterOptions()));

ext_cluster_->Start();

ext_cluster_->AddMaster();

ext_cluster_->AddTabletServer();
ext_cluster_->AddTabletServer();
ext_cluster_->AddTabletServer();
ext_cluster_->AddTabletServer();
ext_cluster_->AddTabletServer();
ext_cluster_->WaitForTabletServerCount(5,MonoDelta::FromSeconds(1));


client::sp::shared_ptr client;

client_= nullptr;
KuduClientBuilder builder;
builder.num_reactors(1);

builder.connection_negotiation_timeout(MonoDelta::FromMilliseconds(1500));
builder.default_rpc_timeout(MonoDelta::FromMilliseconds(500));

ext_cluster_->CreateClient(&builder,&client);
unique_ptr table_creator(client->NewTableCreator());

table_creator->table_name(table_name)
.schema(&schema_)
.num_replicas(5)
.set_range_partition_columns({ "key" })
.timeout(MonoDelta::FromSeconds(60))
.Create();

shared_ptr table;
client->OpenTable(table_name,&table);


unique_ptr insert;
shared_ptr session = client->NewSession();
session->SetTimeoutMillis(6);


insert = BuildTestInsert(table.get(), 1);
ASSERT_OK(session->Apply(insert.release()));
session->Flush();

KuduScanTokenBuilder builder2(table.get());
vector tokens;
ASSERT_OK(builder2.Build(&tokens));

const KuduTablet& tablet = tokens[0]->tablet();
for (const auto& a: tablet.replicas())
{

if (a->is_leader())
{
ext_cluster_->tablet_server_by_uuid(a->ts().uuid())->Pause();
//ext_cluster_->tablet_server_by_uuid(a->ts().uuid())->Shutdown();
break;
}
}
LOG(INFO)<<"paused";
//ext_cluster_->CreateClient(&builder,&client);


session = client->NewSession();
session->SetTimeoutMillis(6);
insert = BuildTestInsert(table.get(), 2);

ASSERT_OK(session->Apply(insert.release()));
}
 {code}

(Don't forget to add the declarations and includes to the start of the file:)
{code:java}
#include "kudu/mini-cluster/external_mini_cluster.h"
using kudu::cluster::ExternalMiniCluster;
using kudu::cluster::ExternalMiniClusterOptions;
 {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3414) Client tries with the same address over and over in case of negotiation timeout

2022-10-26 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3414?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3414:
-
Description: 
Reproduction:
1. Start new master and tablet servers.
2. Create a client and create a table and do a succesfull write operation with 
it.

3. Pause the leader tablet server.
4. Now if you attempt to write with the same client, it will timeout, and it 
will try negotiation with the same server until the session times out. (If a 
new client is created, then the write operation is succesful.)

KUDU-1580  works on a similar issue, the error is retriable for now, but 
[https://gerrit.cloudera.org/#/c/6926/] states that "the Kudu C++ client 
retries an RPC with other tablet replica if the connection negotiation with 
current replica timed out." which doesn't happen right now. (Didn't check if 
the current behaviour is a regression or not.)

Here is a reproduction test code into client-test.cc:
{code:java}
TEST_F(ClientTest, TestDefaultRPCTimeoutSessionTimeoutDifferent) {
cluster_->Shutdown();
FLAGS_raft_heartbeat_interval_ms = 50;
FLAGS_leader_failure_exp_backoff_max_delta_ms = 1000;

const string table_name = "TestTimeout";

unique_ptr ext_cluster_;
ext_cluster_.reset(new ExternalMiniCluster( ExternalMiniClusterOptions()));

ext_cluster_->Start();

ext_cluster_->AddMaster();

ext_cluster_->AddTabletServer();
ext_cluster_->AddTabletServer();
ext_cluster_->AddTabletServer();
ext_cluster_->AddTabletServer();
ext_cluster_->AddTabletServer();
ext_cluster_->WaitForTabletServerCount(5,MonoDelta::FromSeconds(1));


client::sp::shared_ptr client;

client_= nullptr;
KuduClientBuilder builder;
builder.num_reactors(1);

builder.connection_negotiation_timeout(MonoDelta::FromMilliseconds(1500));
builder.default_rpc_timeout(MonoDelta::FromMilliseconds(500));

ext_cluster_->CreateClient(&builder,&client);
unique_ptr table_creator(client->NewTableCreator());

table_creator->table_name(table_name)
.schema(&schema_)
.num_replicas(5)
.set_range_partition_columns({ "key" })
.timeout(MonoDelta::FromSeconds(60))
.Create();

shared_ptr table;
client->OpenTable(table_name,&table);


unique_ptr insert;
shared_ptr session = client->NewSession();
session->SetTimeoutMillis(6);


insert = BuildTestInsert(table.get(), 1);
ASSERT_OK(session->Apply(insert.release()));
session->Flush();

KuduScanTokenBuilder builder2(table.get());
vector tokens;
ASSERT_OK(builder2.Build(&tokens));

const KuduTablet& tablet = tokens[0]->tablet();
for (const auto& a: tablet.replicas())
{

if (a->is_leader())
{
ext_cluster_->tablet_server_by_uuid(a->ts().uuid())->Pause();
//ext_cluster_->tablet_server_by_uuid(a->ts().uuid())->Shutdown();
break;
}
}
LOG(INFO)<<"paused";
//ext_cluster_->CreateClient(&builder,&client);


session = client->NewSession();
session->SetTimeoutMillis(6);
insert = BuildTestInsert(table.get(), 2);

ASSERT_OK(session->Apply(insert.release()));
}
 {code}
(Don't forget to add the declarations and includes to the start of the 
[file:|file:///])
{code:java}
#include "kudu/mini-cluster/external_mini_cluster.h"
using kudu::cluster::ExternalMiniCluster;
using kudu::cluster::ExternalMiniClusterOptions;
 {code}
The logic about deciding when to try different tablet is in: 

RetriableRpcStatus WriteRpc::AnalyzeResponse(const Status& rpc_cb_status) 

  was:
Reproduction:
1. Start new master and tablet servers.
2. Create a client and create a table and do a succesfull write operation with 
it.

3. Pause the leader tablet server.
4. Now if you attempt to write with the same client, it will timeout, and it 
will try negotiation with the same server until the session times out. (If a 
new client is created, then the write operation is succesful.)

KUDU-1580  works on a similar issue, the error is retriable for now, but 
[https://gerrit.cloudera.org/#/c/6926/] states that "the Kudu C++ client 
retries an RPC with other tablet replica if the connection negotiation with 
current replica timed out." which doesn't happen right now. (Didn't check if 
the current behaviour is a regression or not.)

Here is a reproduction test code into client-test.cc:
{code:java}
TEST_F(ClientTest, TestDefaultRPCTimeoutSessionTimeoutDifferent) {
cluster_->Shutdown();
FLAGS_raft_heartbeat_interval_ms = 50;
FLAGS_leader_failure_exp_backoff_max_delta_ms = 1000;

const string table_name = "TestTimeout";

unique_ptr ext_cluster_;

[jira] [Commented] (KUDU-1698) Kudu C++ client: add a new unit test to make sure default_rpc_timeout and session timeout are separate entities

2022-10-28 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-1698?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17625692#comment-17625692
 ] 

Bakai Ádám commented on KUDU-1698:
--

I tried to recreate the exact steps in a test, but it failed, because the 
client didn't rediscover, but rather tried over and over again until session 
timeout.  I talked with [~aserbin] and we came to the conclusion to create a 
new issue for the not rediscovering behaviour, and test the session and rpc 
timeout in an ohter way. The new issue is: KUDU-3414 . 
The new idea to test the separate entity property:
* Make the tablet lookup artificially slow by adding latency.
* See that the rpc is timing out but retries.
* Remove the artifical delay 
* Check that the operation was succesful in the end, and tablet look up 
happened twice.

> Kudu C++ client: add a new unit test to make sure default_rpc_timeout and 
> session timeout are separate entities
> ---
>
> Key: KUDU-1698
> URL: https://issues.apache.org/jira/browse/KUDU-1698
> Project: Kudu
>  Issue Type: Task
>  Components: client, test
>Reporter: Alexey Serbin
>Assignee: Bakai Ádám
>Priority: Minor
>  Labels: newbie
>
> We need a new unit test that makes sure there is a difference between 
> top-level operation timeout and per-call RPC timeout in Kudu C++ client 
> library.  Prior to change introduced in 
> 5195ce573850653e0e53094cdd35a1da93d33444 it was the same (which was a bug).
> The test should:
> * set  per-call RPC timeout when creating KuduClient object
> * set KuduSession::SetTimeoutMillis() for the target session: the value 
> should be 2 times of per-call RPC timeout or such.
> * create a tablet with replication factor of 2 at least.
> * find current tablet replica leader and pause it (send SIGSTOP)
> * make a write into the table
> * make sure the write operation was successful
> Prior to change introduced in 5195ce573850653e0e53094cdd35a1da93d33444 such a 
> test would fail because the C++ client used the full operation deadline on 
> every RPC call.
> I.e., it would wait till the call to current leader times out, and that would 
> consume time budget of the whole operation.  Once RPC timeout is less thatn 
> the timeout for the whole write operation, the call to the frozen tablet 
> server should timeout, and the client should re-discover a new tablet 
> replicate leader and complete the write operation successfully.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (KUDU-3387) Enable TCP keepalive for connections created by Java client

2022-11-04 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3387?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17628948#comment-17628948
 ] 

Bakai Ádám commented on KUDU-3387:
--

In the commit 57dda5d4868d29f68de4aa0ac516ca390333e6be , the possibility was 
explored, and the timeout is enabled, but the timeout is not set:
AsyncKuduClient.java:~3030
{code:java}
+  // Unfortunately there is no way to override the keep-alive timeout in
+  // Java since the JRE doesn't expose any way to call setsockopt() with
+  // TCP_KEEPIDLE. And of course the default timeout is >2h. Sigh.
+  b.option(ChannelOption.SO_KEEPALIVE, true);
{code}
As far as I understand netty doesn't expose the Socket itself, but since then 
it is possible to change the timeout interval on some platform with 
ExtendedSocketOptions 
(https://docs.oracle.com/en/java/javase/11/docs/api/jdk.net/jdk/net/ExtendedSocketOptions.html).
  Here is an example how to set it: 
https://stackoverflow.com/questions/22472844/how-to-set-socket-option-tcp-keepcnt-tcp-keepidle-tcp-keepintvl-in-java-or-n

So in theory It may be possible to set options on the SocketChannel (on Mac and 
Linux, source: https://bugs.openjdk.org/browse/JDK-8194298) , but not with 
io.netty.channel.socket.nio.NioSocketChannel
. These are the options that can be configured: 
https://javadoc.io/static/io.netty/netty-transport/5.0.0.Alpha2/io/netty/channel/socket/SocketChannelConfig.html
 . This is the highest version of netty I could find at the date of the comment.

So, in short: the TCP keepalive is set to true already, but the timeout 
interval is not explicitly set, and it can be very high (2 hours). Now Java 
support changing the keepalive time on Mac and Linux(keep in mind the current 
solution is Linux-only (Socket::SetTcpKeepAlive)), but netty doesn't support 
it, so it can not be changed. 

> Enable TCP keepalive for connections created by Java client
> ---
>
> Key: KUDU-3387
>     URL: https://issues.apache.org/jira/browse/KUDU-3387
> Project: Kudu
>  Issue Type: Improvement
>  Components: client, java
>Reporter: Alexey Serbin
>Priority: Major
>  Labels: newbie
>
> https://github.com/apache/kudu/commit/89c02fded7595b4712b465bfb939e4f3035b2e75
>  enabled TCP keepalive for outbound connections; and the C++ client since 
> then is able to detect a dropped off connection when server didn't send a TCP 
> RST.
> It's would be great to enable TCP keepalive for connections created by Java 
> client as well.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (KUDU-3300) Include the full path of the container in the error message

2022-11-07 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3300:


Assignee: Bakai Ádám

> Include the full path of the container in the error message
> ---
>
> Key: KUDU-3300
> URL: https://issues.apache.org/jira/browse/KUDU-3300
> Project: Kudu
>  Issue Type: Improvement
>  Components: cfile
>Reporter: Abhishek
>Assignee: Bakai Ádám
>Priority: Minor
>  Labels: newbie
>
> If there are multiple data directories configured, having the linux path to 
> the full container file will help to locate the file without having to search 
> for the file 
> Check failed: _s.ok() Bad status: Corruption: Failed to load FS layout: Could 
> not open container 26f5cbd97dfe4cb98f49bb0a6a494e8f: Invalid magic number: 
> Expected: kuducntr, found: \000\000\020\001\030▒▒▒



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-2981) Push predicate evaluation into more CFile decoders

2022-11-08 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-2981?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-2981:
-
Labels: performance  (was: newbie performance)

> Push predicate evaluation into more CFile decoders
> --
>
> Key: KUDU-2981
> URL: https://issues.apache.org/jira/browse/KUDU-2981
> Project: Kudu
>  Issue Type: Improvement
>  Components: cfile, perf
>Reporter: Bankim Bhavsar
>Priority: Major
>  Labels: performance
>
> Commit c0f3727 added an optimization to push predicate evaluation into the 
> CFile decoders without fully materializing the contents of each cblock. It 
> did this with dictionary-encoded blocks, but the optimization can be applied 
> to any other encoding types too.
> KUDU-736 also notes that we may be able to apply some predicates on 
> bitshuffled data.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3421) Make commane line error message more useful

2022-11-10 Thread Jira
Bakai Ádám created KUDU-3421:


 Summary: Make commane line error message more useful
 Key: KUDU-3421
 URL: https://issues.apache.org/jira/browse/KUDU-3421
 Project: Kudu
  Issue Type: Improvement
  Components: CLI
Reporter: Bakai Ádám


If the user runs this command for example:
{code:java}
/data/9/abakai/kudu/build/debug% 
/data/9/abakai/kudu/build/debug/bin/kudu-tserver 
--fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
--fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
--log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
--rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
--unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
--tserver_master_addrs=127.0.0.1:18764 
--webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
--memory_limit_hard_bytes=72050161527 -block_manager log  {code}
it works fine. But if the user adds "-logtostderr true", it will throw an error:
{code:java}
[abakai@va1022]/data/9/abakai/kudu/build/debug% 
/data/9/abakai/kudu/build/debug/bin/kudu-tserver 
--fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
--fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
--log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
--rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
--unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
--tserver_master_addrs=127.0.0.1:18764 
--webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
--memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true
usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code}
Which doesn't really help to identify which command line option is incorrect 
(and maybe how is it incorrect)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3421) Make commane line error message more useful

2022-11-10 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3421:
-
Description: 
If the user runs this command for example:
{code:java}
/data/9/abakai/kudu/build/debug% 
/data/9/abakai/kudu/build/debug/bin/kudu-tserver 
--fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
--fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
--log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
--rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
--unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
--tserver_master_addrs=127.0.0.1:18764 
--webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
--memory_limit_hard_bytes=72050161527 -block_manager log  {code}
it works fine. But if the user adds "-logtostderr true", it will throw an error:
{code:java}
[abakai@va1022]/data/9/abakai/kudu/build/debug% 
/data/9/abakai/kudu/build/debug/bin/kudu-tserver 
--fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
--fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
--log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
--rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
--unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
--tserver_master_addrs=127.0.0.1:18764 
--webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
--memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true
usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code}
Which doesn't really help to identify which command line option is incorrect 
(and maybe how is it incorrect)
(By the way the problem with the second command is that logtostderr doesn't 
need true or false option, it's existence indicates that logtostderr should be 
switched to true)

  was:
If the user runs this command for example:
{code:java}
/data/9/abakai/kudu/build/debug% 
/data/9/abakai/kudu/build/debug/bin/kudu-tserver 
--fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
--fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
--log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
--rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
--unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
--tserver_master_addrs=127.0.0.1:18764 
--webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
--memory_limit_hard_bytes=72050161527 -block_manager log  {code}
it works fine. But if the user adds "-logtostderr true", it will throw an error:
{code:java}
[abakai@va1022]/data/9/abakai/kudu/build/debug% 
/data/9/abakai/kudu/build/debug/bin/kudu-tserver 
--fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
--fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
--log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
--rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
--unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
--tserver_master_addrs=127.0.0.1:18764 
--webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
--memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true
usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code}
Which doesn't really help to identify which command line option is incorrect 
(and maybe how is it incorrect)


> Make commane line error message more useful
> ---
>
> Key: KUDU-3421
> URL: https://issues.apache.org/jira/browse/KUDU-3421
> Project: Kudu
>  Issue Type: Improvement
>  Components: CLI
>Reporter: Bakai Ádám
>Priority: Major
>
> If the user runs this command for example:
> {code:java}
> /data/9/abakai/kudu/build/debug% 
> /data/9/abakai/kudu/build/debug/bin/kudu-tserver 
> --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
> --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
> --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
> --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
> --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
> --tserver_master_addrs=127.0.0.1:18764 
> --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
> --memory_limit_hard_bytes=72050161527 -block_manager log  {code}
> it works fine. But if the user adds "-logtostderr true", it will throw an 
> error:
> {code:java}
> [abakai@va1022]/data/9/abakai/kudu/build/debug% 
> /data/9/abakai/kudu/build/debug/bin/kudu-tserver 
> --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
> --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
> --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
> --rpc_bind_addresses=127.0.0.1:19874 --t

[jira] [Commented] (KUDU-3210) Support FIPS approved mode

2022-11-10 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3210?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17631907#comment-17631907
 ] 

Bakai Ádám commented on KUDU-3210:
--

On OpenSSL 3, FIPS_mode is removed 
([https://wiki.openssl.org/index.php/OpenSSL_3.0#Upgrading_from_the_OpenSSL_2.0_FIPS_Object_Module)]
 and we use the system openssl, it may be a good idea to create a solution that 
works on both openssl versions.

> Support FIPS approved mode
> --
>
> Key: KUDU-3210
> URL: https://issues.apache.org/jira/browse/KUDU-3210
> Project: Kudu
>  Issue Type: Improvement
>Reporter: Attila Bukor
>Assignee: Attila Bukor
>Priority: Major
> Fix For: 1.14.0
>
>
> FIPS 140-2 is a standard used to approve cryptographic modules. Some versions 
> of OpenSSL support a "FIPS mode" where only approved algorithms and key sizes 
> are enabled. Kudu should be able to run when FIPS mode is enabled and should 
> provide a way for admins to require that FIPS mode is enabled.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3421) Make commane line error message more useful

2022-11-15 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3421:
-
Labels: newbie  (was: )

> Make commane line error message more useful
> ---
>
> Key: KUDU-3421
> URL: https://issues.apache.org/jira/browse/KUDU-3421
> Project: Kudu
>  Issue Type: Improvement
>  Components: CLI
>Reporter: Bakai Ádám
>Priority: Major
>  Labels: newbie
>
> If the user runs this command for example:
> {code:java}
> /data/9/abakai/kudu/build/debug% 
> /data/9/abakai/kudu/build/debug/bin/kudu-tserver 
> --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
> --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
> --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
> --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
> --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
> --tserver_master_addrs=127.0.0.1:18764 
> --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
> --memory_limit_hard_bytes=72050161527 -block_manager log  {code}
> it works fine. But if the user adds "-logtostderr true", it will throw an 
> error:
> {code:java}
> [abakai@va1022]/data/9/abakai/kudu/build/debug% 
> /data/9/abakai/kudu/build/debug/bin/kudu-tserver 
> --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
> --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
> --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
> --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
> --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
> --tserver_master_addrs=127.0.0.1:18764 
> --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
> --memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true
> usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code}
> Which doesn't really help to identify which command line option is incorrect 
> (and maybe how is it incorrect)
> (By the way the problem with the second command is that logtostderr doesn't 
> need true or false option, it's existence indicates that logtostderr should 
> be switched to true)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3421) Make command line error message more useful

2022-11-16 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3421:
-
Summary: Make command line error message more useful  (was: Make commane 
line error message more useful)

> Make command line error message more useful
> ---
>
> Key: KUDU-3421
> URL: https://issues.apache.org/jira/browse/KUDU-3421
> Project: Kudu
>  Issue Type: Improvement
>  Components: CLI
>Reporter: Bakai Ádám
>Priority: Major
>  Labels: newbie
>
> If the user runs this command for example:
> {code:java}
> /data/9/abakai/kudu/build/debug% 
> /data/9/abakai/kudu/build/debug/bin/kudu-tserver 
> --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
> --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
> --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
> --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
> --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
> --tserver_master_addrs=127.0.0.1:18764 
> --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
> --memory_limit_hard_bytes=72050161527 -block_manager log  {code}
> it works fine. But if the user adds "-logtostderr true", it will throw an 
> error:
> {code:java}
> [abakai@va1022]/data/9/abakai/kudu/build/debug% 
> /data/9/abakai/kudu/build/debug/bin/kudu-tserver 
> --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
> --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
> --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
> --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
> --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
> --tserver_master_addrs=127.0.0.1:18764 
> --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
> --memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true
> usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code}
> Which doesn't really help to identify which command line option is incorrect 
> (and maybe how is it incorrect)
> (By the way the problem with the second command is that logtostderr doesn't 
> need true or false option, it's existence indicates that logtostderr should 
> be switched to true)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (KUDU-3421) Make command line error message more useful

2022-11-28 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám closed KUDU-3421.


> Make command line error message more useful
> ---
>
> Key: KUDU-3421
> URL: https://issues.apache.org/jira/browse/KUDU-3421
> Project: Kudu
>  Issue Type: Improvement
>  Components: CLI
>Reporter: Bakai Ádám
>Priority: Major
>  Labels: newbie
> Fix For: 1.17.0
>
>
> If the user runs this command for example:
> {code:java}
> /data/9/abakai/kudu/build/debug% 
> /data/9/abakai/kudu/build/debug/bin/kudu-tserver 
> --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
> --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
> --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
> --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
> --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
> --tserver_master_addrs=127.0.0.1:18764 
> --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
> --memory_limit_hard_bytes=72050161527 -block_manager log  {code}
> it works fine. But if the user adds "-logtostderr true", it will throw an 
> error:
> {code:java}
> [abakai@va1022]/data/9/abakai/kudu/build/debug% 
> /data/9/abakai/kudu/build/debug/bin/kudu-tserver 
> --fs_data_dirs=/data/9/abakai/kudu/build/debug/tserver-2/data 
> --fs_wal_dir=/data/9/abakai/kudu/build/debug/tserver-2/wal 
> --log_dir=/data/9/abakai/kudu/build/debug/tserver-2/log 
> --rpc_bind_addresses=127.0.0.1:19874 --time_source=system_unsync 
> --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=19875 
> --tserver_master_addrs=127.0.0.1:18764 
> --webserver_doc_root=/data/9/abakai/kudu/build/debug/../../www 
> --memory_limit_hard_bytes=72050161527 -block_manager log -logtostderr true
> usage: /data/9/abakai/kudu/build/debug/bin/kudu-tserver {code}
> Which doesn't really help to identify which command line option is incorrect 
> (and maybe how is it incorrect)
> (By the way the problem with the second command is that logtostderr doesn't 
> need true or false option, it's existence indicates that logtostderr should 
> be switched to true)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (KUDU-3300) Include the full path of the container in the error message

2022-11-28 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3300:


Assignee: Abhishek  (was: Bakai Ádám)

> Include the full path of the container in the error message
> ---
>
> Key: KUDU-3300
> URL: https://issues.apache.org/jira/browse/KUDU-3300
> Project: Kudu
>  Issue Type: Improvement
>  Components: cfile
>Reporter: Abhishek
>Assignee: Abhishek
>Priority: Minor
>  Labels: newbie
> Fix For: 1.17.0
>
>
> If there are multiple data directories configured, having the linux path to 
> the full container file will help to locate the file without having to search 
> for the file 
> Check failed: _s.ok() Bad status: Corruption: Failed to load FS layout: Could 
> not open container 26f5cbd97dfe4cb98f49bb0a6a494e8f: Invalid magic number: 
> Expected: kuducntr, found: \000\000\020\001\030▒▒▒



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (KUDU-3300) Include the full path of the container in the error message

2022-11-28 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám resolved KUDU-3300.
--
Fix Version/s: 1.17.0
   Resolution: Fixed

> Include the full path of the container in the error message
> ---
>
> Key: KUDU-3300
> URL: https://issues.apache.org/jira/browse/KUDU-3300
> Project: Kudu
>  Issue Type: Improvement
>  Components: cfile
>Reporter: Abhishek
>Assignee: Bakai Ádám
>Priority: Minor
>  Labels: newbie
> Fix For: 1.17.0
>
>
> If there are multiple data directories configured, having the linux path to 
> the full container file will help to locate the file without having to search 
> for the file 
> Check failed: _s.ok() Bad status: Corruption: Failed to load FS layout: Could 
> not open container 26f5cbd97dfe4cb98f49bb0a6a494e8f: Invalid magic number: 
> Expected: kuducntr, found: \000\000\020\001\030▒▒▒



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (KUDU-3292) Show non-default flags on varz Web UI

2023-01-05 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3292:


Assignee: Bakai Ádám  (was: Zoltan Chovan)

> Show non-default flags on varz Web UI
> -
>
> Key: KUDU-3292
> URL: https://issues.apache.org/jira/browse/KUDU-3292
> Project: Kudu
>  Issue Type: Improvement
>  Components: ui
>Reporter: Grant Henke
>Assignee: Bakai Ádám
>Priority: Minor
>  Labels: beginner, newbie, newbie++, trivial
>
> Currently each Kudu server has a /varz webpage (the Flags tab) showing all of 
> the flags set on the server. It would be a nice usability change to include a 
> seperate section showing only the non-default flags. This should be super 
> straigtforward given we have the ability to get all the non-default flags via 
> GetNonDefaultFlags or GetNonDefaultFlagsMap in flags.cc.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (KUDU-1698) Kudu C++ client: add a new unit test to make sure default_rpc_timeout and session timeout are separate entities

2023-01-10 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-1698?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám resolved KUDU-1698.
--
Fix Version/s: n/a
   Resolution: Fixed

> Kudu C++ client: add a new unit test to make sure default_rpc_timeout and 
> session timeout are separate entities
> ---
>
> Key: KUDU-1698
> URL: https://issues.apache.org/jira/browse/KUDU-1698
> Project: Kudu
>  Issue Type: Task
>  Components: client, test
>Reporter: Alexey Serbin
>Assignee: Bakai Ádám
>Priority: Minor
>  Labels: newbie
> Fix For: n/a
>
>
> We need a new unit test that makes sure there is a difference between 
> top-level operation timeout and per-call RPC timeout in Kudu C++ client 
> library.  Prior to change introduced in 
> 5195ce573850653e0e53094cdd35a1da93d33444 it was the same (which was a bug).
> The test should:
> * set  per-call RPC timeout when creating KuduClient object
> * set KuduSession::SetTimeoutMillis() for the target session: the value 
> should be 2 times of per-call RPC timeout or such.
> * create a tablet with replication factor of 2 at least.
> * find current tablet replica leader and pause it (send SIGSTOP)
> * make a write into the table
> * make sure the write operation was successful
> Prior to change introduced in 5195ce573850653e0e53094cdd35a1da93d33444 such a 
> test would fail because the C++ client used the full operation deadline on 
> every RPC call.
> I.e., it would wait till the call to current leader times out, and that would 
> consume time budget of the whole operation.  Once RPC timeout is less thatn 
> the timeout for the whole write operation, the call to the frozen tablet 
> server should timeout, and the client should re-discover a new tablet 
> replicate leader and complete the write operation successfully.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (KUDU-3292) Show non-default flags on varz Web UI

2023-01-11 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3292?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17674833#comment-17674833
 ] 

Bakai Ádám commented on KUDU-3292:
--

[~laiyingchun]  I'm not familiar with impala, and you are right, it looks more 
informative, but refactoring it is a way bigger task then what this issue was 
about originally.  

> Show non-default flags on varz Web UI
> -
>
> Key: KUDU-3292
> URL: https://issues.apache.org/jira/browse/KUDU-3292
> Project: Kudu
>  Issue Type: Improvement
>  Components: ui
>Reporter: Grant Henke
>Assignee: Bakai Ádám
>Priority: Minor
>  Labels: beginner, newbie, newbie++, trivial
> Attachments: image-2023-01-10-11-57-13-209.png
>
>
> Currently each Kudu server has a /varz webpage (the Flags tab) showing all of 
> the flags set on the server. It would be a nice usability change to include a 
> seperate section showing only the non-default flags. This should be super 
> straigtforward given we have the ability to get all the non-default flags via 
> GetNonDefaultFlags or GetNonDefaultFlagsMap in flags.cc.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (KUDU-3292) Show non-default flags on varz Web UI

2023-01-11 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám resolved KUDU-3292.
--
Fix Version/s: n/a
   Resolution: Fixed

> Show non-default flags on varz Web UI
> -
>
> Key: KUDU-3292
> URL: https://issues.apache.org/jira/browse/KUDU-3292
> Project: Kudu
>  Issue Type: Improvement
>  Components: ui
>Reporter: Grant Henke
>Assignee: Bakai Ádám
>Priority: Minor
>  Labels: beginner, newbie, newbie++, trivial
> Fix For: n/a
>
> Attachments: image-2023-01-10-11-57-13-209.png
>
>
> Currently each Kudu server has a /varz webpage (the Flags tab) showing all of 
> the flags set on the server. It would be a nice usability change to include a 
> seperate section showing only the non-default flags. This should be super 
> straigtforward given we have the ability to get all the non-default flags via 
> GetNonDefaultFlags or GetNonDefaultFlagsMap in flags.cc.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (KUDU-3431) There is no binary package suitable for osx-aarch

2023-01-16 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3431?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17677407#comment-17677407
 ] 

Bakai Ádám commented on KUDU-3431:
--

[~healchow] I created a short documentary review request about how to run the 
java examples on M1 Mac, it may help: [https://gerrit.cloudera.org/#/c/19418/] 
If it didn't help in your case, please describe it more in details where it 
causes problem that the MAC ARM64 jar does not exist. 

> There is no binary package suitable for osx-aarch
> -
>
> Key: KUDU-3431
> URL: https://issues.apache.org/jira/browse/KUDU-3431
> Project: Kudu
>  Issue Type: Bug
>Affects Versions: 1.15.0, 1.16.0
> Environment: MacBook M1 chip, OS Ventura 13.1
>Reporter: Heal Chow
>Priority: Major
> Attachments: image-2022-12-29-20-43-52-324.png
>
>
> My development machine is MacBook of Apple M1 chip. During the development 
> process, I found that kudu binary does not adapt to osx aarch_ 64 
> architecture.
>  
> In this repository, I didn't find what I wanted.
> [https://repo.maven.apache.org/maven2/org/apache/kudu/kudu-binary/1.16.0/]
> !image-2022-12-29-20-43-52-324.png|width=577,height=287!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3434) Couldn't close scanner : Service unavailable: reactor is shutting down (error 58) warning at big scans

2023-01-17 Thread Jira
Bakai Ádám created KUDU-3434:


 Summary: Couldn't close scanner : Service unavailable: reactor is 
shutting down (error 58) warning at big scans
 Key: KUDU-3434
 URL: https://issues.apache.org/jira/browse/KUDU-3434
 Project: Kudu
  Issue Type: Bug
  Components: client
 Environment: MAC Arm64 development environment
Reporter: Bakai Ádám


I ran 
{code:java}
./bin/kudu table scan  localhost:8764  test_table {code}
The test_table has 372000 and the command takes about 1.90579 seconds to 
execute. 
Most of the time it ends like this:
{code:java}
...(rows)...
(int32 key=371998, int32 integer_val=743996, int32 
non_null_with_default=1859990, bool another_val=NULL)
(int32 key=371999, int32 integer_val=743998, int32 
non_null_with_default=1859995, bool another_val=NULL)
Total count 372000 cost 1.90579 seconds
W20230117 13:23:30.705157 1238496 client.cc:1955] Couldn't close scanner 
a92a2d1cc6394cc3837d67286ce25575: Service unavailable: reactor is shutting down 
(error 58) {code}
If I run with the java-example test table(150 rows, around 0.003 seconds to 
execute), it never shows the warning.

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3436) build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1 ARM64

2023-01-23 Thread Jira
Bakai Ádám created KUDU-3436:


 Summary: build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1 
ARM64
 Key: KUDU-3436
 URL: https://issues.apache.org/jira/browse/KUDU-3436
 Project: Kudu
  Issue Type: Bug
Reporter: Bakai Ádám


 
{code:java}
build_mini_cluster_binaries.sh {code}
returns the following error:
{code:java}
Traceback (most recent call last):
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 503, in 
main()
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 500, in main
relocate_deps(target_src, target_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 408, in relocate_deps
return relocate_deps_macos(target_src, target_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 397, in relocate_deps_macos
copy_file(dep_src, dep_dst)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 287, in copy_file
shutil.copyfile(src, dest)
  File 
"/opt/homebrew/Cellar/python@2/2.7.18/Frameworks/Python.framework/Versions/2.7/lib/python2.7/shutil.py",
 line 96, in copyfile
with open(src, 'rb') as fsrc:
IOError: [Errno 2] No such file or directory: u'/usr/lib/libc++abi.dylib' {code}
After further investigation, it looks like libc++abi.dylib is in the 
uninstrumented lib, but otool -L always gives back a path for 
/usr/lib/libc++abi.dylib . Simply adding the dylib into the 
PAT_MACOS_LIB_EXCLUDE list doesn't work: it creates a jar file, but the 
binaries can not be started.
It is probably due to the changes in how dynamic linking works in newer MacOS: 
https://stackoverflow.com/questions/70581876/macos-dynamic-linker-reports-it-loaded-library-which-doesnt-exist

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3436) build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1

2023-01-24 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3436?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3436:
-
Description: 
 
{code:java}
build_mini_cluster_binaries.sh {code}
returns the following error:
{code:java}
Traceback (most recent call last):
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 503, in 
main()
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 500, in main
relocate_deps(target_src, target_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 408, in relocate_deps
return relocate_deps_macos(target_src, target_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 397, in relocate_deps_macos
copy_file(dep_src, dep_dst)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 287, in copy_file
shutil.copyfile(src, dest)
  File 
"/opt/homebrew/Cellar/python@2/2.7.18/Frameworks/Python.framework/Versions/2.7/lib/python2.7/shutil.py",
 line 96, in copyfile
with open(src, 'rb') as fsrc:
IOError: [Errno 2] No such file or directory: u'/usr/lib/libc++abi.dylib' {code}
After further investigation, it looks like libc+{+}abi.dylib is in the 
uninstrumented lib, but otool -L always gives back a path for 
/usr/lib/libc{+}+abi.dylib . Simply adding the dylib into the 
PAT_MACOS_LIB_EXCLUDE list doesn't work: it creates a jar file, but the 
binaries can not be started.
It is probably due to the changes in how dynamic linking works in newer MacOS: 
[https://stackoverflow.com/questions/70581876/macos-dynamic-linker-reports-it-loaded-library-which-doesnt-exist]

It happens both on ARM64 and X86

  was:
 
{code:java}
build_mini_cluster_binaries.sh {code}
returns the following error:
{code:java}
Traceback (most recent call last):
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 503, in 
main()
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 500, in main
relocate_deps(target_src, target_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 408, in relocate_deps
return relocate_deps_macos(target_src, target_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 398, in relocate_deps_macos
relocate_deps_macos(dep_src, dep_dst, config)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 397, in relocate_deps_macos
copy_file(dep_src, dep_dst)
  File 
"/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
 line 287, in copy_file
shutil.copyfile(src, dest)
  File 
"/opt/homebrew/Cellar/python@2/2.7.18/Frameworks/Python.framework/Versions/2.7/lib/python2.7/shutil.py",
 line 96, in copyfile
with open(src, 'rb') as fsrc:
IOError: [Errno 2] No such file or dir

[jira] [Updated] (KUDU-3436) build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1

2023-01-24 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3436?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3436:
-
Summary: build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1  (was: 
build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1 ARM64)

> build_mini_cluster_binaries.sh doesn't work on Mac 13.0.1
> -
>
> Key: KUDU-3436
> URL: https://issues.apache.org/jira/browse/KUDU-3436
> Project: Kudu
>  Issue Type: Bug
>Reporter: Bakai Ádám
>Priority: Major
>
>  
> {code:java}
> build_mini_cluster_binaries.sh {code}
> returns the following error:
> {code:java}
> Traceback (most recent call last):
>   File 
> "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
>  line 503, in 
> main()
>   File 
> "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
>  line 500, in main
> relocate_deps(target_src, target_dst, config)
>   File 
> "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
>  line 408, in relocate_deps
> return relocate_deps_macos(target_src, target_dst, config)
>   File 
> "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
>  line 398, in relocate_deps_macos
> relocate_deps_macos(dep_src, dep_dst, config)
>   File 
> "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
>  line 398, in relocate_deps_macos
> relocate_deps_macos(dep_src, dep_dst, config)
>   File 
> "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
>  line 398, in relocate_deps_macos
> relocate_deps_macos(dep_src, dep_dst, config)
>   File 
> "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
>  line 398, in relocate_deps_macos
> relocate_deps_macos(dep_src, dep_dst, config)
>   File 
> "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
>  line 397, in relocate_deps_macos
> copy_file(dep_src, dep_dst)
>   File 
> "/Users/adambakai/CLionProjects/kudu/build-support/mini-cluster/relocate_binaries_for_mini_cluster.py",
>  line 287, in copy_file
> shutil.copyfile(src, dest)
>   File 
> "/opt/homebrew/Cellar/python@2/2.7.18/Frameworks/Python.framework/Versions/2.7/lib/python2.7/shutil.py",
>  line 96, in copyfile
> with open(src, 'rb') as fsrc:
> IOError: [Errno 2] No such file or directory: u'/usr/lib/libc++abi.dylib' 
> {code}
> After further investigation, it looks like libc++abi.dylib is in the 
> uninstrumented lib, but otool -L always gives back a path for 
> /usr/lib/libc++abi.dylib . Simply adding the dylib into the 
> PAT_MACOS_LIB_EXCLUDE list doesn't work: it creates a jar file, but the 
> binaries can not be started.
> It is probably due to the changes in how dynamic linking works in newer 
> MacOS: 
> https://stackoverflow.com/questions/70581876/macos-dynamic-linker-reports-it-loaded-library-which-doesnt-exist
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3437) Starting Master server runs into error first time

2023-01-27 Thread Jira
Bakai Ádám created KUDU-3437:


 Summary: Starting Master server runs into error first time
 Key: KUDU-3437
 URL: https://issues.apache.org/jira/browse/KUDU-3437
 Project: Kudu
  Issue Type: Bug
  Components: cfile
 Environment: Macos 13.0.1 Arm64, fastdebug build
Reporter: Bakai Ádám


Reproduction script:
{code:java}
#Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in 
your path to run this script.
export PATH=$PATH:./bin/stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data 
master-3/wal master-3/log; start_kudu.sh -m 3echo just to make sure it returns 
three masters as expected
kudu master list localhost:8764,localhost:8766,localhost:8768echo This returns 
error:
kudu-master 
--master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 
--fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
--log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
--time_source=system_unsync --unlock_unsafe_flags 
--webserver_interface=127.0.0.1 --webserver_port=8771echo just to make sure it 
returns three masters
kudu master list localhost:8764,localhost:8766,localhost:8768
echo Now it should be running fine:
kudu-master 
--master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 
--fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
--log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
--time_source=system_unsync --unlock_unsafe_flags 
--webserver_interface=127.0.0.1 --webserver_port=8771 &sleep 2 
echo just to make sure it returns four masters
kudu master list localhost:8764,localhost:8766,localhost:8768
 {code}
My error for the first run:
{code:java}
F20230127 15:58:28.923543 360774 cache.cc:579] Check failed: IsGTest() Metrics 
should only be set once per Cache
*** Check failure stack trace: ***
    @        0x1025950d4  google::LogMessageFatal::~LogMessageFatal()
    @        0x102591930  google::LogMessageFatal::~LogMessageFatal()
    @        0x10b56c4a8  kudu::(anonymous 
namespace)::ShardedCache<>::SetMetrics()
    @        0x103522c1c  kudu::cfile::BlockCache::StartInstrumentation()
    @        0x101249230  kudu::master::Master::Init()
    @        0x10126cbf0  kudu::master::RunMasterServer()
    @        0x1005eae64  kudu::master::MasterMain()
    @        0x1005ead10  main {code}
 If I uncomment the line
{code:java}
CHECK(IsGTest()) << "Metrics should only be set once per Cache";
 {code}
from kudu/util/cache.cc:579  ( CHECK(IsGTest()) << "Metrics should only be set 
once per Cache"<

[jira] [Updated] (KUDU-3437) Starting Master server runs into error first time

2023-01-27 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3437:
-
Description: 
Reproduction script:
{noformat}
#Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in 
your path to run this script.
export PATH=$PATH:./bin/

stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data master-3/wal master-3/log; 
start_kudu.sh -m 3

echo just to make sure it returns three masters as expected
kudu master list localhost:8764,localhost:8766,localhost:8768

echo This returns error:
kudu-master 
--master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 
--fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
--log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
--time_source=system_unsync --unlock_unsafe_flags 
--webserver_interface=127.0.0.1 --webserver_port=8771

echo just to make sure it returns three masters
kudu master list localhost:8764,localhost:8766,localhost:8768


echo Now it should be running fine:
kudu-master 
--master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 
--fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
--log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
--time_source=system_unsync --unlock_unsafe_flags 
--webserver_interface=127.0.0.1 --webserver_port=8771 &

sleep 2 
echo just to make sure it returns four masters
kudu master list localhost:8764,localhost:8766,localhost:8768
 {noformat}
My error for the first run:
{code:java}
F20230127 15:58:28.923543 360774 cache.cc:579] Check failed: IsGTest() Metrics 
should only be set once per Cache
*** Check failure stack trace: ***
    @        0x1025950d4  google::LogMessageFatal::~LogMessageFatal()
    @        0x102591930  google::LogMessageFatal::~LogMessageFatal()
    @        0x10b56c4a8  kudu::(anonymous 
namespace)::ShardedCache<>::SetMetrics()
    @        0x103522c1c  kudu::cfile::BlockCache::StartInstrumentation()
    @        0x101249230  kudu::master::Master::Init()
    @        0x10126cbf0  kudu::master::RunMasterServer()
    @        0x1005eae64  kudu::master::MasterMain()
    @        0x1005ead10  main {code}
 If I uncomment the line
{code:java}
CHECK(IsGTest()) << "Metrics should only be set once per Cache";
 {code}
from kudu/util/cache.cc:579  ( CHECK(IsGTest()) << "Metrics should only be set 
once per Cache"<::SetMetrics()
    @        0x103522c1c  kudu::cfile::BlockCache::StartInstrumentation()
    @        0x101249230  kudu::master::Master::Init()
    @        0x10126cbf0  kudu::master::RunMasterServer()
    @        0x1005eae64  kudu::master::MasterMain()
    @        0x1005ead10  main {code}
 If I uncomment the line
{code:java}
CHECK(IsGTest()) << "Metrics should only be set once per Cache";
 {code}
from kudu/util/cache.cc:579  ( CHECK(IsGTest()) << "Metrics should only be set 
once per Cache"< Starting Master server runs into error first time
> -
>
> Key: KUDU-3437
> URL: https://issues.apache.org/jira/browse/KUDU-3437
> Project: Kudu
>  Issue Type: Bug
>  Components: cfile
> Environment: Macos 13.0.1 Arm64, fastdebug build
>Reporter: Bakai Ádám
>Priority: Major
>
> Reproduction script:
> {noformat}
> #Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in 
> your path to run this script.
> export PATH=$PATH:./bin/
> stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data master-3/wal 
> master-3/log; start_kudu.sh -m 3
> echo just to make sure it returns three masters as expected
> kudu master list localhost:8764,localhost:8766,localhost:8768
> echo This returns error:
> kudu-master 
> --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770
>  --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
> --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
> --time_source=system_unsync --unlock_unsafe_flags 
> --webserver_interface=127.0.0.1 --webserver_port=8771
> echo just to make sure it returns three masters
> kudu master list localhost:8764,localhost:8766,localhost:8768
> echo Now it should be running fine:
> kudu-master 
> --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770
>  --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
> --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
> --time_source=system_unsync --unlock_unsafe_flags 
> --webserver_interface=127.0.0.1 --webserver_port=8771 &
> sleep 2 
> echo just to make sure it returns four masters
> kudu master list localhost:8764,localhost:8766,localhost:8768
>  

[jira] [Updated] (KUDU-3437) Starting Master server runs into error first time

2023-01-27 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3437:
-
Description: 
Reproduction script:
{noformat}
#Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in 
your path to run this script.
export PATH=$PATH:./bin/

stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data master-3/wal master-3/log; 
start_kudu.sh -m 3

echo just to make sure it returns three masters as expected
kudu master list localhost:8764,localhost:8766,localhost:8768

echo This returns error:
kudu-master 
--master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 
--fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
--log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
--time_source=system_unsync --unlock_unsafe_flags 
--webserver_interface=127.0.0.1 --webserver_port=8771

echo just to make sure it returns three masters
kudu master list localhost:8764,localhost:8766,localhost:8768


echo Now it should be running fine:
kudu-master 
--master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 
--fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
--log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
--time_source=system_unsync --unlock_unsafe_flags 
--webserver_interface=127.0.0.1 --webserver_port=8771 &

sleep 2 
echo just to make sure it returns four masters
kudu master list localhost:8764,localhost:8766,localhost:8768
 {noformat}
My error for the first run:
{code:java}
F20230127 15:58:28.923543 360774 cache.cc:579] Check failed: IsGTest() Metrics 
should only be set once per Cache
*** Check failure stack trace: ***
    @        0x1025950d4  google::LogMessageFatal::~LogMessageFatal()
    @        0x102591930  google::LogMessageFatal::~LogMessageFatal()
    @        0x10b56c4a8  kudu::(anonymous 
namespace)::ShardedCache<>::SetMetrics()
    @        0x103522c1c  kudu::cfile::BlockCache::StartInstrumentation()
    @        0x101249230  kudu::master::Master::Init()
    @        0x10126cbf0  kudu::master::RunMasterServer()
    @        0x1005eae64  kudu::master::MasterMain()
    @        0x1005ead10  main {code}
 If I uncomment the line
{code:java}
CHECK(IsGTest()) << "Metrics should only be set once per Cache";
 {code}
from kudu/util/cache.cc:579  it seems to work fine.

  was:
Reproduction script:
{noformat}
#Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in 
your path to run this script.
export PATH=$PATH:./bin/

stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data master-3/wal master-3/log; 
start_kudu.sh -m 3

echo just to make sure it returns three masters as expected
kudu master list localhost:8764,localhost:8766,localhost:8768

echo This returns error:
kudu-master 
--master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 
--fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
--log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
--time_source=system_unsync --unlock_unsafe_flags 
--webserver_interface=127.0.0.1 --webserver_port=8771

echo just to make sure it returns three masters
kudu master list localhost:8764,localhost:8766,localhost:8768


echo Now it should be running fine:
kudu-master 
--master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770 
--fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
--log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
--time_source=system_unsync --unlock_unsafe_flags 
--webserver_interface=127.0.0.1 --webserver_port=8771 &

sleep 2 
echo just to make sure it returns four masters
kudu master list localhost:8764,localhost:8766,localhost:8768
 {noformat}
My error for the first run:
{code:java}
F20230127 15:58:28.923543 360774 cache.cc:579] Check failed: IsGTest() Metrics 
should only be set once per Cache
*** Check failure stack trace: ***
    @        0x1025950d4  google::LogMessageFatal::~LogMessageFatal()
    @        0x102591930  google::LogMessageFatal::~LogMessageFatal()
    @        0x10b56c4a8  kudu::(anonymous 
namespace)::ShardedCache<>::SetMetrics()
    @        0x103522c1c  kudu::cfile::BlockCache::StartInstrumentation()
    @        0x101249230  kudu::master::Master::Init()
    @        0x10126cbf0  kudu::master::RunMasterServer()
    @        0x1005eae64  kudu::master::MasterMain()
    @        0x1005ead10  main {code}
 If I uncomment the line
{code:java}
CHECK(IsGTest()) << "Metrics should only be set once per Cache";
 {code}
from kudu/util/cache.cc:579  ( CHECK(IsGTest()) << "Metrics should only be set 
once per Cache"< Starting Master server runs into error first time
> -
>
> Key: KUDU-3437
> URL: https://issues.apache.org/jira/browse/KUDU-3437
> Project: Kud

[jira] [Created] (KUDU-3439) make cpp client example error message for mistyped master easier to understand

2023-01-31 Thread Jira
Bakai Ádám created KUDU-3439:


 Summary: make cpp client example error message for mistyped master 
easier to understand
 Key: KUDU-3439
 URL: https://issues.apache.org/jira/browse/KUDU-3439
 Project: Kudu
  Issue Type: Wish
  Components: client
Reporter: Bakai Ádám


If the user mistypes the master address, then he gets a very long error 
message. It definitely contains the useful part ("Could not connect to the 
cluster: unable to resolve address for localhosta: nodename nor servname 
provided, or not known") , but there are so many lines it's kind of hard to 
find: 
{noformat}
adambakai@abakai-MBP16 examples % ./example localhosta:8764
Running with Kudu client version: kudu 1.17.0-SNAPSHOT (rev 
244b21efb87927de795206b6825b64eb76cacfd9-dirty)
Long version info: kudu 1.17.0-SNAPSHOT
revision 244b21efb87927de795206b6825b64eb76cacfd9-dirty
build type FASTDEBUG
built by adambakai at 31 Jan 2023 11:06:55 CET on abakai-MBP16
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/openssl_util.cc
 Line number: 109
 Time: Tue Jan 31 11:15:39 2023
 Message: FIPS mode is disabled.
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/security/tls_context.cc
 Line number: 297
 Time: Tue Jan 31 11:15:39 2023
 Message: OpenSSL security level is 1
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/thread.cc
 Line number: 658
 Time: Tue Jan 31 11:15:39 2023
 Message: Started thread 740961 - kernel-watchdog: kernel-watcher
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/thread.cc
 Line number: 658
 Time: Tue Jan 31 11:15:39 2023
 Message: Started thread 740962 - reactor: rpc reactor
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/thread.cc
 Line number: 658
 Time: Tue Jan 31 11:15:39 2023
 Message: Started thread 740963 - reactor: rpc reactor
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/thread.cc
 Line number: 658
 Time: Tue Jan 31 11:15:39 2023
 Message: Started thread 740964 - reactor: rpc reactor
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/thread.cc
 Line number: 658
 Time: Tue Jan 31 11:15:39 2023
 Message: Started thread 740965 - reactor: rpc reactor
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/cache.cc
 Line number: 541
 Time: Tue Jan 31 11:15:39 2023
 Message: Will use 16 shards for recency list cache.
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/mem_tracker.cc
 Line number: 83
 Time: Tue Jan 31 11:15:39 2023
 Message: Creating tracker root
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/mem_tracker.cc
 Line number: 83
 Time: Tue Jan 31 11:15:39 2023
 Message: Creating tracker ttl-cache-sharded_fifo_cache->root
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/ttl_cache.h
 Line number: 134
 Time: Tue Jan 31 11:15:39 2023
 Message: constructed TTL cache 'ttl-cache' with capacity of 1048576
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/util/mem_tracker.cc
 Line number: 87
 Time: Tue Jan 31 11:15:39 2023
 Message: Destroying tracker ttl-cache-sharded_fifo_cache->root
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/rpc/messenger.cc
 Line number: 185
 Time: Tue Jan 31 11:15:39 2023
 Message: shutting down messenger client
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/rpc/reactor.cc
 Line number: 236
 Time: Tue Jan 31 11:15:39 2023
 Message: client_R000: shutting down Reactor thread.
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/rpc/reactor.cc
 Line number: 236
 Time: Tue Jan 31 11:15:39 2023
 Message: client_R001: shutting down Reactor thread.
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/rpc/reactor.cc
 Line number: 251
 Time: Tue Jan 31 11:15:39 2023
 Message: client_R000: tearing down outbound TCP connections...
Received log message from Kudu client library
 Severity: 0
 Filename: /Users/adambakai/CLionProjects/kudu/src/kudu/rpc/reactor.cc
 Line number: 236

[jira] [Assigned] (KUDU-3437) Starting Master server runs into error first time

2023-03-02 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3437:


Assignee: Bakai Ádám

> Starting Master server runs into error first time
> -
>
> Key: KUDU-3437
> URL: https://issues.apache.org/jira/browse/KUDU-3437
> Project: Kudu
>  Issue Type: Bug
>  Components: cfile
> Environment: Macos 13.0.1 Arm64, fastdebug build
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>
> Reproduction script:
> {noformat}
> #Please make sure that stop_kudu.sh and start_kudu.sh and kudu command is in 
> your path to run this script.
> export PATH=$PATH:./bin/
> stop_kudu.sh;rm -rf master-*; mkdir -p master-3/data master-3/wal 
> master-3/log; start_kudu.sh -m 3
> echo just to make sure it returns three masters as expected
> kudu master list localhost:8764,localhost:8766,localhost:8768
> echo This returns error:
> kudu-master 
> --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770
>  --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
> --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
> --time_source=system_unsync --unlock_unsafe_flags 
> --webserver_interface=127.0.0.1 --webserver_port=8771
> echo just to make sure it returns three masters
> kudu master list localhost:8764,localhost:8766,localhost:8768
> echo Now it should be running fine:
> kudu-master 
> --master_addresses=127.0.0.1:8764,127.0.0.1:8766,127.0.0.1:8768,localhost:8770
>  --fs_data_dirs=$PWD/master-3/data --fs_wal_dir=$PWD/master-3/wal 
> --log_dir=$PWD/master-3/log --rpc_bind_addresses=127.0.0.1:8770 
> --time_source=system_unsync --unlock_unsafe_flags 
> --webserver_interface=127.0.0.1 --webserver_port=8771 &
> sleep 2 
> echo just to make sure it returns four masters
> kudu master list localhost:8764,localhost:8766,localhost:8768
>  {noformat}
> My error for the first run:
> {code:java}
> F20230127 15:58:28.923543 360774 cache.cc:579] Check failed: IsGTest() 
> Metrics should only be set once per Cache
> *** Check failure stack trace: ***
>     @        0x1025950d4  google::LogMessageFatal::~LogMessageFatal()
>     @        0x102591930  google::LogMessageFatal::~LogMessageFatal()
>     @        0x10b56c4a8  kudu::(anonymous 
> namespace)::ShardedCache<>::SetMetrics()
>     @        0x103522c1c  kudu::cfile::BlockCache::StartInstrumentation()
>     @        0x101249230  kudu::master::Master::Init()
>     @        0x10126cbf0  kudu::master::RunMasterServer()
>     @        0x1005eae64  kudu::master::MasterMain()
>     @        0x1005ead10  main {code}
>  If I uncomment the line
> {code:java}
> CHECK(IsGTest()) << "Metrics should only be set once per Cache";
>  {code}
> from kudu/util/cache.cc:579  it seems to work fine.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3491) MiniDumpExceptionHandler assert randomly fails

2023-07-17 Thread Jira
Bakai Ádám created KUDU-3491:


 Summary: MiniDumpExceptionHandler assert randomly fails 
 Key: KUDU-3491
 URL: https://issues.apache.org/jira/browse/KUDU-3491
 Project: Kudu
  Issue Type: Bug
Affects Versions: 1.15.0
Reporter: Bakai Ádám


When starting master, this error randomly happens on my system:
{code:java}
+ exec 
/opt/cloudera/parcels/CDH-7.2.18-1.cdh7.2.18.p0.43161468/lib/kudu/sbin/kudu-master
 
--master_addresses=abakai-1.abakai.root.hwx.site,abakai-2.abakai.root.hwx.site,abakai-3.abakai.root.hwx.site
 
--location_mapping_cmd=/var/run/cloudera-scm-agent/process/126-kudu-KUDU_MASTER/topology.py
 --flagfile=/var/run/cloudera-scm-agent/process/126-kudu-KUDU_MASTER/gflagfile
F20230717 10:37:23.626719 101405 minidump.cc:273] Check failed: 0 == 
MinidumpExceptionHandler::current_num_instances_.fetch_add(1) (0 vs. 1) 
*** Check failure stack trace: ***
Wrote minidump to 
/var/log/kudu/minidumps/kudu-master/1664cde8-f41a-4b7f-121578b0-55ba68a6.dmp
*** Aborted at 1689590243 (unix time) try "date -d @1689590243" if you are 
using GNU date ***
PC: @                0x0 (unknown)
*** SIGABRT (@0x9c2900018c1d) received by PID 101405 (TID 0x7f204a7cda00) from 
PID 101405; stack trace: ***
    @           0xe4df76 google::(anonymous namespace)::FailureSignalHandler()
    @     0x7f204a19e6d0 (unknown)
    @     0x7f20483a6277 __GI_raise
    @     0x7f20483a7968 __GI_abort
    @           0xcb50af kudu::AbortFailureFunction()
    @           0xe4296d google::LogMessage::Fail()
    @           0xe4584a google::LogMessage::SendToLog()
    @           0xe4249e google::LogMessage::Flush()
    @           0xe439d9 google::LogMessageFatal::~LogMessageFatal()
    @          0x319a0a1 
kudu::MinidumpExceptionHandler::RegisterMinidumpExceptionHandler()
    @          0x319a107 
kudu::MinidumpExceptionHandler::MinidumpExceptionHandler()
    @          0x12821f0 kudu::server::ServerBase::ServerBase()
    @          0x124577e kudu::kserver::KuduServer::KuduServer()
    @           0xddc9ce kudu::master::Master::Master()
    @           0xd5732b kudu::master::RunMasterServer()
    @           0xd51d8a kudu::master::MasterMain()
    @     0x7f2048392445 __libc_start_main
    @           0xd51b64 (unknown) {code}
Version information:
{code:java}
[root@abakai-1 ~]# 
/opt/cloudera/parcels/CDH-7.2.18-1.cdh7.2.18.p0.43161468/lib/kudu/sbin-release/kudu-master
 -version
kudu 1.15.0.7.2.18.0-205
revision 7e222133b1a13ce6c212ffb32d8ceaa0c6a8545a
build type RELEASE
built by None at 14 Jul 2023 06:51:13 UTC on re-centos-slave-large-2wczs
build id 1178567 {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3505) kudu ksck fails if healthy master in healthy cluster is started after the command

2023-08-24 Thread Jira
Bakai Ádám created KUDU-3505:


 Summary: kudu ksck fails if healthy master in healthy cluster is 
started after the command
 Key: KUDU-3505
 URL: https://issues.apache.org/jira/browse/KUDU-3505
 Project: Kudu
  Issue Type: Bug
  Components: master
Reporter: Bakai Ádám


Environment: single master configuration
If the master is not running and the user starts a kudu cluster ksck command 
then it will try to connect to the master over and over again. Once the master 
is started, the ksck command is executed and it shows a bunch of errors:
{code:java}
adambakai@abakai-MBP16 d % kudu cluster ksck localhost:8764 -ksck_format 
plain_full
Master Summary
               UUID               |    Address     | Status
--++-
 f41052a1ba8242d49ee5e16c0d60558a | localhost:8764 | HEALTHY
All reported replicas are:
  A = f41052a1ba8242d49ee5e16c0d60558a
The consensus matrix is:
 Config source | Replicas | Current term | Config index | Committed?
---+--+--+--+
 A             | A*       | 16           | -1           | YesFlags of checked 
categories for Master:
        Flag         |                            Value                         
   |         Master
-+-+-
 builtin_ntp_servers | 
0.pool.ntp.org,1.pool.ntp.org,2.pool.ntp.org,3.pool.ntp.org | all 1 server(s) 
checked
 time_source         | system_unsync                                            
   | all 1 server(s) checkedTablet Server Summary
               UUID               |    Address     | Status  | Location | 
Tablet Leaders | Active Scanners
--++-+--++-
 1938796538bf483f9bcd133e29aa645b | 127.0.0.1:9878 | HEALTHY |    |       
0        |       0
 8080a72aeb714c5087b8c515f21b1735 | 127.0.0.1:9870 | HEALTHY |    |       
1        |       0
 9f86252d00814cb3ae0ef6858ee31a02 | 127.0.0.1:9874 | HEALTHY |    |       
0        |       0
 c23de9c2b3e1448fa8dde2bb1a292388 | 127.0.0.1:9872 | HEALTHY |    |       
0        |       0
 fb700997c9274a9d8287eb3c765606d2 | 127.0.0.1:9876 | HEALTHY |    |       
0        |       0Tablet Server Location Summary
 Location |  Count
--+-
    |       5Flags of checked categories for Tablet Server:
        Flag         |                            Value                         
   |      Tablet Server
-+-+-
 builtin_ntp_servers | 
0.pool.ntp.org,1.pool.ntp.org,2.pool.ntp.org,3.pool.ntp.org | all 5 server(s) 
checked
 time_source         | system_unsync                                            
   | all 5 server(s) checkedVersion Summary
     Version     |         Servers
-+-
 1.18.0-SNAPSHOT | all 6 server(s) checkedTablet Summary
Tablet 5d87f015c3a2438c8cec6e84796f9ecb of table 'db.test_table' is healthy.
  8080a72aeb714c5087b8c515f21b1735 (127.0.0.1:9870): RUNNING [LEADER]
  c23de9c2b3e1448fa8dde2bb1a292388 (127.0.0.1:9872): RUNNING
  9f86252d00814cb3ae0ef6858ee31a02 (127.0.0.1:9874): RUNNING
All reported replicas are:
  A = 8080a72aeb714c5087b8c515f21b1735
  B = c23de9c2b3e1448fa8dde2bb1a292388
  C = 9f86252d00814cb3ae0ef6858ee31a02
The consensus matrix is:
 Config source |   Replicas   | Current term | Config index | Committed?
---+--+--+--+
 master        | A*  B   C    |              |              | Yes
 A             | A*  B   C    | 3            | -1           | Yes
 B             | A*  B   C    | 3            | -1           | Yes
 C             | A*  B   C    | 3            | -1           | YesThe cluster 
doesn't have any matching system tables
Summary by table
     Name      | RF | Status  | Total Tablets | Healthy | Recovering | 
Under-replicated | Unavailable
---++-+---+-++--+-
 db.test_table | 3  | HEALTHY | 1             | 1       | 0          | 0        
        | 0Tablet Replica Count Summary
   Statistic    | Replica Count
+---
 Minimum        | 0
 First Quartile | 0
 Median         | 1
 Third Quartile | 1
 Maximum        | 1Tablet Replica Count by Tablet Server
               UUID               |      Host      | Replica Count
--++---
 1938796538bf483f9bcd133e29aa645b | 127.0.0.1:9878 | 0
 8080a72aeb714c5087b8c515f21b1735 | 127.0.0.1:9870 | 1
 9f86252d00814cb3ae0ef6858ee31a02 | 127.0.0.1:9874 | 1
 c23de9c2b3e1448fa8dde2bb1a292388 | 127.0.0.1:9872 | 1
 fb700997c9274a9d8287eb3c765606d2 |

[jira] [Created] (KUDU-3547) kudu master add failes if new master port is not the default value

2024-01-24 Thread Jira
Bakai Ádám created KUDU-3547:


 Summary: kudu master add failes if new master port is not the 
default value
 Key: KUDU-3547
 URL: https://issues.apache.org/jira/browse/KUDU-3547
 Project: Kudu
  Issue Type: Bug
  Components: master
Reporter: Bakai Ádám


Tried this script on de87aca3dd29177658790cffeb7a1de0c8d69231 :
{code:java}
PATH=$PATH:./binPATH=$PATH:../../src/kudu/scripts/
cleanup(){kill %1 %2 %3 %4 %5 %6 %7 %8 %9 %10pkill kudu# sudo fuser -k -9 
9873/tcpfuser -k -9 8765/tcp
pkill kudupkill kudu-masterpkill kudu-tserver../src/kudu/scripts/stop_kudu.sh
sleep 1}
cleanup
rm -rf master-0/ tserver-0/ tserver-1 tserver-2 master-1 
master-2create_stuff(){rm -rf  $1/data $1/wal $1/log
mkdir -p $1/data $1/wal $1/log}
create_stuff master-0create_stuff tserver-0create_stuff tserver-1create_stuff 
tserver-2create_stuff master-1
sync

$PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-0/data 
--fs_wal_dir=$PWD/tserver-0/wal --log_dir=$PWD/tserver-0/log 
--rpc_bind_addresses=127.0.0.1:9870 --time_source=system_unsync 
--unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9871 
--tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 
&$PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-1/data 
--fs_wal_dir=$PWD/tserver-1/wal --log_dir=$PWD/tserver-1/log 
--rpc_bind_addresses=127.0.0.1:9872 --time_source=system_unsync 
--unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9873 
--tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 
&$PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-2/data 
--fs_wal_dir=$PWD/tserver-2/wal --log_dir=$PWD/tserver-2/log 
--rpc_bind_addresses=127.0.0.1:9874 --time_source=system_unsync 
--unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9875 
--tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 &
sleep 1
$PWD/bin/kudu-master --master_addresses=127.0.0.1:8764 
--fs_data_dirs=$PWD/master-0/data --fs_wal_dir=$PWD/master-0/wal 
--log_dir=$PWD/master-0/log --rpc_bind_addresses=127.0.0.1:8764 
--time_source=system_unsync --unlock_unsafe_flags  --webserver_port=8765 &

#/home/abakai/kudu_1_16_2/kudu/a/b/bin/kudu-master 
--master_addresses=127.0.0.1:8764 
--fs_data_dirs=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/data 
--fs_wal_dir=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/wal 
--log_dir=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/log 
--rpc_bind_addresses=127.0.0.1:8764 --time_source=system_unsync 
--unlock_unsafe_flags --webserver_port=8765 
--webserver_doc_root=/home/abakai/kudu_1_16_2/kudu/a/b/../../www 
kudu master list localhost:8764
sleep 1kudu master add 127.0.0.1:8764 127.0.0.1:8766 
--fs_data_dirs=$PWD/master-1/data --fs_wal_dir=$PWD/master-1/wal -logtostderr 
-v 99 &>test123
kudu master list localhost:8764
 {code}
New master is not added to the master quorum. 
There is a workaround, adding rpc_bind_address to the arguments:
{code:java}
kudu master add 127.0.0.1:8764 127.0.0.1:8766 --fs_data_dirs=$PWD/master-1/data 
--fs_wal_dir=$PWD/master-1/wal -logtostderr -v 1 
--rpc_bind_addresses=127.0.0.1:8766 
 {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3547) kudu master add fails if new master port is not the default value

2024-01-24 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3547?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3547:
-
Summary: kudu master add fails if new master port is not the default value  
(was: kudu master add failes if new master port is not the default value)

> kudu master add fails if new master port is not the default value
> -
>
> Key: KUDU-3547
> URL: https://issues.apache.org/jira/browse/KUDU-3547
> Project: Kudu
>  Issue Type: Bug
>  Components: master
>Reporter: Bakai Ádám
>Priority: Minor
>
> Tried this script on de87aca3dd29177658790cffeb7a1de0c8d69231 :
> {code:java}
> PATH=$PATH:./binPATH=$PATH:../../src/kudu/scripts/
> cleanup(){kill %1 %2 %3 %4 %5 %6 %7 %8 %9 %10pkill kudu# sudo fuser -k -9 
> 9873/tcpfuser -k -9 8765/tcp
> pkill kudupkill kudu-masterpkill kudu-tserver../src/kudu/scripts/stop_kudu.sh
> sleep 1}
> cleanup
> rm -rf master-0/ tserver-0/ tserver-1 tserver-2 master-1 
> master-2create_stuff(){rm -rf  $1/data $1/wal $1/log
> mkdir -p $1/data $1/wal $1/log}
> create_stuff master-0create_stuff tserver-0create_stuff tserver-1create_stuff 
> tserver-2create_stuff master-1
> sync
> $PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-0/data 
> --fs_wal_dir=$PWD/tserver-0/wal --log_dir=$PWD/tserver-0/log 
> --rpc_bind_addresses=127.0.0.1:9870 --time_source=system_unsync 
> --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9871 
> --tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 
> &$PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-1/data 
> --fs_wal_dir=$PWD/tserver-1/wal --log_dir=$PWD/tserver-1/log 
> --rpc_bind_addresses=127.0.0.1:9872 --time_source=system_unsync 
> --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9873 
> --tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 
> &$PWD/bin/kudu-tserver --fs_data_dirs=$PWD/tserver-2/data 
> --fs_wal_dir=$PWD/tserver-2/wal --log_dir=$PWD/tserver-2/log 
> --rpc_bind_addresses=127.0.0.1:9874 --time_source=system_unsync 
> --unlock_unsafe_flags --webserver_interface=127.0.0.1 --webserver_port=9875 
> --tserver_master_addrs=127.0.0.1:8764 --memory_limit_hard_bytes=17916750506 &
> sleep 1
> $PWD/bin/kudu-master --master_addresses=127.0.0.1:8764 
> --fs_data_dirs=$PWD/master-0/data --fs_wal_dir=$PWD/master-0/wal 
> --log_dir=$PWD/master-0/log --rpc_bind_addresses=127.0.0.1:8764 
> --time_source=system_unsync --unlock_unsafe_flags  --webserver_port=8765 &
> #/home/abakai/kudu_1_16_2/kudu/a/b/bin/kudu-master 
> --master_addresses=127.0.0.1:8764 
> --fs_data_dirs=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/data 
> --fs_wal_dir=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/wal 
> --log_dir=/home/abakai/kudu_1_16_2/kudu/a/b/master-0/log 
> --rpc_bind_addresses=127.0.0.1:8764 --time_source=system_unsync 
> --unlock_unsafe_flags --webserver_port=8765 
> --webserver_doc_root=/home/abakai/kudu_1_16_2/kudu/a/b/../../www 
> kudu master list localhost:8764
> sleep 1kudu master add 127.0.0.1:8764 127.0.0.1:8766 
> --fs_data_dirs=$PWD/master-1/data --fs_wal_dir=$PWD/master-1/wal -logtostderr 
> -v 99 &>test123
> kudu master list localhost:8764
>  {code}
> New master is not added to the master quorum. 
> There is a workaround, adding rpc_bind_address to the arguments:
> {code:java}
> kudu master add 127.0.0.1:8764 127.0.0.1:8766 
> --fs_data_dirs=$PWD/master-1/data --fs_wal_dir=$PWD/master-1/wal -logtostderr 
> -v 1 --rpc_bind_addresses=127.0.0.1:8766 
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3548) kudu master add times out if new master argument is an ip address that can be resolved

2024-01-24 Thread Jira
Bakai Ádám created KUDU-3548:


 Summary: kudu master add times out if new master argument is an ip 
address that can be resolved
 Key: KUDU-3548
 URL: https://issues.apache.org/jira/browse/KUDU-3548
 Project: Kudu
  Issue Type: Bug
Reporter: Bakai Ádám


It is produced on kudu on de87aca3dd29177658790cffeb7a1de0c8d69231:

{code:java}
kudu master add 127.0.0.1:8764 127.0.0.1 --fs_data_dirs=$PWD/master-1/data 
--fs_wal_dir=$PWD/master-1/wal -logtostderr -v 1 {code}
In CheckMasterVoterAndHealthy
the elements in 
master.registration().rpc_addresses() are:  127.0.0.1:8764 and 
adam-7070-ubuntu:7051 (adam-7070 is my hostname), and the 
new_master_hp.ToString() is 127.0.0.1:7051 which won't match to 
adam-7070-ubuntu:7051 even though they are the same. 
This makes CheckMasterVoterAndHealthy time out, but the new master still gets 
into the raft quorum. 
 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-2203) java.io.FileNotFoundException when trying to initialize MiniKuduCluster

2017-10-27 Thread JIRA
Nacho García Fernández created KUDU-2203:


 Summary:  java.io.FileNotFoundException when trying to initialize 
MiniKuduCluster
 Key: KUDU-2203
 URL: https://issues.apache.org/jira/browse/KUDU-2203
 Project: Kudu
  Issue Type: Bug
  Components: java, test
Affects Versions: 1.5.0
Reporter: Nacho García Fernández


I'm getting the following error when I try to create a new instance of 
MiniKuduCluster:
{code:java}
java.io.FileNotFoundException: Cannot find binary kudu-master in binary 
directory null

at org.apache.kudu.client.TestUtils.findBinary(TestUtils.java:159)
at 
org.apache.kudu.client.MiniKuduCluster.startMasters(MiniKuduCluster.java:210)
at 
org.apache.kudu.client.MiniKuduCluster.startCluster(MiniKuduCluster.java:153)
at 
org.apache.kudu.client.MiniKuduCluster.start(MiniKuduCluster.java:117)
at 
org.apache.kudu.client.MiniKuduCluster.access$300(MiniKuduCluster.java:50)
at 
org.apache.kudu.client.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:661)
at org.apache.kudu.client.BaseKuduTest.doSetup(BaseKuduTest.java:113)
at 
org.apache.kudu.client.BaseKuduTest.setUpBeforeClass(BaseKuduTest.java:76)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
at 
com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:68)
at 
com.intellij.rt.execution.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:47)
at 
com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:242)
at 
com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:70)
{code}

This is happening when a base class try to extend from BaseKuduTest (which 
internally instantiates a MiniKuduCluster).

My simple test class:

{code:java}

import org.apache.kudu.client.BaseKuduTest;
import org.junit.Before;
import org.junit.Test;

public class KuduInputFormatTest extends BaseKuduTest {


@Before
public void initialize() throws Exception {
}

@Test
public void test() throws Exception {
System.out.println("The error occurred before this message is printed");
}
{code}


Current POM dependencies:

{noformat}


org.apache.kudu
kudu-client
${kudu.version}



org.apache.kudu
kudu-client
${kudu.version}
test-jar
test

{noformat}

where ${kudu.version} is 1.5.0





--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (KUDU-2203) java.io.FileNotFoundException when trying to initialize MiniKuduCluster

2017-10-27 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/KUDU-2203?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nacho García Fernández updated KUDU-2203:
-
Description: 
I'm getting the following error when I try to create a new instance of 
MiniKuduCluster:
{code:java}
java.io.FileNotFoundException: Cannot find binary kudu-master in binary 
directory null

at org.apache.kudu.client.TestUtils.findBinary(TestUtils.java:159)
at 
org.apache.kudu.client.MiniKuduCluster.startMasters(MiniKuduCluster.java:210)
at 
org.apache.kudu.client.MiniKuduCluster.startCluster(MiniKuduCluster.java:153)
at 
org.apache.kudu.client.MiniKuduCluster.start(MiniKuduCluster.java:117)
at 
org.apache.kudu.client.MiniKuduCluster.access$300(MiniKuduCluster.java:50)
at 
org.apache.kudu.client.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:661)
at org.apache.kudu.client.BaseKuduTest.doSetup(BaseKuduTest.java:113)
at 
org.apache.kudu.client.BaseKuduTest.setUpBeforeClass(BaseKuduTest.java:76)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
at 
com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:68)
at 
com.intellij.rt.execution.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:47)
at 
com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:242)
at 
com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:70)
{code}

This is happening when a base class try to extend from BaseKuduTest (which 
internally instantiates a MiniKuduCluster).

My simple test class:

{code:java}

import org.apache.kudu.client.BaseKuduTest;
import org.junit.Before;
import org.junit.Test;

public class KuduInputFormatTest extends BaseKuduTest {


@Before
public void initialize() throws Exception {
}

@Test
public void test() throws Exception {
System.out.println("The error occurs before this message is printed");
}
{code}


Current POM dependencies:

{noformat}


org.apache.kudu
kudu-client
${kudu.version}



org.apache.kudu
kudu-client
${kudu.version}
test-jar
test

{noformat}

where kudu.version is 1.5.0



  was:
I'm getting the following error when I try to create a new instance of 
MiniKuduCluster:
{code:java}
java.io.FileNotFoundException: Cannot find binary kudu-master in binary 
directory null

at org.apache.kudu.client.TestUtils.findBinary(TestUtils.java:159)
at 
org.apache.kudu.client.MiniKuduCluster.startMasters(MiniKuduCluster.java:210)
at 
org.apache.kudu.client.MiniKuduCluster.startCluster(MiniKuduCluster.java:153)
at 
org.apache.kudu.client.MiniKuduCluster.start(MiniKuduCluster.java:117)
at 
org.apache.kudu.client.MiniKuduCluster.access$300(MiniKuduCluster.java:50)
at 
org.apache.kudu.client.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:661)
at org.apache.kudu.client.BaseKuduTest.doSetup(BaseKuduTest.java:113)
at 
org.apache.kudu.client.BaseKuduTest.setUpBeforeClass(BaseKuduTest.java:76)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters

[jira] [Commented] (KUDU-2203) java.io.FileNotFoundException when trying to initialize MiniKuduCluster

2017-10-30 Thread JIRA

[ 
https://issues.apache.org/jira/browse/KUDU-2203?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16224480#comment-16224480
 ] 

Nacho García Fernández commented on KUDU-2203:
--

Hi Jean-Daniel.

Thanks, it was useful. Is there any plan to provide an embedded kudu driver to 
perform tests without needing a real kudu deployment? (I can move this question 
to Stackoverflow if needed).

>  java.io.FileNotFoundException when trying to initialize MiniKuduCluster
> 
>
> Key: KUDU-2203
> URL: https://issues.apache.org/jira/browse/KUDU-2203
> Project: Kudu
>  Issue Type: Bug
>  Components: java, test
>Affects Versions: 1.5.0
>Reporter: Nacho García Fernández
>
> I'm getting the following error when I try to create a new instance of 
> MiniKuduCluster:
> {code:java}
> java.io.FileNotFoundException: Cannot find binary kudu-master in binary 
> directory null
>   at org.apache.kudu.client.TestUtils.findBinary(TestUtils.java:159)
>   at 
> org.apache.kudu.client.MiniKuduCluster.startMasters(MiniKuduCluster.java:210)
>   at 
> org.apache.kudu.client.MiniKuduCluster.startCluster(MiniKuduCluster.java:153)
>   at 
> org.apache.kudu.client.MiniKuduCluster.start(MiniKuduCluster.java:117)
>   at 
> org.apache.kudu.client.MiniKuduCluster.access$300(MiniKuduCluster.java:50)
>   at 
> org.apache.kudu.client.MiniKuduCluster$MiniKuduClusterBuilder.build(MiniKuduCluster.java:661)
>   at org.apache.kudu.client.BaseKuduTest.doSetup(BaseKuduTest.java:113)
>   at 
> org.apache.kudu.client.BaseKuduTest.setUpBeforeClass(BaseKuduTest.java:76)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
>   at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
>   at 
> com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:68)
>   at 
> com.intellij.rt.execution.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:47)
>   at 
> com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:242)
>   at 
> com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:70)
> {code}
> This is happening when a base class try to extend from BaseKuduTest (which 
> internally instantiates a MiniKuduCluster).
> My simple test class:
> {code:java}
> import org.apache.kudu.client.BaseKuduTest;
> import org.junit.Before;
> import org.junit.Test;
> public class KuduInputFormatTest extends BaseKuduTest {
> @Before
> public void initialize() throws Exception {
> }
> @Test
> public void test() throws Exception {
> System.out.println("The error occurs before this message is printed");
> }
> {code}
> Current POM dependencies:
> {noformat}
> 
> org.apache.kudu
> kudu-client
> ${kudu.version}
> 
> 
> org.apache.kudu
> kudu-client
> ${kudu.version}
> test-jar
> test
> 
> {noformat}
> where kudu.version is 1.5.0



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0

2017-11-21 Thread JIRA
Nacho García Fernández created KUDU-2223:


 Summary: Failed to add distributed masters: Unable to start Master 
at index 0
 Key: KUDU-2223
 URL: https://issues.apache.org/jira/browse/KUDU-2223
 Project: Kudu
  Issue Type: Bug
  Components: build, java
Affects Versions: 1.5.0
Reporter: Nacho García Fernández


After successfuly building Kudu on my OSX, I try to mvn verify in the java 
submodule, but I get the following exception:

{code:java}
[INFO] ---
[INFO]  T E S T S
[INFO] ---
[INFO] Running org.apache.kudu.client.TestAlterTable
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.42 s 
<<< FAILURE! - in org.apache.kudu.client.TestAlterTable
[ERROR] org.apache.kudu.client.TestAlterTable  Time elapsed: 0.42 s  <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6

[INFO] Running org.apache.kudu.client.TestAsyncKuduClient
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.644 s 
<<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduClient
[ERROR] org.apache.kudu.client.TestAsyncKuduClient  Time elapsed: 0.644 s  <<< 
ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestAsyncKuduClient.setUpBeforeClass(TestAsyncKuduClient.java:45)

[INFO] Running org.apache.kudu.client.TestAsyncKuduSession
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.478 s 
<<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduSession
[ERROR] org.apache.kudu.client.TestAsyncKuduSession  Time elapsed: 0.478 s  <<< 
ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestAsyncKuduSession.setUpBeforeClass(TestAsyncKuduSession.java:59)

[INFO] Running org.apache.kudu.client.TestAuthnTokenReacquire
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.601 s 
<<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquire
[ERROR] org.apache.kudu.client.TestAuthnTokenReacquire  Time elapsed: 0.601 s  
<<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestAuthnTokenReacquire.setUpBeforeClass(TestAuthnTokenReacquire.java:55)

[INFO] Running org.apache.kudu.client.TestAuthnTokenReacquireOpen
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.865 s 
<<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquireOpen
[ERROR] org.apache.kudu.client.TestAuthnTokenReacquireOpen  Time elapsed: 0.865 
s  <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to start a single 
Master: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: 
process exited on signal 6
at 
org.apache.kudu.client.TestAuthnTokenReacquireOpen.setUpBeforeClass(TestAuthnTokenReacquireOpen.java:58)

[INFO] Running org.apache.kudu.client.TestBitSet
[INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.061 s 
- in org.apache.kudu.client.TestBitSet
[INFO] Running org.apache.kudu.client.TestBytes
[INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.077 s 
- in org.apache.kudu.client.TestBytes
[INFO] Running org.apache.kudu.client.TestClientFailoverSupport
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.399 s 
<<< FAILURE! - in org.apache.kudu.client.TestClientFailoverSupport
[ERROR] org.apache.kudu.client.TestClientFailoverSupport  Time elapsed: 0.399 s 
 <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestClientFailoverSupport.setUpBeforeClass(TestClientFailoverSupport.java:35)

[INFO] Running org.apache.kudu.client.TestColumnRangePredicate
[INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.106 s 
- in org.apache.kudu.client.TestColumnRangePredicate
[INFO] Running org.apache.kudu.client.Tes

[jira] [Updated] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0

2017-11-21 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nacho García Fernández updated KUDU-2223:
-
Description: 
After successfully building Kudu on my OSX, I try to mvn verify in the java 
submodule, but I get the following exception:

{code:java}
[INFO] ---
[INFO]  T E S T S
[INFO] ---
[INFO] Running org.apache.kudu.client.TestAlterTable
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.42 s 
<<< FAILURE! - in org.apache.kudu.client.TestAlterTable
[ERROR] org.apache.kudu.client.TestAlterTable  Time elapsed: 0.42 s  <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6

[INFO] Running org.apache.kudu.client.TestAsyncKuduClient
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.644 s 
<<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduClient
[ERROR] org.apache.kudu.client.TestAsyncKuduClient  Time elapsed: 0.644 s  <<< 
ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestAsyncKuduClient.setUpBeforeClass(TestAsyncKuduClient.java:45)

[INFO] Running org.apache.kudu.client.TestAsyncKuduSession
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.478 s 
<<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduSession
[ERROR] org.apache.kudu.client.TestAsyncKuduSession  Time elapsed: 0.478 s  <<< 
ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestAsyncKuduSession.setUpBeforeClass(TestAsyncKuduSession.java:59)

[INFO] Running org.apache.kudu.client.TestAuthnTokenReacquire
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.601 s 
<<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquire
[ERROR] org.apache.kudu.client.TestAuthnTokenReacquire  Time elapsed: 0.601 s  
<<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestAuthnTokenReacquire.setUpBeforeClass(TestAuthnTokenReacquire.java:55)

[INFO] Running org.apache.kudu.client.TestAuthnTokenReacquireOpen
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.865 s 
<<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquireOpen
[ERROR] org.apache.kudu.client.TestAuthnTokenReacquireOpen  Time elapsed: 0.865 
s  <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to start a single 
Master: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: 
process exited on signal 6
at 
org.apache.kudu.client.TestAuthnTokenReacquireOpen.setUpBeforeClass(TestAuthnTokenReacquireOpen.java:58)

[INFO] Running org.apache.kudu.client.TestBitSet
[INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.061 s 
- in org.apache.kudu.client.TestBitSet
[INFO] Running org.apache.kudu.client.TestBytes
[INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.077 s 
- in org.apache.kudu.client.TestBytes
[INFO] Running org.apache.kudu.client.TestClientFailoverSupport
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.399 s 
<<< FAILURE! - in org.apache.kudu.client.TestClientFailoverSupport
[ERROR] org.apache.kudu.client.TestClientFailoverSupport  Time elapsed: 0.399 s 
 <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestClientFailoverSupport.setUpBeforeClass(TestClientFailoverSupport.java:35)

[INFO] Running org.apache.kudu.client.TestColumnRangePredicate
[INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.106 s 
- in org.apache.kudu.client.TestColumnRangePredicate
[INFO] Running org.apache.kudu.client.TestConnectionCache
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.43 s 
<<< FAILURE! - in org.apache.kudu.client.TestConnectionCache
[ERROR] test(org.apache.kudu.client.TestConnectio

[jira] [Updated] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0

2017-11-21 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nacho García Fernández updated KUDU-2223:
-
Description: 
After successfully building Kudu on my OSX, I try to run the  mvn verify 
command in the java submodule, but I get the following exception:

{code:java}
[INFO] ---
[INFO]  T E S T S
[INFO] ---
[INFO] Running org.apache.kudu.client.TestAlterTable
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.42 s 
<<< FAILURE! - in org.apache.kudu.client.TestAlterTable
[ERROR] org.apache.kudu.client.TestAlterTable  Time elapsed: 0.42 s  <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6

[INFO] Running org.apache.kudu.client.TestAsyncKuduClient
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.644 s 
<<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduClient
[ERROR] org.apache.kudu.client.TestAsyncKuduClient  Time elapsed: 0.644 s  <<< 
ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestAsyncKuduClient.setUpBeforeClass(TestAsyncKuduClient.java:45)

[INFO] Running org.apache.kudu.client.TestAsyncKuduSession
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.478 s 
<<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduSession
[ERROR] org.apache.kudu.client.TestAsyncKuduSession  Time elapsed: 0.478 s  <<< 
ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestAsyncKuduSession.setUpBeforeClass(TestAsyncKuduSession.java:59)

[INFO] Running org.apache.kudu.client.TestAuthnTokenReacquire
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.601 s 
<<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquire
[ERROR] org.apache.kudu.client.TestAuthnTokenReacquire  Time elapsed: 0.601 s  
<<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestAuthnTokenReacquire.setUpBeforeClass(TestAuthnTokenReacquire.java:55)

[INFO] Running org.apache.kudu.client.TestAuthnTokenReacquireOpen
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.865 s 
<<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquireOpen
[ERROR] org.apache.kudu.client.TestAuthnTokenReacquireOpen  Time elapsed: 0.865 
s  <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to start a single 
Master: /Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: 
process exited on signal 6
at 
org.apache.kudu.client.TestAuthnTokenReacquireOpen.setUpBeforeClass(TestAuthnTokenReacquireOpen.java:58)

[INFO] Running org.apache.kudu.client.TestBitSet
[INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.061 s 
- in org.apache.kudu.client.TestBitSet
[INFO] Running org.apache.kudu.client.TestBytes
[INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.077 s 
- in org.apache.kudu.client.TestBytes
[INFO] Running org.apache.kudu.client.TestClientFailoverSupport
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.399 s 
<<< FAILURE! - in org.apache.kudu.client.TestClientFailoverSupport
[ERROR] org.apache.kudu.client.TestClientFailoverSupport  Time elapsed: 0.399 s 
 <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/n.garcia.fernandez/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestClientFailoverSupport.setUpBeforeClass(TestClientFailoverSupport.java:35)

[INFO] Running org.apache.kudu.client.TestColumnRangePredicate
[INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.106 s 
- in org.apache.kudu.client.TestColumnRangePredicate
[INFO] Running org.apache.kudu.client.TestConnectionCache
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.43 s 
<<< FAILURE! - in org.apache.kudu.client.TestConnectionCache
[ERROR] test(org.apache.kudu.clie

[jira] [Updated] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0

2017-11-21 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nacho García Fernández updated KUDU-2223:
-
Description: 
After successfully building Kudu on my OSX, I try to run the  mvn verify 
command in the java submodule, but I get the following exception:

{code:java}
[INFO] ---
[INFO]  T E S T S
[INFO] ---
[INFO] Running org.apache.kudu.client.TestAlterTable
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.42 s 
<<< FAILURE! - in org.apache.kudu.client.TestAlterTable
[ERROR] org.apache.kudu.client.TestAlterTable  Time elapsed: 0.42 s  <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited on 
signal 6

[INFO] Running org.apache.kudu.client.TestAsyncKuduClient
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.644 s 
<<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduClient
[ERROR] org.apache.kudu.client.TestAsyncKuduClient  Time elapsed: 0.644 s  <<< 
ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited on 
signal 6
at 
org.apache.kudu.client.TestAsyncKuduClient.setUpBeforeClass(TestAsyncKuduClient.java:45)

[INFO] Running org.apache.kudu.client.TestAsyncKuduSession
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.478 s 
<<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduSession
[ERROR] org.apache.kudu.client.TestAsyncKuduSession  Time elapsed: 0.478 s  <<< 
ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited on 
signal 6
at 
org.apache.kudu.client.TestAsyncKuduSession.setUpBeforeClass(TestAsyncKuduSession.java:59)

[INFO] Running org.apache.kudu.client.TestAuthnTokenReacquire
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.601 s 
<<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquire
[ERROR] org.apache.kudu.client.TestAuthnTokenReacquire  Time elapsed: 0.601 s  
<<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited on 
signal 6
at 
org.apache.kudu.client.TestAuthnTokenReacquire.setUpBeforeClass(TestAuthnTokenReacquire.java:55)

[INFO] Running org.apache.kudu.client.TestAuthnTokenReacquireOpen
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.865 s 
<<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquireOpen
[ERROR] org.apache.kudu.client.TestAuthnTokenReacquireOpen  Time elapsed: 0.865 
s  <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to start a single 
Master: /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process 
exited on signal 6
at 
org.apache.kudu.client.TestAuthnTokenReacquireOpen.setUpBeforeClass(TestAuthnTokenReacquireOpen.java:58)

[INFO] Running org.apache.kudu.client.TestBitSet
[INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.061 s 
- in org.apache.kudu.client.TestBitSet
[INFO] Running org.apache.kudu.client.TestBytes
[INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.077 s 
- in org.apache.kudu.client.TestBytes
[INFO] Running org.apache.kudu.client.TestClientFailoverSupport
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.399 s 
<<< FAILURE! - in org.apache.kudu.client.TestClientFailoverSupport
[ERROR] org.apache.kudu.client.TestClientFailoverSupport  Time elapsed: 0.399 s 
 <<< ERROR!
org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
masters: Unable to start Master at index 0: 
/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited on 
signal 6
at 
org.apache.kudu.client.TestClientFailoverSupport.setUpBeforeClass(TestClientFailoverSupport.java:35)

[INFO] Running org.apache.kudu.client.TestColumnRangePredicate
[INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.106 s 
- in org.apache.kudu.client.TestColumnRangePredicate
[INFO] Running org.apache.kudu.client.TestConnectionCache
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.43 s 
<<< FAILURE! - in org.apache.kudu.client.TestConnectionCache
[ERROR] test(org.apache.kudu.client.TestConnectionCache)  Time elapsed: 0.375 s 
 <<<

[jira] [Updated] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0

2017-11-21 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nacho García Fernández updated KUDU-2223:
-
Affects Version/s: (was: 1.5.0)
   1.6.0

> Failed to add distributed masters: Unable to start Master at index 0
> 
>
> Key: KUDU-2223
> URL: https://issues.apache.org/jira/browse/KUDU-2223
> Project: Kudu
>  Issue Type: Bug
>  Components: build, java
>Affects Versions: 1.6.0
>Reporter: Nacho García Fernández
>
> After successfully building Kudu on my OSX, I try to run the  mvn verify 
> command in the java submodule, but I get the following exception:
> {code:java}
> [INFO] ---
> [INFO]  T E S T S
> [INFO] ---
> [INFO] Running org.apache.kudu.client.TestAlterTable
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.42 
> s <<< FAILURE! - in org.apache.kudu.client.TestAlterTable
> [ERROR] org.apache.kudu.client.TestAlterTable  Time elapsed: 0.42 s  <<< 
> ERROR!
> org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
> masters: Unable to start Master at index 0: 
> /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited 
> on signal 6
> [INFO] Running org.apache.kudu.client.TestAsyncKuduClient
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.644 
> s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduClient
> [ERROR] org.apache.kudu.client.TestAsyncKuduClient  Time elapsed: 0.644 s  
> <<< ERROR!
> org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
> masters: Unable to start Master at index 0: 
> /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited 
> on signal 6
>   at 
> org.apache.kudu.client.TestAsyncKuduClient.setUpBeforeClass(TestAsyncKuduClient.java:45)
> [INFO] Running org.apache.kudu.client.TestAsyncKuduSession
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.478 
> s <<< FAILURE! - in org.apache.kudu.client.TestAsyncKuduSession
> [ERROR] org.apache.kudu.client.TestAsyncKuduSession  Time elapsed: 0.478 s  
> <<< ERROR!
> org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
> masters: Unable to start Master at index 0: 
> /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited 
> on signal 6
>   at 
> org.apache.kudu.client.TestAsyncKuduSession.setUpBeforeClass(TestAsyncKuduSession.java:59)
> [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquire
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.601 
> s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquire
> [ERROR] org.apache.kudu.client.TestAuthnTokenReacquire  Time elapsed: 0.601 s 
>  <<< ERROR!
> org.apache.kudu.client.NonRecoverableException: Failed to add distributed 
> masters: Unable to start Master at index 0: 
> /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process exited 
> on signal 6
>   at 
> org.apache.kudu.client.TestAuthnTokenReacquire.setUpBeforeClass(TestAuthnTokenReacquire.java:55)
> [INFO] Running org.apache.kudu.client.TestAuthnTokenReacquireOpen
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.865 
> s <<< FAILURE! - in org.apache.kudu.client.TestAuthnTokenReacquireOpen
> [ERROR] org.apache.kudu.client.TestAuthnTokenReacquireOpen  Time elapsed: 
> 0.865 s  <<< ERROR!
> org.apache.kudu.client.NonRecoverableException: Failed to start a single 
> Master: /Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master: process 
> exited on signal 6
>   at 
> org.apache.kudu.client.TestAuthnTokenReacquireOpen.setUpBeforeClass(TestAuthnTokenReacquireOpen.java:58)
> [INFO] Running org.apache.kudu.client.TestBitSet
> [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.061 
> s - in org.apache.kudu.client.TestBitSet
> [INFO] Running org.apache.kudu.client.TestBytes
> [INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.077 
> s - in org.apache.kudu.client.TestBytes
> [INFO] Running org.apache.kudu.client.TestClientFailoverSupport
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.399 
> s <<< FAILURE! - in org.apache.kudu.client.TestClientFailoverSupport
> [ERROR] org.apache.kudu.client.TestClientFailoverSupport  Time elapsed: 0.399 
> s  <<<

[jira] [Commented] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0

2017-11-21 Thread JIRA

[ 
https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16261314#comment-16261314
 ] 

Nacho García Fernández commented on KUDU-2223:
--

Example of surefire logs:

{code:java}
19:58:04.892 [INFO - Time-limited test] (MiniKuduCluster.java:145) Starting 
process: [/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu, test, 
mini_cluster, --serialization=pb]
19:58:05.054 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: 
create_cluster {
  num_masters: 3
  num_tservers: 0
  enable_kerberos: false
}

19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:123) Response:
19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: 
start_cluster {
}

19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 
19:58:05.079746 3212440512 external_mini_cluster.cc:771] Running 
/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master
19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master
19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--fs_wal_dir=/tmp/kudutest-214522293/minicluster-data/master-0/wal
19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/master-0/data
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--block_manager=file
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--webserver_interface=localhost
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--ipki_ca_key_size=1024
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--tsk_num_rsa_bits=512
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--rpc_bind_addresses=127.0.0.1:11030
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--webserver_interface=127.0.0.1
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--webserver_port=0
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--never_fsync
19:58:05.082 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--ipki_server_key_size=1024
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--enable_minidumps=false
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--redact=none
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--metrics_log_interval_ms=1000
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--logtostderr
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--logbuflevel=-1
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--log_dir=/tmp/kudutest-214522293/minicluster-data/master-0/logs
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--server_dump_info_path=/tmp/kudutest-214522293/minicluster-data/master-0/data/info.pb
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--server_dump_info_format=pb
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--rpc_server_allow_ephemeral_ports
19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--unlock_experimental_flags
19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--unlock_unsafe_flags
19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--master_addresses=127.0.0.1:11030,127.0.0.1:11031,127.0.0.1:11032 with env {}
19:58:05.095 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
WARNING: Logging before InitGoogleLogging() is written to STDERR
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.095039 3212440512 flags.cc:391] Enabled unsafe flag: 
--rpc_server_allow_ephemeral_ports=true
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.095845 3212440512 flags.cc:391] Enabled unsafe flag: --never_fsync=true
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.096652 3212440512 flags.cc:391] Enabled experimental flag: 
--ipki_ca_key_size=1024
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.096664 3212440512 flags.cc:391] Enabled experimental flag: 
--ipki_server_key_size=1024
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.096679 3212440512 flags.cc:391] Enabled experimental flag: 
--tsk_num_rsa_bits=512
19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 
19:58:05.097530 3212440512 master_main.cc:66] Master server non-default flags:
19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/master-0/data
19:58:05.097 [INFO - cluster stderr pri

[jira] [Comment Edited] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0

2017-11-21 Thread JIRA

[ 
https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16261314#comment-16261314
 ] 

Nacho García Fernández edited comment on KUDU-2223 at 11/21/17 7:11 PM:


Example of surefire logs:

{code:java}
19:58:04.892 [INFO - Time-limited test] (MiniKuduCluster.java:145) Starting 
process: [/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu, test, 
mini_cluster, --serialization=pb]
19:58:05.054 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: 
create_cluster {
  num_masters: 3
  num_tservers: 0
  enable_kerberos: false
}

19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:123) Response:
19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: 
start_cluster {
}

19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 
19:58:05.079746 3212440512 external_mini_cluster.cc:771] Running 
/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master
19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master
19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--fs_wal_dir=/tmp/kudutest-214522293/minicluster-data/master-0/wal
19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/master-0/data
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--block_manager=file
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--webserver_interface=localhost
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--ipki_ca_key_size=1024
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--tsk_num_rsa_bits=512
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--rpc_bind_addresses=127.0.0.1:11030
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--webserver_interface=127.0.0.1
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--webserver_port=0
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--never_fsync
19:58:05.082 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--ipki_server_key_size=1024
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--enable_minidumps=false
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--redact=none
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--metrics_log_interval_ms=1000
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--logtostderr
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--logbuflevel=-1
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--log_dir=/tmp/kudutest-214522293/minicluster-data/master-0/logs
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--server_dump_info_path=/tmp/kudutest-214522293/minicluster-data/master-0/data/info.pb
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--server_dump_info_format=pb
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--rpc_server_allow_ephemeral_ports
19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--unlock_experimental_flags
19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--unlock_unsafe_flags
19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--master_addresses=127.0.0.1:11030,127.0.0.1:11031,127.0.0.1:11032 with env {}
19:58:05.095 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
WARNING: Logging before InitGoogleLogging() is written to STDERR
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.095039 3212440512 flags.cc:391] Enabled unsafe flag: 
--rpc_server_allow_ephemeral_ports=true
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.095845 3212440512 flags.cc:391] Enabled unsafe flag: --never_fsync=true
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.096652 3212440512 flags.cc:391] Enabled experimental flag: 
--ipki_ca_key_size=1024
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.096664 3212440512 flags.cc:391] Enabled experimental flag: 
--ipki_server_key_size=1024
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.096679 3212440512 flags.cc:391] Enabled experimental flag: 
--tsk_num_rsa_bits=512
19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 
19:58:05.097530 3212440512 master_main.cc:66] Master server non-default flags:
19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/mast

[jira] [Comment Edited] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0

2017-11-21 Thread JIRA

[ 
https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16261314#comment-16261314
 ] 

Nacho García Fernández edited comment on KUDU-2223 at 11/21/17 7:12 PM:


Example of surefire logs:

{code:java}
19:58:04.892 [INFO - Time-limited test] (MiniKuduCluster.java:145) Starting 
process: [/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu, test, 
mini_cluster, --serialization=pb]
19:58:05.054 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: 
create_cluster {
  num_masters: 3
  num_tservers: 0
  enable_kerberos: false
}

19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:123) Response:
19:58:05.063 [DEBUG - Time-limited test] (MiniKuduCluster.java:113) Request: 
start_cluster {
}

19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 
19:58:05.079746 3212440512 external_mini_cluster.cc:771] Running 
/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master
19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
/Users/0xNacho/dev/github/kudu/build/latest/bin/kudu-master
19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--fs_wal_dir=/tmp/kudutest-214522293/minicluster-data/master-0/wal
19:58:05.080 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/master-0/data
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--block_manager=file
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--webserver_interface=localhost
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--ipki_ca_key_size=1024
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--tsk_num_rsa_bits=512
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--rpc_bind_addresses=127.0.0.1:11030
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--webserver_interface=127.0.0.1
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--webserver_port=0
19:58:05.081 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--never_fsync
19:58:05.082 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--ipki_server_key_size=1024
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--enable_minidumps=false
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--redact=none
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--metrics_log_interval_ms=1000
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--logtostderr
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--logbuflevel=-1
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--log_dir=/tmp/kudutest-214522293/minicluster-data/master-0/logs
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--server_dump_info_path=/tmp/kudutest-214522293/minicluster-data/master-0/data/info.pb
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--server_dump_info_format=pb
19:58:05.083 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--rpc_server_allow_ephemeral_ports
19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--unlock_experimental_flags
19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--unlock_unsafe_flags
19:58:05.084 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--master_addresses=127.0.0.1:11030,127.0.0.1:11031,127.0.0.1:11032 with env {}
19:58:05.095 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
WARNING: Logging before InitGoogleLogging() is written to STDERR
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.095039 3212440512 flags.cc:391] Enabled unsafe flag: 
--rpc_server_allow_ephemeral_ports=true
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.095845 3212440512 flags.cc:391] Enabled unsafe flag: --never_fsync=true
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.096652 3212440512 flags.cc:391] Enabled experimental flag: 
--ipki_ca_key_size=1024
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.096664 3212440512 flags.cc:391] Enabled experimental flag: 
--ipki_server_key_size=1024
19:58:05.096 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) W1121 
19:58:05.096679 3212440512 flags.cc:391] Enabled experimental flag: 
--tsk_num_rsa_bits=512
19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) I1121 
19:58:05.097530 3212440512 master_main.cc:66] Master server non-default flags:
19:58:05.097 [INFO - cluster stderr printer] (MiniKuduCluster.java:437) 
--fs_data_dirs=/tmp/kudutest-214522293/minicluster-data/mast

[jira] [Commented] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0

2017-11-22 Thread JIRA

[ 
https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16262426#comment-16262426
 ] 

Nacho García Fernández commented on KUDU-2223:
--

I've just run the tests one more time after removing the /tmp/master directory. 
However, some of them are still failing with the same exception:

{code:java}
0xNacho  ~/dev/github/kudu/java   master ●  mvn test
[INFO] Scanning for projects...
[WARNING]
[WARNING] Some problems were encountered while building the effective model for 
org.apache.kudu:kudu-spark2-tools_2.11:jar:1.6.0-SNAPSHOT
[WARNING] 'artifactId' contains an expression but should be a constant. @ 
org.apache.kudu:kudu-${spark.version.label}-tools_${scala.binary.version}:[unknown-version],
 /Users/0xNacho/dev/github/kudu/java/kudu-spark-tools/pom.xml, line 30, column 
17
[WARNING]
[WARNING] Some problems were encountered while building the effective model for 
org.apache.kudu:kudu-spark2_2.11:jar:1.6.0-SNAPSHOT
[WARNING] 'artifactId' contains an expression but should be a constant. @ 
org.apache.kudu:kudu-${spark.version.label}_${scala.binary.version}:[unknown-version],
 /Users/0xNacho/dev/github/kudu/java/kudu-spark/pom.xml, line 23, column 17
[WARNING]
[WARNING] It is highly recommended to fix these problems because they threaten 
the stability of your build.
[WARNING]
[WARNING] For this reason, future Maven versions might no longer support 
building such malformed projects.
[WARNING]
[INFO] 
[INFO] Detecting the operating system and CPU architecture
[INFO] 
[INFO] os.detected.name: osx
[INFO] os.detected.arch: x86_64
[INFO] os.detected.version: 10.12
[INFO] os.detected.version.major: 10
[INFO] os.detected.version.minor: 12
[INFO] os.detected.classifier: osx-x86_64
[INFO] 
[INFO] Reactor Build Order:
[INFO]
[INFO] Kudu
[INFO] Kudu Java Client
[INFO] Kudu MapReduce bindings
[INFO] Kudu Client Tools
[INFO] Kudu Flume NG Sink
[INFO] Kudu Hive
[INFO] Kudu Spark Bindings
[INFO] Kudu Spark Tools
[INFO] Kudu Flink bindings
[INFO]
[INFO] 
[INFO] Building Kudu 1.6.0-SNAPSHOT
[INFO] 
[INFO]
[INFO] --- maven-enforcer-plugin:1.4.1:enforce (default) @ kudu-parent ---
[INFO]
[INFO] --- maven-checkstyle-plugin:2.17:check (validate) @ kudu-parent ---
[INFO] Starting audit...
Audit done.
[INFO]
[INFO] --- maven-remote-resources-plugin:1.5:process (process-resource-bundles) 
@ kudu-parent ---
[INFO]
[INFO] --- animal-sniffer-maven-plugin:1.16:check (check-java7-apis) @ 
kudu-parent ---
[INFO] Checking unresolved references to org.codehaus.mojo.signature:java17:1.0
[INFO]
[INFO] --- maven-antrun-plugin:1.8:run (default) @ kudu-parent ---
[INFO] Executing tasks

main:
[mkdir] Created dir: /Users/0xNacho/dev/github/kudu/java/target/testdata
[INFO] Executed tasks
[INFO]
[INFO] 
[INFO] Building Kudu Java Client 1.6.0-SNAPSHOT
[INFO] 
[INFO]
[INFO] --- maven-checkstyle-plugin:2.17:check (validate) @ kudu-client ---
[INFO] Starting audit...
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:490:
 'method def modifier' have incorrect indentation level 6, expected level 
should be one of the following: 8, 10, 12. [Indentation]
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:492:
 'method def' child have incorrect indentation level 8, expected level should 
be one of the following: 10, 12, 14. [Indentation]
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:494:
 'method def' child have incorrect indentation level 8, expected level should 
be one of the following: 10, 12, 14. [Indentation]
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:495:
 'method def rcurly' have incorrect indentation level 6, expected level should 
be one of the following: 8, 10, 12. [Indentation]
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:496:
 'object def rcurly' have incorrect indentation level 4, expected level should 
be one of the following: 6, 8, 10. [Indentation]
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:1103:
 Line is longer than 100 characters (f

[jira] [Comment Edited] (KUDU-2223) Failed to add distributed masters: Unable to start Master at index 0

2017-11-22 Thread JIRA

[ 
https://issues.apache.org/jira/browse/KUDU-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16262426#comment-16262426
 ] 

Nacho García Fernández edited comment on KUDU-2223 at 11/22/17 12:23 PM:
-

I've just run the tests one more time after removing the /tmp/master directory. 
However, some of them are still failing with the same exception:

{code:java}
0xNacho  ~/dev/github/kudu/java   master ●  mvn test
[INFO] Scanning for projects...
[WARNING]
[WARNING] Some problems were encountered while building the effective model for 
org.apache.kudu:kudu-spark2-tools_2.11:jar:1.6.0-SNAPSHOT
[WARNING] 'artifactId' contains an expression but should be a constant. @ 
org.apache.kudu:kudu-${spark.version.label}-tools_${scala.binary.version}:[unknown-version],
 /Users/0xNacho/dev/github/kudu/java/kudu-spark-tools/pom.xml, line 30, column 
17
[WARNING]
[WARNING] Some problems were encountered while building the effective model for 
org.apache.kudu:kudu-spark2_2.11:jar:1.6.0-SNAPSHOT
[WARNING] 'artifactId' contains an expression but should be a constant. @ 
org.apache.kudu:kudu-${spark.version.label}_${scala.binary.version}:[unknown-version],
 /Users/0xNacho/dev/github/kudu/java/kudu-spark/pom.xml, line 23, column 17
[WARNING]
[WARNING] It is highly recommended to fix these problems because they threaten 
the stability of your build.
[WARNING]
[WARNING] For this reason, future Maven versions might no longer support 
building such malformed projects.
[WARNING]
[INFO] 
[INFO] Detecting the operating system and CPU architecture
[INFO] 
[INFO] os.detected.name: osx
[INFO] os.detected.arch: x86_64
[INFO] os.detected.version: 10.12
[INFO] os.detected.version.major: 10
[INFO] os.detected.version.minor: 12
[INFO] os.detected.classifier: osx-x86_64
[INFO] 
[INFO] Reactor Build Order:
[INFO]
[INFO] Kudu
[INFO] Kudu Java Client
[INFO] Kudu MapReduce bindings
[INFO] Kudu Client Tools
[INFO] Kudu Flume NG Sink
[INFO] Kudu Hive
[INFO] Kudu Spark Bindings
[INFO] Kudu Spark Tools
[INFO] Kudu Flink bindings
[INFO]
[INFO] 
[INFO] Building Kudu 1.6.0-SNAPSHOT
[INFO] 
[INFO]
[INFO] --- maven-enforcer-plugin:1.4.1:enforce (default) @ kudu-parent ---
[INFO]
[INFO] --- maven-checkstyle-plugin:2.17:check (validate) @ kudu-parent ---
[INFO] Starting audit...
Audit done.
[INFO]
[INFO] --- maven-remote-resources-plugin:1.5:process (process-resource-bundles) 
@ kudu-parent ---
[INFO]
[INFO] --- animal-sniffer-maven-plugin:1.16:check (check-java7-apis) @ 
kudu-parent ---
[INFO] Checking unresolved references to org.codehaus.mojo.signature:java17:1.0
[INFO]
[INFO] --- maven-antrun-plugin:1.8:run (default) @ kudu-parent ---
[INFO] Executing tasks

main:
[mkdir] Created dir: /Users/0xNacho/dev/github/kudu/java/target/testdata
[INFO] Executed tasks
[INFO]
[INFO] 
[INFO] Building Kudu Java Client 1.6.0-SNAPSHOT
[INFO] 
[INFO]
[INFO] --- maven-checkstyle-plugin:2.17:check (validate) @ kudu-client ---
[INFO] Starting audit...
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:490:
 'method def modifier' have incorrect indentation level 6, expected level 
should be one of the following: 8, 10, 12. [Indentation]
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:492:
 'method def' child have incorrect indentation level 8, expected level should 
be one of the following: 10, 12, 14. [Indentation]
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:494:
 'method def' child have incorrect indentation level 8, expected level should 
be one of the following: 10, 12, 14. [Indentation]
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:495:
 'method def rcurly' have incorrect indentation level 6, expected level should 
be one of the following: 8, 10, 12. [Indentation]
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKuduClient.java:496:
 'object def rcurly' have incorrect indentation level 4, expected level should 
be one of the following: 6, 8, 10. [Indentation]
[WARN] 
/Users/0xNacho/dev/github/kudu/java/kudu-client/src/main/java/org/apache/kudu/client/AsyncKud

[jira] [Created] (KUDU-2329) Random RPC timeout errors when inserting rows in a Kudu table

2018-03-02 Thread JIRA
Héctor Gutiérrez created KUDU-2329:
--

 Summary: Random RPC timeout errors when inserting rows in a Kudu 
table
 Key: KUDU-2329
 URL: https://issues.apache.org/jira/browse/KUDU-2329
 Project: Kudu
  Issue Type: Bug
  Components: rpc, server
Affects Versions: 1.5.0
Reporter: Héctor Gutiérrez


When executing inserts into a Kudu table, we are experiencing errors at random 
times. The first time we found one of these errors was during a bulk update of 
a Kudu table via Spark (in Scala):

{{kuduContext.updateRows(dataFrame, "table_name")}}

The error message in Spark was the following:

{{java.lang.RuntimeException: failed to write 579 rows from DataFrame to Kudu; 
sample errors: Timed out: can not complete before timeout: Batch{operations=6, 
tablet="cd1e33fce0114c9bbd9c14e2559e7591" [0x000F, 0x0010), 
ignoreAllDuplicateRows=false, rpc=KuduRpc(method=Write, 
tablet=cd1e33fce0114c9bbd9c14e2559e7591, attempt=3, 
DeadlineTracker(timeout=3, elapsed=30090), Traces: [0ms] sending RPC to 
server 6f273933b4d5498e87aadfb99b054a21, [10011ms] received from server 
6f273933b4d5498e87aadfb99b054a21 response Network error: [peer 
6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the 
channel, [10011ms] delaying RPC due to Network error: [peer 
6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the 
channel, [10033ms] sending RPC to server 6f273933b4d5498e87aadfb99b054a21, 
[20050ms] received from server 6f273933b4d5498e87aadfb99b054a21 response 
Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read 
timeout; closing the channel, [20050ms] delaying RPC due to Network error: 
[peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the 
channel, [20072ms] sending RPC to server 6f273933b4d5498e87aadfb99b054a21, 
[30090ms] received from server 6f273933b4d5498e87aadfb99b054a21 response 
Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read 
timeout; closing the channel, [30090ms] delaying RPC due to Network error: 
[peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the 
channel)}}}

(+ 4 more errors similar to this one in the error message)

We first thought it was actually a problem with our Spark code, but when we 
tried to execute a simple "INSERT INTO" query from the impala shell into a Kudu 
table, we got the following error:

{{[.] > insert into test_kudu values (282, 
'hola');}}
{{ Query: insert into test_kudu values (282, 'hola')}}
{{ Query submitted at: ..}}
{{ Query progress can be monitored at: }}
{{ WARNINGS: Kudu error(s) reported, first error: Timed out: Failed to write 
batch of 1 ops to tablet 9c295e90811e483a9550bfd75abcf666 after 1 attempt(s): 
Failed to write to server: 071bcafbb1644678a697c474662047b7 
(.:7050): Write RPC to :7050 timed 
out after 179.949s (SENT)}}

{{Error in Kudu table 'impala:kudu_db.test_kudu': Timed out: Failed to write 
batch of 1 ops to tablet 9c295e90811e483a9550bfd75abcf666 after 1 attempt(s): 
Failed to write to server: 071bcafbb1644678a697c474662047b7 
(...:7050): Write RPC to ..:7050 
timed out after 179.949s (SENT)}}

To make things even more confusing, despite getting this error in the impala 
shell, after a while (and not immediately), the inserted rows ended up in the 
table, so somehow they were actually inserted.

We also tried tweaking the Kudu timeout configuration values that we had 
previously set, but it didn't solve anything and the problem kept appearing.

Furthermore, we don't always get these errors, they only appear at random 
times. For example, right now we're just getting errors in that update we have 
in the Spark code, but we are not experiencing issues when working from the 
impala shell.

After all that we have tried, we are pretty certain that this is a bug in Kudu, 
although we think it is a bit strange that it is undocumented and certainly 
it's hard to reproduce.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (KUDU-2329) Random RPC timeout errors when inserting rows in a Kudu table

2018-03-02 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/KUDU-2329?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Héctor Gutiérrez updated KUDU-2329:
---
Description: 
When executing inserts into a Kudu table, we are experiencing errors at random 
times. The first time we found one of these errors was during a bulk update of 
a Kudu table via Spark (in Scala):

{{kuduContext.updateRows(dataFrame, "table_name")}}

The error message in Spark was the following:

{{java.lang.RuntimeException: failed to write 579 rows from DataFrame to Kudu; 
sample errors: Timed out: can not complete before timeout: Batch

{operations=6, tablet="cd1e33fce0114c9bbd9c14e2559e7591" [0x000F, 
0x0010), ignoreAllDuplicateRows=false, rpc=KuduRpc(method=Write, 
tablet=cd1e33fce0114c9bbd9c14e2559e7591, attempt=3, 
DeadlineTracker(timeout=3, elapsed=30090), Traces: [0ms] sending RPC to 
server 6f273933b4d5498e87aadfb99b054a21, [10011ms] received from server 
6f273933b4d5498e87aadfb99b054a21 response Network error: [peer 
6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the 
channel, [10011ms] delaying RPC due to Network error: [peer 
6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the 
channel, [10033ms] sending RPC to server 6f273933b4d5498e87aadfb99b054a21, 
[20050ms] received from server 6f273933b4d5498e87aadfb99b054a21 response 
Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read 
timeout; closing the channel, [20050ms] delaying RPC due to Network error: 
[peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the 
channel, [20072ms] sending RPC to server 6f273933b4d5498e87aadfb99b054a21, 
[30090ms] received from server 6f273933b4d5498e87aadfb99b054a21 response 
Network error: [peer 6f273933b4d5498e87aadfb99b054a21] encountered a read 
timeout; closing the channel, [30090ms] delaying RPC due to Network error: 
[peer 6f273933b4d5498e87aadfb99b054a21] encountered a read timeout; closing the 
channel)}

}}

(+ 4 more errors similar to this one in the error message)

We first thought it was actually a problem with our Spark code, but when we 
tried to execute a simple "INSERT INTO" query from the impala shell into a Kudu 
table, we got the following error:

{{[.] > insert into test_kudu values (282, 
'hola');}}
{{ \{{ Query: insert into test_kudu values (282, 'hola')
{{ \{{ Query submitted at: ..
{{ \{{ Query progress can be monitored at: 
{{ \{{ WARNINGS: Kudu error(s) reported, first error: Timed out: Failed to 
write batch of 1 ops to tablet 9c295e90811e483a9550bfd75abcf666 after 1 
attempt(s): Failed to write to server: 071bcafbb1644678a697c474662047b7 
(.:7050): Write RPC to :7050 timed 
out after 179.949s (SENT)

{{Error in Kudu table 'impala:kudu_db.test_kudu': Timed out: Failed to write 
batch of 1 ops to tablet 9c295e90811e483a9550bfd75abcf666 after 1 attempt(s): 
Failed to write to server: 071bcafbb1644678a697c474662047b7 
(...:7050): Write RPC to ..:7050 
timed out after 179.949s (SENT)}}

To make things even more confusing, despite getting this error in the impala 
shell, after a while (and not immediately), the inserted rows ended up in the 
table, so somehow they were actually inserted.

We also tried tweaking the Kudu timeout configuration values that we had 
previously set, but it didn't solve anything and the problem kept appearing.

Furthermore, we don't always get these errors, they only appear at random 
times. For example, right now we're just getting errors in that update we have 
in the Spark code, but we are not experiencing issues when working from the 
impala shell.

After all that we have tried, we are pretty certain that this is a bug in Kudu, 
although we think it is a bit strange that it is undocumented and certainly 
it's hard to reproduce.

  was:
When executing inserts into a Kudu table, we are experiencing errors at random 
times. The first time we found one of these errors was during a bulk update of 
a Kudu table via Spark (in Scala):

{{kuduContext.updateRows(dataFrame, "table_name")}}

The error message in Spark was the following:

{{java.lang.RuntimeException: failed to write 579 rows from DataFrame to Kudu; 
sample errors: Timed out: can not complete before timeout: Batch{operations=6, 
tablet="cd1e33fce0114c9bbd9c14e2559e7591" [0x000F, 0x0010), 
ignoreAllDuplicateRows=false, rpc=KuduRpc(method=Write, 
tablet=cd1e33fce0114c9bbd9c14e2559e7591, attempt=3, 
DeadlineTracker(timeout=3, elapsed=30090), Traces: [0ms] sending RPC to 
server 6f273933b4d5498e87aadfb99b054a21, [10011ms] received from server 
6f273933b4d5498e87aadfb99b054a21 response Network error: [peer 
6f27393

[jira] [Created] (KUDU-2332) Error when importing kudu package in Python Anaconda distribution

2018-03-08 Thread JIRA
Michał Sznajder created KUDU-2332:
-

 Summary: Error when importing kudu package in Python Anaconda 
distribution
 Key: KUDU-2332
 URL: https://issues.apache.org/jira/browse/KUDU-2332
 Project: Kudu
  Issue Type: Bug
  Components: python
Affects Versions: 1.2.0
 Environment: ProductName:Mac OS X
ProductVersion:10.13.3
BuildVersion:17D102
Anaconda Python distribution
Reporter: Michał Sznajder
 Attachments: error.png

I tried to install Kudu on my local machine:

ProductName:    Mac OS X
ProductVersion:    10.13.3
BuildVersion:    17D102

I followed all the steps to build Kudu 1.6 from source:

1. brew install autoconf automake cmake libtool pkg-config pstree
2. git clone https://github.com/apache/incubator-kudu kudu
2. cd kudu
3. PKG_CONFIG_PATH=/usr/local/Cellar/openssl/1.0.2n/lib/pkgconfig 
thirdparty/build-if-necessary.sh
4. mkdir -p build/release
5. cd build/release
6. PKG_CONFIG_PATH=/usr/local/Cellar/openssl/1.0.2n/lib/pkgconfig 
../../thirdparty/installed/common/bin/cmake \
  -DCMAKE_BUILD_TYPE=release \
  -DOPENSSL_ROOT_DIR=/usr/local/opt/openssl \
  ../..
7. make -j4
8. sudo make install

This resulted with following libraries installed:

 

/usr/local/include/kudu
/usr/local/include/kudu/util/kudu_export.h
/usr/local/lib/libkudu_client.0.1.0.dylib
/usr/local/lib/libkudu_client.dylib
/usr/local/lib/libkudu_client.0.dylib
/usr/local/share/kuduClient
/usr/local/share/kuduClient/cmake/kuduClientTargets.cmake
/usr/local/share/kuduClient/cmake/kuduClientTargets-release.cmake
/usr/local/share/kuduClient/cmake/kuduClientConfig.cmake
/usr/local/share/doc/kuduClient

 

Then I followed steps to instal kudu-python package using pip:

1. clean pip cache to make sure it is clean
2. pip install -v kudu-python

Then after calling:

import kudu

I got error like in the attachment "error.png".

As first line of this screen states it was Anaconda Python distribution.

After removing Anaconda Python and installing Python using Homebrew and again 
following above steps - all worked.

My conclusion: there is some kind of issue happening between Anaconda Python 
and kudu-python package.

Some more details are also in 
[https://getkudu.slack.com|https://getkudu.slack.com/] slack channel on 
#kudu-general channel.

I am reachable there under msznajder nickname.

 

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (KUDU-2487) kudu quote escape issues in select query

2019-01-03 Thread JIRA


[ 
https://issues.apache.org/jira/browse/KUDU-2487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16733145#comment-16733145
 ] 

Tiago Simões commented on KUDU-2487:


Isn't this an Impala issue?

> kudu quote escape issues in select query
> 
>
> Key: KUDU-2487
> URL: https://issues.apache.org/jira/browse/KUDU-2487
> Project: Kudu
>  Issue Type: Bug
>  Components: impala
>Affects Versions: 1.7.0
> Environment: kudu-1.7.0
> cdh5.15.0
> impala-2.12.0
>Reporter: fangwan
>Priority: Critical
>
> From impala shell, I am getting wrong results by using escape single quote 
> for query, details explained below:
>  1. create a kudu table like:
> {code:java}
> CREATE TABLE test_cf_escape(
>  x INT NOT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION,
> y STRING NOT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION,
> z INT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION,
> PRIMARY KEY(x, y)
> )
> STORED AS KUDU
> TBLPROPERTIES ('kudu.master_addresses'='hadoop-nn.xxx.com'){code}
> 2. insert data to the table:
> {code:java}
> insert into test_cf_escape (x,y,z) values (1, "chili\\\'s", 0);
> insert into test_cf_escape (x,y,z) values (1, "chili\'s", 2);
> select * from test_cf_escape;
> =>
> +---+--+---+
> | x | y| z |
> +---+--+---+
> | 1 | chili's  | 2 |
> | 1 | chili\'s | 0 |
> +---+--+---+
> {code}
> 3. we should get the same row by using below two select query, but failed:
> {code:java}
> select * from test_cf_escape where y = 'chili\'s'; => (1, 'chili\'s', 0) 
> (supposed to read data as (1, "chili's", 2))
> select * from test_cf_escape where y = "chili's"; => (1, "chili's", 2)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (KUDU-2487) kudu quote escape issues in select query

2019-01-03 Thread JIRA


[ 
https://issues.apache.org/jira/browse/KUDU-2487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16733145#comment-16733145
 ] 

Tiago Simões edited comment on KUDU-2487 at 1/3/19 3:18 PM:


Isn't this an Impala issue?

Have you tried to do the same using a parquet file?


was (Author: tiagorodriguessimoes):
Isn't this an Impala issue?

> kudu quote escape issues in select query
> 
>
> Key: KUDU-2487
> URL: https://issues.apache.org/jira/browse/KUDU-2487
> Project: Kudu
>  Issue Type: Bug
>  Components: impala
>Affects Versions: 1.7.0
> Environment: kudu-1.7.0
> cdh5.15.0
> impala-2.12.0
>Reporter: fangwan
>Priority: Critical
>
> From impala shell, I am getting wrong results by using escape single quote 
> for query, details explained below:
>  1. create a kudu table like:
> {code:java}
> CREATE TABLE test_cf_escape(
>  x INT NOT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION,
> y STRING NOT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION,
> z INT NULL ENCODING AUTO_ENCODING COMPRESSION DEFAULT_COMPRESSION,
> PRIMARY KEY(x, y)
> )
> STORED AS KUDU
> TBLPROPERTIES ('kudu.master_addresses'='hadoop-nn.xxx.com'){code}
> 2. insert data to the table:
> {code:java}
> insert into test_cf_escape (x,y,z) values (1, "chili\\\'s", 0);
> insert into test_cf_escape (x,y,z) values (1, "chili\'s", 2);
> select * from test_cf_escape;
> =>
> +---+--+---+
> | x | y| z |
> +---+--+---+
> | 1 | chili's  | 2 |
> | 1 | chili\'s | 0 |
> +---+--+---+
> {code}
> 3. we should get the same row by using below two select query, but failed:
> {code:java}
> select * from test_cf_escape where y = 'chili\'s'; => (1, 'chili\'s', 0) 
> (supposed to read data as (1, "chili's", 2))
> select * from test_cf_escape where y = "chili's"; => (1, "chili's", 2)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KUDU-3588) SecurityITest.TestNonDefaultPrincipalMultipleMaster flakiness

2024-07-16 Thread Jira
Bakai Ádám created KUDU-3588:


 Summary: SecurityITest.TestNonDefaultPrincipalMultipleMaster 
flakiness
 Key: KUDU-3588
 URL: https://issues.apache.org/jira/browse/KUDU-3588
 Project: Kudu
  Issue Type: Bug
Reporter: Bakai Ádám






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (KUDU-3588) SecurityITest.TestNonDefaultPrincipalMultipleMaster flakiness

2024-07-16 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3588?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám closed KUDU-3588.

Resolution: Duplicate

> SecurityITest.TestNonDefaultPrincipalMultipleMaster flakiness
> -
>
> Key: KUDU-3588
> URL: https://issues.apache.org/jira/browse/KUDU-3588
> Project: Kudu
>  Issue Type: Bug
>Reporter: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3589) SecurityITest.TestNonDefaultPrincipalMultipleMaster is flaky

2024-07-16 Thread Jira
Bakai Ádám created KUDU-3589:


 Summary: SecurityITest.TestNonDefaultPrincipalMultipleMaster is 
flaky
 Key: KUDU-3589
 URL: https://issues.apache.org/jira/browse/KUDU-3589
 Project: Kudu
  Issue Type: Sub-task
Reporter: Bakai Ádám






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3590) update certs in test_certs.cc

2024-07-16 Thread Jira
Bakai Ádám created KUDU-3590:


 Summary: update certs in test_certs.cc
 Key: KUDU-3590
 URL: https://issues.apache.org/jira/browse/KUDU-3590
 Project: Kudu
  Issue Type: Sub-task
Reporter: Bakai Ádám






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (KUDU-3590) update certs in test_certs.cc

2024-07-16 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3590?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3590:


Assignee: Bakai Ádám

> update certs in test_certs.cc
> -
>
> Key: KUDU-3590
> URL: https://issues.apache.org/jira/browse/KUDU-3590
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (KUDU-3589) SecurityITest.TestNonDefaultPrincipalMultipleMaster is flaky

2024-07-16 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3589:


Assignee: Bakai Ádám

> SecurityITest.TestNonDefaultPrincipalMultipleMaster is flaky
> 
>
> Key: KUDU-3589
> URL: https://issues.apache.org/jira/browse/KUDU-3589
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3605) org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally is flaky

2024-08-06 Thread Jira
Bakai Ádám created KUDU-3605:


 Summary: 
org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally
 is flaky
 Key: KUDU-3605
 URL: https://issues.apache.org/jira/browse/KUDU-3605
 Project: Kudu
  Issue Type: Sub-task
Reporter: Bakai Ádám






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (KUDU-3605) org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally is flaky

2024-08-06 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3605?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3605:


Assignee: Bakai Ádám

> org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally
>  is flaky
> -
>
> Key: KUDU-3605
> URL: https://issues.apache.org/jira/browse/KUDU-3605
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (KUDU-3605) org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally is flaky

2024-08-08 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3605?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17872068#comment-17872068
 ] 

Bakai Ádám commented on KUDU-3605:
--

Found the probable cause of issue, which is missing data on the disk for 
syscatalog:
{code:java}
13:14:38.568 [INFO - cluster stderr printer] (MiniKuduCluster.java:751) 
E20240808 13:14:38.568001  1477 master.cc:392] Unable to init master catalog 
manager: Not found: Unable to initialize catalog manager: Failed to initialize 
sys tables async: Unable to load consensus metadata for tablet 
: Unable to load consensus metadata for tablet 
: 
/tmp/dist-test-task6KabZO/test-tmp/mini-kudu-cluster692925529015678/master-1/wal/consensus-meta/:
 No such file or directory (error 2)
13:14:38.568 [INFO - cluster stderr printer] (MiniKuduCluster.java:751) 
I20240808 13:14:38.568156  1409 master.cc:537] Master@127.1.11.125:39515 
shutting down...
13:14:38.585 [INFO - cluster stderr printer] (MiniKuduCluster.java:751) 
I20240808 13:14:38.585610  1409 master.cc:552] Master@127.1.11.125:39515 
shutdown complete. {code}

> org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally
>  is flaky
> -
>
> Key: KUDU-3605
> URL: https://issues.apache.org/jira/browse/KUDU-3605
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (KUDU-3589) SecurityITest.TestNonDefaultPrincipalMultipleMaster is flaky

2024-11-04 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám resolved KUDU-3589.
--
Fix Version/s: n/a
   Resolution: Fixed

> SecurityITest.TestNonDefaultPrincipalMultipleMaster is flaky
> 
>
> Key: KUDU-3589
> URL: https://issues.apache.org/jira/browse/KUDU-3589
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
> Fix For: n/a
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (KUDU-3605) org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally is flaky

2024-11-04 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3605?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám resolved KUDU-3605.
--
Fix Version/s: n/a
   Resolution: Fixed

> org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally
>  is flaky
> -
>
> Key: KUDU-3605
> URL: https://issues.apache.org/jira/browse/KUDU-3605
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
> Fix For: n/a
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (KUDU-3624) EchoSubprocessTest.TestSubprocessMetricsOnError is flaky

2024-11-04 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3624?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3624:


Assignee: Bakai Ádám

> EchoSubprocessTest.TestSubprocessMetricsOnError is flaky
> 
>
> Key: KUDU-3624
> URL: https://issues.apache.org/jira/browse/KUDU-3624
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>
> {code:java}
> I20241104 14:09:34.411099 543202 server.cc:273] Received an EOF from the 
> subprocess
> W20241104 14:09:34.411275 543165 server.cc:408] The subprocess has exited 
> with status 9
> I20241104 14:09:34.417060 543203 server.cc:440] outbound queue shut down: 
> Aborted:
> I20241104 14:09:34.417075 543200 server.cc:366] get failed, inbound queue 
> shut down: Aborted:
> I20241104 14:09:34.417109 543201 server.cc:366] get failed, inbound queue 
> shut down: Aborted:
> I20241104 14:09:34.417068 543199 server.cc:366] get failed, inbound queue 
> shut down: Aborted:
> I20241104 14:09:37.790342 543244 server.cc:273] Received an EOF from the 
> subprocess
> F20241104 14:09:37.790630 543207 server.cc:401] Check failed: _s.ok() Bad 
> status: Runtime error: Unable to wait on child: No child processes (error 10)
> *** Check failure stack trace: ***
> I20241104 14:09:37.790678 543242 server.cc:366] get failed, inbound queue 
> shut down: Aborted:
> I20241104 14:09:37.790673 543243 server.cc:366] get failed, inbound queue 
> shut down: Aborted:
> *** Aborted at 1730729377 (unix time) try "date -d @1730729377" if you are 
> using GNU date ***
> I20241104 14:09:37.790684 543241 server.cc:366] get failed, inbound queue 
> shut down: Aborted:
> I20241104 14:09:37.790699 543245 server.cc:440] outbound queue shut down: 
> Aborted:
> PC: @                0x0 (unknown)
> *** SIGABRT (@0x848a9) received by PID 542889 (TID 0x704a47600700) from PID 
> 542889; stack trace: ***
>     @     0x704a4d478980 (unknown)
>     @     0x704a4d0b3e87 gsignal
>     @     0x704a4d0b57f1 abort
>     @     0x704a4e171d8d google::LogMessage::Fail()
>     @     0x704a4e175b53 google::LogMessage::SendToLog()
>     @     0x704a4e17178c google::LogMessage::Flush()
>     @     0x704a4e172f19 google::LogMessageFatal::~LogMessageFatal()
>     @     0x704a4f855a7e 
> kudu::subprocess::SubprocessServer::ExitCheckerThread()
>     @     0x704a4f8529bd 
> _ZZN4kudu10subprocess16SubprocessServer4InitEvENKUlvE0_clEv
>     @     0x704a4f8568ab 
> _ZNSt17_Function_handlerIFvvEZN4kudu10subprocess16SubprocessServer4InitEvEUlvE0_E9_M_invokeERKSt9_Any_data
>     @     0x704a4f8cb06a std::function<>::operator()()
>     @     0x704a4eff682b kudu::Thread::SuperviseThread()
>     @     0x704a4d46d6db start_thread
>     @     0x704a4d19661f clone
> zsh: abort (core dumped)  ./bin/subprocess_proxy-test --gtest_repeat=999 
> --gtest_break_on_failure {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3624) EchoSubprocessTest.TestSubprocessMetricsOnError is flaky

2024-11-04 Thread Jira
Bakai Ádám created KUDU-3624:


 Summary: EchoSubprocessTest.TestSubprocessMetricsOnError is flaky
 Key: KUDU-3624
 URL: https://issues.apache.org/jira/browse/KUDU-3624
 Project: Kudu
  Issue Type: Sub-task
Reporter: Bakai Ádám


{code:java}
I20241104 14:09:34.411099 543202 server.cc:273] Received an EOF from the 
subprocess
W20241104 14:09:34.411275 543165 server.cc:408] The subprocess has exited with 
status 9
I20241104 14:09:34.417060 543203 server.cc:440] outbound queue shut down: 
Aborted:
I20241104 14:09:34.417075 543200 server.cc:366] get failed, inbound queue shut 
down: Aborted:
I20241104 14:09:34.417109 543201 server.cc:366] get failed, inbound queue shut 
down: Aborted:
I20241104 14:09:34.417068 543199 server.cc:366] get failed, inbound queue shut 
down: Aborted:
I20241104 14:09:37.790342 543244 server.cc:273] Received an EOF from the 
subprocess
F20241104 14:09:37.790630 543207 server.cc:401] Check failed: _s.ok() Bad 
status: Runtime error: Unable to wait on child: No child processes (error 10)
*** Check failure stack trace: ***
I20241104 14:09:37.790678 543242 server.cc:366] get failed, inbound queue shut 
down: Aborted:
I20241104 14:09:37.790673 543243 server.cc:366] get failed, inbound queue shut 
down: Aborted:
*** Aborted at 1730729377 (unix time) try "date -d @1730729377" if you are 
using GNU date ***
I20241104 14:09:37.790684 543241 server.cc:366] get failed, inbound queue shut 
down: Aborted:
I20241104 14:09:37.790699 543245 server.cc:440] outbound queue shut down: 
Aborted:
PC: @                0x0 (unknown)
*** SIGABRT (@0x848a9) received by PID 542889 (TID 0x704a47600700) from PID 
542889; stack trace: ***
    @     0x704a4d478980 (unknown)
    @     0x704a4d0b3e87 gsignal
    @     0x704a4d0b57f1 abort
    @     0x704a4e171d8d google::LogMessage::Fail()
    @     0x704a4e175b53 google::LogMessage::SendToLog()
    @     0x704a4e17178c google::LogMessage::Flush()
    @     0x704a4e172f19 google::LogMessageFatal::~LogMessageFatal()
    @     0x704a4f855a7e kudu::subprocess::SubprocessServer::ExitCheckerThread()
    @     0x704a4f8529bd 
_ZZN4kudu10subprocess16SubprocessServer4InitEvENKUlvE0_clEv
    @     0x704a4f8568ab 
_ZNSt17_Function_handlerIFvvEZN4kudu10subprocess16SubprocessServer4InitEvEUlvE0_E9_M_invokeERKSt9_Any_data
    @     0x704a4f8cb06a std::function<>::operator()()
    @     0x704a4eff682b kudu::Thread::SuperviseThread()
    @     0x704a4d46d6db start_thread
    @     0x704a4d19661f clone
zsh: abort (core dumped)  ./bin/subprocess_proxy-test --gtest_repeat=999 
--gtest_break_on_failure {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (KUDU-3605) org.apache.kudu.client.TestSecurity.testExternallyProvidedSubjectRefreshedExternally is flaky

2024-09-24 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3605?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17884277#comment-17884277
 ] 

Bakai Ádám commented on KUDU-3605:
--

I had to recreate my dev environment, and the same test now fails with a 
different error:
{code:java}
There was 1 failure:
1) 
testExternallyProvidedSubjectRefreshedExternally(org.apache.kudu.client.TestSecurity)
org.apache.kudu.client.NonRecoverableException: cannot complete before timeout: 
KuduRpc(method=ListTabletServers, tablet=null, attempt=24, 
TimeoutTracker(timeout=3, elapsed=28366), Traces: [0ms] refreshing cache 
from master, [30ms] Sub RPC ConnectToMaster: sending RPC to server 
master-127.1.11.62:46867, [105ms] Sub RPC ConnectToMaster: sending RPC to 
server master-127.1.11.60:40051, [107ms] Sub RPC ConnectToMaster: sending RPC 
to server master-127.1.11.61:46759, [144ms] Sub RPC ConnectToMaster: received 
response from server master-127.1.11.62:46867: Network error: Failed to connect 
to peer master-127.1.11.62:46867(127.1.11.62:46867): Connection refused: 
/127.1.11.62:46867, [148ms] Sub RPC ConnectToMaster: received response from 
server master-127.1.11.61:46759: Network error: Failed to connect to peer 
master-127.1.11.61:46759(127.1.11.61:46759): Connection refused: 
/127.1.11.61:46759, [431ms] Sub RPC ConnectToMaster: received response from 
server master-127.1.11.60:40051: OK, [436ms] delaying RPC due to: Service 
unavailable: Master config 
(127.1.11.62:46867,127.1.11.60:40051,127.1.11.61:46759) has no leader. 
Exceptions received: org.apache.kudu.client.RecoverableException: Failed to 
connect to peer master-127.1.11.62:46867(127.1.11.62:46867): Connection 
refused: /127.1.11.62:46867,org.apache.kudu.client.RecoverableException: Failed 
to connect to peer master-127.1.11.61:46759(127.1.11.61:46759): Connection 
refused: /127.1.11.61:46759, [455ms] refreshing cache from master, [456ms] Sub 
RPC ConnectToMaster: sending RPC to server master-127.1.11.62:46867, [458ms] 
Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.60:40051, 
[460ms] Sub RPC ConnectToMaster: received response from server 
master-127.1.11.62:46867: Network error: Failed to connect to peer 
master-127.1.11.62:46867(127.1.11.62:46867): Connection refused: 
/127.1.11.62:46867, [460ms] Sub RPC ConnectToMaster: sending RPC to server 
master-127.1.11.61:46759, [463ms] Sub RPC ConnectToMaster: received response 
from server master-127.1.11.61:46759: Network error: Failed to connect to peer 
master-127.1.11.61:46759(127.1.11.61:46759): Connection refused: 
/127.1.11.61:46759, [465ms] Sub RPC ConnectToMaster: received response from 
server master-127.1.11.60:40051: OK, [466ms] delaying RPC due to: Service 
unavailable: Master config 
(127.1.11.62:46867,127.1.11.60:40051,127.1.11.61:46759) has no leader. 
Exceptions received: org.apache.kudu.client.RecoverableException: Failed to 
connect to peer master-127.1.11.62:46867(127.1.11.62:46867): Connection 
refused: /127.1.11.62:46867,org.apache.kudu.client.RecoverableException: Failed 
to connect to peer master-127.1.11.61:46759(127.1.11.61:46759): Connection 
refused: /127.1.11.61:46759, [475ms] refreshing cache from master, [476ms] Sub 
RPC ConnectToMaster: sending RPC to server master-127.1.11.62:46867, [477ms] 
Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.60:40051, 
[477ms] Sub RPC ConnectToMaster: sending RPC to server 
master-127.1.11.61:46759, [480ms] Sub RPC ConnectToMaster: received response 
from server master-127.1.11.61:46759: Network error: Failed to connect to peer 
master-127.1.11.61:46759(127.1.11.61:46759): Connection refused: 
/127.1.11.61:46759, [481ms] Sub RPC ConnectToMaster: received response from 
server master-127.1.11.60:40051: OK, [482ms] Sub RPC ConnectToMaster: received 
response from server master-127.1.11.62:46867: Network error: Failed to connect 
to peer master-127.1.11.62:46867(127.1.11.62:46867): Connection refused: 
/127.1.11.62:46867, [482ms] delaying RPC due to: Service unavailable: Master 
config (127.1.11.62:46867,127.1.11.60:40051,127.1.11.61:46759) has no leader. 
Exceptions received: org.apache.kudu.client.RecoverableException: Failed to 
connect to peer master-127.1.11.61:46759(127.1.11.61:46759): Connection 
refused: /127.1.11.61:46759,org.apache.kudu.client.RecoverableException: Failed 
to connect to peer master-127.1.11.62:46867(127.1.11.62:46867): Connection 
refused: /127.1.11.62:46867, [495ms] refreshing cache from master, [496ms] Sub 
RPC ConnectToMaster: sending RPC to server master-127.1.11.62:46867, [497ms] 
Sub RPC ConnectToMaster: sending RPC to server master-127.1.11.60:40051, 
[497ms] Sub RPC ConnectToMaster: sending RPC to server 
master-127.1.11.61:46759, [500ms] Sub RPC ConnectToMaster: received response 
from server master-127.1.11.61:46759: Network error: Failed to connect to peer 
master-127.1.11.61:46759(127.1.11.61:46759): Connection ref

[jira] [Assigned] (KUDU-3628) ksck_->CheckMasterConsensus() is flaky in TSAN build

2024-11-26 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3628?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3628:


Assignee: Bakai Ádám

> ksck_->CheckMasterConsensus() is flaky in TSAN build
> 
>
> Key: KUDU-3628
> URL: https://issues.apache.org/jira/browse/KUDU-3628
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3628) ksck_->CheckMasterConsensus() is flaky in TSAN build

2024-11-26 Thread Jira
Bakai Ádám created KUDU-3628:


 Summary: ksck_->CheckMasterConsensus() is flaky in TSAN build
 Key: KUDU-3628
 URL: https://issues.apache.org/jira/browse/KUDU-3628
 Project: Kudu
  Issue Type: Sub-task
Reporter: Bakai Ádám






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (KUDU-3628) ksck_->CheckMasterConsensus() is flaky in TSAN build

2024-12-03 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3628?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17902633#comment-17902633
 ] 

Bakai Ádám commented on KUDU-3628:
--

This issue is a manifestation of KUDU-2709. 

> ksck_->CheckMasterConsensus() is flaky in TSAN build
> 
>
> Key: KUDU-3628
> URL: https://issues.apache.org/jira/browse/KUDU-3628
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (KUDU-3631) TSAN errors in tests when using InternalMiniCluster

2024-12-03 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3631?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17902678#comment-17902678
 ] 

Bakai Ádám commented on KUDU-3631:
--

One example:
{code:java}
WARNING: ThreadSanitizer: destroy of a locked mutex (pid=411)
    #0 pthread_mutex_destroy 
/root/tmp/test123/kudu/thirdparty/src/llvm-11.0.0.src/projects/compiler-rt/lib/tsan/rtl/tsan_interceptors_posix.cpp:1241
 (ksck_remote-test+0x308066)
    #1 kudu::Mutex::~Mutex() 
/root/tmp/test123/kudu/src/kudu/util/mutex.cc:76:12 (libkudu_util.so+0x3df980)
    #2 kudu::CountDownLatch::~CountDownLatch() 
/root/tmp/test123/kudu/src/kudu/util/countdown_latch.h:33:7 
(ksck_remote-test+0x372262)
    #3 kudu::server::ServerBase::~ServerBase() 
/root/tmp/test123/kudu/src/kudu/server/server_base.cc:772:1 
(libserver_process.so+0x1427ef)
    #4 kudu::kserver::KuduServer::~KuduServer() 
/root/tmp/test123/kudu/src/kudu/kserver/kserver.h:41:7 (libmaster.so+0x400b3f)
    #5 kudu::master::Master::~Master() 
/root/tmp/test123/kudu/src/kudu/master/master.cc:263:1 (libmaster.so+0x3f6dc2)
    #6 kudu::master::Master::~Master() 
/root/tmp/test123/kudu/src/kudu/master/master.cc:261:19 (libmaster.so+0x3f70b9)
    #7 
std::__1::default_delete::operator()(kudu::master::Master*)
 const 
/root/tmp/test123/kudu/thirdparty/installed/tsan/include/c++/v1/memory:2262:5 
(libmaster.so+0x434907)
    #8 std::__1::unique_ptr >::reset(kudu::master::Master*) 
/root/tmp/test123/kudu/thirdparty/installed/tsan/include/c++/v1/memory:2517:7 
(libmaster.so+0x42d9cd)
    #9 kudu::master::MiniMaster::Shutdown() 
/root/tmp/test123/kudu/src/kudu/master/mini_master.cc:120:13 
(libmaster.so+0x4c18c2)
    #10 kudu::tools::RemoteKsckTest_TestClusterWithLocation_Test::TestBody() 
/root/tmp/test123/kudu/src/kudu/tools/ksck_remote-test.cc:612:13 
(ksck_remote-test+0x35a1f8)
    #11 void 
testing::internal::HandleSehExceptionsInMethodIfSupported(testing::Test*, void (testing::Test::*)(), char const*) 
/root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2599:10
 (libgtest.so.1.12.1+0x64dbf)
    #12 void 
testing::internal::HandleExceptionsInMethodIfSupported(testing::Test*, void (testing::Test::*)(), char const*) 
/root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2635:14
 (libgtest.so.1.12.1+0x64dbf)
    #13 testing::Test::Run() 
/root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2674:5
 (libgtest.so.1.12.1+0x429c1)
    #14 testing::TestInfo::Run() 
/root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2853:11
 (libgtest.so.1.12.1+0x43cd8)
    #15 testing::TestSuite::Run() 
/root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:3012:30
 (libgtest.so.1.12.1+0x44cb4)
    #16 testing::internal::UnitTestImpl::RunAllTests() 
/root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:5870:44
 (libgtest.so.1.12.1+0x597a4)
    #17 bool 
testing::internal::HandleSehExceptionsInMethodIfSupported(testing::internal::UnitTestImpl*, bool 
(testing::internal::UnitTestImpl::*)(), char const*) 
/root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2599:10
 (libgtest.so.1.12.1+0x65c7f)
    #18 bool 
testing::internal::HandleExceptionsInMethodIfSupported(testing::internal::UnitTestImpl*, bool 
(testing::internal::UnitTestImpl::*)(), char const*) 
/root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:2635:14
 (libgtest.so.1.12.1+0x65c7f)
    #19 testing::UnitTest::Run() 
/root/tmp/test123/kudu/thirdparty/src/googletest-release-1.12.1/googletest/src/gtest.cc:5444:10
 (libgtest.so.1.12.1+0x58d5c)
    #20 RUN_ALL_TESTS() 
/root/tmp/test123/kudu/thirdparty/installed/tsan/include/gtest/gtest.h:2293:73 
(ksck_remote-test+0x38d91b)
    #21 main /root/tmp/test123/kudu/src/kudu/util/test_main.cc:109:10 
(ksck_remote-test+0x38c81c)  and:
    #0 pthread_cond_timedwait 
/root/tmp/test123/kudu/thirdparty/src/llvm-11.0.0.src/projects/compiler-rt/lib/tsan/rtl/tsan_interceptors_posix.cpp:1181
 (ksck_remote-test+0x2f5550)
    #1 kudu::ConditionVariable::WaitUntil(kudu::MonoTime const&) const 
/root/tmp/test123/kudu/src/kudu/util/condition_variable.cc:87:12 
(libkudu_util.so+0x2c400a)
    #2 kudu::CountDownLatch::WaitUntil(kudu::MonoTime const&) const 
/root/tmp/test123/kudu/src/kudu/util/countdown_latch.h:89:18 
(ksck_remote-test+0x38abe8)
    #3 kudu::CountDownLatch::WaitFor(kudu::MonoDelta const&) const 
/root/tmp/test123/kudu/src/kudu/util/countdown_latch.h:99:12 
(ksck_remote-test+0x38ab5e)
    #4 kudu::master::Master::ExpiredReservedTablesDeleterThread() 
/root/tmp/test123/kudu/src/kudu/master/master.cc:568:42 (libmaster.so+0x3f975e)
    #5 
kudu::master::Master::StartExpiredReservedTablesDeleterThread()::$_2::operator()()
 const /root/tmp/te

[jira] [Created] (KUDU-3631) TSAN errors in tests related to InternalMiniCluster

2024-12-03 Thread Jira
Bakai Ádám created KUDU-3631:


 Summary: TSAN errors in tests related to InternalMiniCluster
 Key: KUDU-3631
 URL: https://issues.apache.org/jira/browse/KUDU-3631
 Project: Kudu
  Issue Type: Sub-task
Reporter: Bakai Ádám






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3631) TSAN errors in tests when using InternalMiniCluster

2024-12-03 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3631:
-
Summary: TSAN errors in tests when using InternalMiniCluster  (was: TSAN 
errors in tests related to InternalMiniCluster)

> TSAN errors in tests when using InternalMiniCluster
> ---
>
> Key: KUDU-3631
> URL: https://issues.apache.org/jira/browse/KUDU-3631
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3631) TSAN error in CountDownLatch destructor when using InternalMiniCluster

2024-12-05 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3631:
-
Summary: TSAN error in CountDownLatch destructor when using 
InternalMiniCluster  (was: TSAN errors in tests when using InternalMiniCluster)

> TSAN error in CountDownLatch destructor when using InternalMiniCluster
> --
>
> Key: KUDU-3631
> URL: https://issues.apache.org/jira/browse/KUDU-3631
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (KUDU-3633) Threadpool check flakiness in ksck_remote-test during MiniMaster shutdown

2024-12-05 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3633?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3633:


Assignee: Bakai Ádám

> Threadpool check flakiness in ksck_remote-test during MiniMaster shutdown
> -
>
> Key: KUDU-3633
> URL: https://issues.apache.org/jira/browse/KUDU-3633
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>
> {code:java}
> F20241204 12:57:40.147302 16123 threadpool.cc:391] Check failed: 1 == 
> tokens_.size() (1 vs. 3) Threadpool raft destroyed with 3 allocated tokens
> *** Check failure stack trace: ***\{code}
> {code:java}
>     @     0x7f6b96b2cd64 google::LogMessage::SendToLog() at ??:0
>     @     0x7f6b96b2d910 google::LogMessage::Flush() at ??:0
>     @     0x7f6b96b32a4b google::LogMessageFatal::~LogMessageFatal() at ??:0
>     @     0x7f6b974a777d kudu::ThreadPool::~ThreadPool() at ??:0
> I20241204 12:57:40.556027 23288 raft_consensus.cc:1270] T 
> df574f38d0a746d1929d9494d82da991 P c273df5d41694d4da3bc1b5bc5e81b84 [term 2 
> FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: 
> Log matching property violated. Preceding OpId in replica: term: 1 index: 1. 
> Preceding OpId from leader: term: 2 index: 2. (index mismatch)
> I20241204 12:57:40.558073 23666 consensus_queue.cc:1035] T 
> df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: 
> Connected to new peer: Peer: permanent_uuid: 
> "c273df5d41694d4da3bc1b5bc5e81b84" member_type: VOTER last_known_addr { host: 
> "127.15.190.193" port: 33967 }, Status: LMP_MISMATCH, Last received: 0.0, 
> Next index: 2, Last known committed idx: 1, Time since last communication: 
> 0.000s
>     @     0x7f6b9ff4f6bf std::__1::default_delete<>::operator()() at ??:0
> I20241204 12:57:40.605798 23460 raft_consensus.cc:1270] T 
> df574f38d0a746d1929d9494d82da991 P 87f06d0d674a4791871f81a7af62b7be [term 2 
> FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: 
> Log matching property violated. Preceding OpId in replica: term: 1 index: 1. 
> Preceding OpId from leader: term: 2 index: 2. (index mismatch)
> I20241204 12:57:40.611544 23707 consensus_queue.cc:1035] T 
> df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: 
> Connected to new peer: Peer: permanent_uuid: 
> "87f06d0d674a4791871f81a7af62b7be" member_type: VOTER last_known_addr { host: 
> "127.15.190.195" port: 35365 }, Status: LMP_MISMATCH, Last received: 0.0, 
> Next index: 2, Last known committed idx: 1, Time since last communication: 
> 0.000s
>     @     0x7f6b9ff4f62e std::__1::unique_ptr<>::reset() at ??:0
>     @     0x7f6b9ff0e2cc std::__1::unique_ptr<>::~unique_ptr() at ??:0
>     @     0x7f6b9ffb65b4 kudu::kserver::KuduServer::~KuduServer() at ??:0
>     @     0x7f6b9ffac863 kudu::master::Master::~Master() at ??:0
>     @     0x7f6b9ffacb5a kudu::master::Master::~Master() at ??:0
>     @     0x7f6b9ffea408 std::__1::default_delete<>::operator()() at ??:0
>     @     0x7f6b9ffe34ce std::__1::unique_ptr<>::reset() at ??:0
>     @     0x7f6ba00773c3 kudu::master::MiniMaster::Shutdown() at ??:0
>     @           0x354ea9 
> kudu::tools::RemoteKsckTest_TestClusterWithLocation_Test::TestBody() at 
> /root/tmp/test123/kudu/src/kudu/tools/ksck_remote-test.cc:607
>     @     0x7f6ba045adc0 
> testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0
>     @     0x7f6ba04389c2 testing::Test::Run() at ??:0
>     @     0x7f6ba0439cd9 testing::TestInfo::Run() at ??:0
>     @     0x7f6ba043acb5 testing::TestSuite::Run() at ??:0
>     @     0x7f6ba044f7a5 testing::internal::UnitTestImpl::RunAllTests() at 
> ??:0
>     @     0x7f6ba045bc80 
> testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0
>     @     0x7f6ba044ed5d testing::UnitTest::Run() at ??:0
>     @           0x3801bc RUN_ALL_TESTS() at 
> /root/tmp/test123/kudu/thirdparty/installed/tsan/include/gtest/gtest.h:?
>     @           0x37f0bd main at 
> /root/tmp/test123/kudu/src/kudu/util/test_main.cc:?
>     @     0x7f6b93f58bf7 __libc_start_main at ??:0
>     @           0x298ada _start at ??:? {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3633) Threadpool check flakiness in ksck_remote-test during MiniMaster shutdown

2024-12-04 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3633?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3633:
-
Summary: Threadpool check flakiness in ksck_remote-test during MiniMaster 
shutdown  (was: Threadpool check flakiness in ksck_remote-test)

> Threadpool check flakiness in ksck_remote-test during MiniMaster shutdown
> -
>
> Key: KUDU-3633
> URL: https://issues.apache.org/jira/browse/KUDU-3633
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Priority: Major
>
> {code:java}
> F20241204 12:57:40.147302 16123 threadpool.cc:391] Check failed: 1 == 
> tokens_.size() (1 vs. 3) Threadpool raft destroyed with 3 allocated tokens
> *** Check failure stack trace: ***\{code}
> {code:java}
>     @     0x7f6b96b2cd64 google::LogMessage::SendToLog() at ??:0
>     @     0x7f6b96b2d910 google::LogMessage::Flush() at ??:0
>     @     0x7f6b96b32a4b google::LogMessageFatal::~LogMessageFatal() at ??:0
>     @     0x7f6b974a777d kudu::ThreadPool::~ThreadPool() at ??:0
> I20241204 12:57:40.556027 23288 raft_consensus.cc:1270] T 
> df574f38d0a746d1929d9494d82da991 P c273df5d41694d4da3bc1b5bc5e81b84 [term 2 
> FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: 
> Log matching property violated. Preceding OpId in replica: term: 1 index: 1. 
> Preceding OpId from leader: term: 2 index: 2. (index mismatch)
> I20241204 12:57:40.558073 23666 consensus_queue.cc:1035] T 
> df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: 
> Connected to new peer: Peer: permanent_uuid: 
> "c273df5d41694d4da3bc1b5bc5e81b84" member_type: VOTER last_known_addr { host: 
> "127.15.190.193" port: 33967 }, Status: LMP_MISMATCH, Last received: 0.0, 
> Next index: 2, Last known committed idx: 1, Time since last communication: 
> 0.000s
>     @     0x7f6b9ff4f6bf std::__1::default_delete<>::operator()() at ??:0
> I20241204 12:57:40.605798 23460 raft_consensus.cc:1270] T 
> df574f38d0a746d1929d9494d82da991 P 87f06d0d674a4791871f81a7af62b7be [term 2 
> FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: 
> Log matching property violated. Preceding OpId in replica: term: 1 index: 1. 
> Preceding OpId from leader: term: 2 index: 2. (index mismatch)
> I20241204 12:57:40.611544 23707 consensus_queue.cc:1035] T 
> df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: 
> Connected to new peer: Peer: permanent_uuid: 
> "87f06d0d674a4791871f81a7af62b7be" member_type: VOTER last_known_addr { host: 
> "127.15.190.195" port: 35365 }, Status: LMP_MISMATCH, Last received: 0.0, 
> Next index: 2, Last known committed idx: 1, Time since last communication: 
> 0.000s
>     @     0x7f6b9ff4f62e std::__1::unique_ptr<>::reset() at ??:0
>     @     0x7f6b9ff0e2cc std::__1::unique_ptr<>::~unique_ptr() at ??:0
>     @     0x7f6b9ffb65b4 kudu::kserver::KuduServer::~KuduServer() at ??:0
>     @     0x7f6b9ffac863 kudu::master::Master::~Master() at ??:0
>     @     0x7f6b9ffacb5a kudu::master::Master::~Master() at ??:0
>     @     0x7f6b9ffea408 std::__1::default_delete<>::operator()() at ??:0
>     @     0x7f6b9ffe34ce std::__1::unique_ptr<>::reset() at ??:0
>     @     0x7f6ba00773c3 kudu::master::MiniMaster::Shutdown() at ??:0
>     @           0x354ea9 
> kudu::tools::RemoteKsckTest_TestClusterWithLocation_Test::TestBody() at 
> /root/tmp/test123/kudu/src/kudu/tools/ksck_remote-test.cc:607
>     @     0x7f6ba045adc0 
> testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0
>     @     0x7f6ba04389c2 testing::Test::Run() at ??:0
>     @     0x7f6ba0439cd9 testing::TestInfo::Run() at ??:0
>     @     0x7f6ba043acb5 testing::TestSuite::Run() at ??:0
>     @     0x7f6ba044f7a5 testing::internal::UnitTestImpl::RunAllTests() at 
> ??:0
>     @     0x7f6ba045bc80 
> testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0
>     @     0x7f6ba044ed5d testing::UnitTest::Run() at ??:0
>     @           0x3801bc RUN_ALL_TESTS() at 
> /root/tmp/test123/kudu/thirdparty/installed/tsan/include/gtest/gtest.h:?
>     @           0x37f0bd main at 
> /root/tmp/test123/kudu/src/kudu/util/test_main.cc:?
>     @     0x7f6b93f58bf7 __libc_start_main at ??:0
>     @           0x298ada _start at ??:? {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3633) Threadpool check flakiness in ksck_remote-test

2024-12-04 Thread Jira
Bakai Ádám created KUDU-3633:


 Summary: Threadpool check flakiness in ksck_remote-test
 Key: KUDU-3633
 URL: https://issues.apache.org/jira/browse/KUDU-3633
 Project: Kudu
  Issue Type: Sub-task
Reporter: Bakai Ádám






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3633) Threadpool check flakiness in ksck_remote-test

2024-12-04 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3633?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3633:
-
Description: 
{code:java}
F20241204 12:57:40.147302 16123 threadpool.cc:391] Check failed: 1 == 
tokens_.size() (1 vs. 3) Threadpool raft destroyed with 3 allocated tokens
*** Check failure stack trace: ***\{code}
{code:java}
    @     0x7f6b96b2cd64 google::LogMessage::SendToLog() at ??:0
    @     0x7f6b96b2d910 google::LogMessage::Flush() at ??:0
    @     0x7f6b96b32a4b google::LogMessageFatal::~LogMessageFatal() at ??:0
    @     0x7f6b974a777d kudu::ThreadPool::~ThreadPool() at ??:0
I20241204 12:57:40.556027 23288 raft_consensus.cc:1270] T 
df574f38d0a746d1929d9494d82da991 P c273df5d41694d4da3bc1b5bc5e81b84 [term 2 
FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: 
Log matching property violated. Preceding OpId in replica: term: 1 index: 1. 
Preceding OpId from leader: term: 2 index: 2. (index mismatch)
I20241204 12:57:40.558073 23666 consensus_queue.cc:1035] T 
df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: 
Connected to new peer: Peer: permanent_uuid: "c273df5d41694d4da3bc1b5bc5e81b84" 
member_type: VOTER last_known_addr { host: "127.15.190.193" port: 33967 }, 
Status: LMP_MISMATCH, Last received: 0.0, Next index: 2, Last known committed 
idx: 1, Time since last communication: 0.000s
    @     0x7f6b9ff4f6bf std::__1::default_delete<>::operator()() at ??:0
I20241204 12:57:40.605798 23460 raft_consensus.cc:1270] T 
df574f38d0a746d1929d9494d82da991 P 87f06d0d674a4791871f81a7af62b7be [term 2 
FOLLOWER]: Refusing update from remote peer 2e54eeefd5f947279415fb606d3fe035: 
Log matching property violated. Preceding OpId in replica: term: 1 index: 1. 
Preceding OpId from leader: term: 2 index: 2. (index mismatch)
I20241204 12:57:40.611544 23707 consensus_queue.cc:1035] T 
df574f38d0a746d1929d9494d82da991 P 2e54eeefd5f947279415fb606d3fe035 [LEADER]: 
Connected to new peer: Peer: permanent_uuid: "87f06d0d674a4791871f81a7af62b7be" 
member_type: VOTER last_known_addr { host: "127.15.190.195" port: 35365 }, 
Status: LMP_MISMATCH, Last received: 0.0, Next index: 2, Last known committed 
idx: 1, Time since last communication: 0.000s
    @     0x7f6b9ff4f62e std::__1::unique_ptr<>::reset() at ??:0
    @     0x7f6b9ff0e2cc std::__1::unique_ptr<>::~unique_ptr() at ??:0
    @     0x7f6b9ffb65b4 kudu::kserver::KuduServer::~KuduServer() at ??:0
    @     0x7f6b9ffac863 kudu::master::Master::~Master() at ??:0
    @     0x7f6b9ffacb5a kudu::master::Master::~Master() at ??:0
    @     0x7f6b9ffea408 std::__1::default_delete<>::operator()() at ??:0
    @     0x7f6b9ffe34ce std::__1::unique_ptr<>::reset() at ??:0
    @     0x7f6ba00773c3 kudu::master::MiniMaster::Shutdown() at ??:0
    @           0x354ea9 
kudu::tools::RemoteKsckTest_TestClusterWithLocation_Test::TestBody() at 
/root/tmp/test123/kudu/src/kudu/tools/ksck_remote-test.cc:607
    @     0x7f6ba045adc0 
testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0
    @     0x7f6ba04389c2 testing::Test::Run() at ??:0
    @     0x7f6ba0439cd9 testing::TestInfo::Run() at ??:0
    @     0x7f6ba043acb5 testing::TestSuite::Run() at ??:0
    @     0x7f6ba044f7a5 testing::internal::UnitTestImpl::RunAllTests() at ??:0
    @     0x7f6ba045bc80 
testing::internal::HandleExceptionsInMethodIfSupported<>() at ??:0
    @     0x7f6ba044ed5d testing::UnitTest::Run() at ??:0
    @           0x3801bc RUN_ALL_TESTS() at 
/root/tmp/test123/kudu/thirdparty/installed/tsan/include/gtest/gtest.h:?
    @           0x37f0bd main at 
/root/tmp/test123/kudu/src/kudu/util/test_main.cc:?
    @     0x7f6b93f58bf7 __libc_start_main at ??:0
    @           0x298ada _start at ??:? {code}

> Threadpool check flakiness in ksck_remote-test
> ------
>
> Key: KUDU-3633
> URL: https://issues.apache.org/jira/browse/KUDU-3633
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Priority: Major
>
> {code:java}
> F20241204 12:57:40.147302 16123 threadpool.cc:391] Check failed: 1 == 
> tokens_.size() (1 vs. 3) Threadpool raft destroyed with 3 allocated tokens
> *** Check failure stack trace: ***\{code}
> {code:java}
>     @     0x7f6b96b2cd64 google::LogMessage::SendToLog() at ??:0
>     @     0x7f6b96b2d910 google::LogMessage::Flush() at ??:0
>     @     0x7f6b96b32a4b google::LogMessageFatal::~LogMessageFatal() at ??:0
>     @     0x7f6b974a777d kudu::ThreadPool::~ThreadPool() at ??:0
> I20241204 12:57:40.556027 23288 raft_consensus.cc:1270] T 
> df574f38d0a746d1929d9494d82da991 P c273df5d41694d4da3bc1b5bc5e81b84 [term 2 
> FOLLOWER]: Refusing update from remote peer 2e54eeefd5f94

[jira] [Resolved] (KUDU-3563) Output tablet-level metrics in Prometheus format

2025-02-11 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3563?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám resolved KUDU-3563.
--
Fix Version/s: 1.18.0
   Resolution: Fixed

> Output tablet-level metrics in Prometheus format
> 
>
> Key: KUDU-3563
> URL: https://issues.apache.org/jira/browse/KUDU-3563
> Project: Kudu
>  Issue Type: Improvement
>  Components: master, server, tserver
>Reporter: Alexey Serbin
>Assignee: Bakai Ádám
>Priority: Major
>  Labels: metrics, prometheus, supportability, visibility
> Fix For: 1.18.0
>
>
> The request to support outputting Kudu metrics in Prometheus format is 
> tracked [KUDU-3375|https://issues.apache.org/jira/browse/KUDU-3375].  The 
> [first take on 
> this|https://github.com/apache/kudu/commit/00efc6826ac9a1f5d10750296c7357790a041fec]
>  has taken care of the server-level metrics, ignoring all the tablet-level 
> metrics.
> In the scope of this JIRA item, it's necessary to output all the tablet-level 
> metrics in Prometheus format as well.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3634) check error in ksck_results.cc in TSAN build

2024-12-18 Thread Jira
Bakai Ádám created KUDU-3634:


 Summary: check error in ksck_results.cc in TSAN build
 Key: KUDU-3634
 URL: https://issues.apache.org/jira/browse/KUDU-3634
 Project: Kudu
  Issue Type: Sub-task
Reporter: Bakai Ádám






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3634) check error in ksck_results.cc in TSAN build

2024-12-18 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3634:
-
Description: 
{code:java}
F20241218 13:05:04.229065  4953 ksck_results.cc:160] Check failed: 
servers.size() <= server_count (6 vs. 3) 
*** Check failure stack trace: ***
*** Aborted at 1734527104 (unix time) try "date -d @1734527104" if you are 
using GNU date ***
PC: @                0x0 (unknown)
*** SIGABRT (@0x3e81359) received by PID 4953 (TID 0x7f5270649880) from PID 
4953; stack trace: ***
    @           0x2d8140 __tsan::CallUserSignalHandler() at 
/root/tmp/test123/kudu/thirdparty/src/llvm-11.0.0.src/projects/compiler-rt/lib/tsan/rtl/tsan_interceptors_posix.cpp:1915
    @           0x2da623 rtl_sigaction() at 
/root/tmp/test123/kudu/thirdparty/src/llvm-11.0.0.src/projects/compiler-rt/lib/tsan/rtl/tsan_interceptors_posix.cpp:2005
I20241218 13:05:04.427536 14086 raft_consensus.cc:1270] T 
5d3037cff4d843b0a7dbc9ff78a339ba P 3e6bbc8984eb4fbfbefd27d5036db51a [term 5 
FOLLOWER]: Refusing update from remote peer 1a37ef9543954c8e823eeae2d8592107: 
Log matching property violated. Preceding OpId in replica: term: 4 index: 3. 
Preceding OpId from leader: term: 5 index: 4. (index mismatch)
I20241218 13:05:04.429109 14549 consensus_queue.cc:1035] T 
5d3037cff4d843b0a7dbc9ff78a339ba P 1a37ef9543954c8e823eeae2d8592107 [LEADER]: 
Connected to new peer: Peer: permanent_uuid: "3e6bbc8984eb4fbfbefd27d5036db51a" 
member_type: VOTER last_known_addr { host: "127.4.214.67" port: 37251 }, 
Status: LMP_MISMATCH, Last received: 0.0, Next index: 4, Last known committed 
idx: 3, Time since last communication: 0.000s
    @     0x7f527808f980 (unknown) at ??:0
I20241218 13:05:04.509820 14545 raft_consensus.cc:491] T 
5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a [term 5 
FOLLOWER]: Starting pre-election (no leader contacted us within the election 
timeout)
I20241218 13:05:04.510313 14545 raft_consensus.cc:513] T 
5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a [term 5 
FOLLOWER]: Starting pre-election with config: opid_index: -1 OBSOLETE_local: 
false peers { permanent_uuid: "0bfce15bbb5f4f0aa5a9e60872491f3a" member_type: 
VOTER last_known_addr { host: "127.4.214.65" port: 36387 } } peers { 
permanent_uuid: "1a37ef9543954c8e823eeae2d8592107" member_type: VOTER 
last_known_addr { host: "127.4.214.66" port: 36001 } } peers { permanent_uuid: 
"3e6bbc8984eb4fbfbefd27d5036db51a" member_type: VOTER last_known_addr { host: 
"127.4.214.67" port: 37251 } }
I20241218 13:05:04.513247 14001 tablet_service.cc:1812] Received 
RequestConsensusVote() RPC: tablet_id: "5d3037cff4d843b0a7dbc9ff78a339ba" 
candidate_uuid: "0bfce15bbb5f4f0aa5a9e60872491f3a" candidate_term: 6 
candidate_status { last_received { term: 3 index: 2 } } ignore_live_leader: 
false dest_uuid: "1a37ef9543954c8e823eeae2d8592107" is_pre_election: true
I20241218 13:05:04.516223 14089 tablet_service.cc:1812] Received 
RequestConsensusVote() RPC: tablet_id: "5d3037cff4d843b0a7dbc9ff78a339ba" 
candidate_uuid: "0bfce15bbb5f4f0aa5a9e60872491f3a" candidate_term: 6 
candidate_status { last_received { term: 3 index: 2 } } ignore_live_leader: 
false dest_uuid: "3e6bbc8984eb4fbfbefd27d5036db51a" is_pre_election: true
I20241218 13:05:04.517762 13891 leader_election.cc:304] T 
5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a 
[CANDIDATE]: Term 6 pre-election: Election decided. Result: candidate lost. 
Election summary: received 3 responses out of 3 voters: 1 yes votes; 2 no 
votes. yes voters: 0bfce15bbb5f4f0aa5a9e60872491f3a; no voters: 
1a37ef9543954c8e823eeae2d8592107, 3e6bbc8984eb4fbfbefd27d5036db51a
I20241218 13:05:04.518554 14506 raft_consensus.cc:2743] T 
5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a [term 5 
FOLLOWER]: Leader pre-election lost for term 6. Reason: could not achieve 
majority
I20241218 13:05:04.518522 14545 leader_election.cc:290] T 
5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a 
[CANDIDATE]: Term 6 pre-election: Requested pre-vote from peers 
1a37ef9543954c8e823eeae2d8592107 (127.4.214.66:36001), 
3e6bbc8984eb4fbfbefd27d5036db51a (127.4.214.67:37251)
    @     0x7f5271309fb7 gsignal at ??:0
I20241218 13:05:04.633380 13919 raft_consensus.cc:1270] T 
5d3037cff4d843b0a7dbc9ff78a339ba P 0bfce15bbb5f4f0aa5a9e60872491f3a [term 5 
FOLLOWER]: Refusing update from remote peer 1a37ef9543954c8e823eeae2d8592107: 
Log matching property violated. Preceding OpId in replica: term: 3 index: 2. 
Preceding OpId from leader: term: 5 index: 4. (index mismatch)
I20241218 13:05:04.634791 14549 consensus_queue.cc:1035] T 
5d3037cff4d843b0a7dbc9ff78a339ba P 1a37ef9543954c8e823eeae2d8592107 [LEADER]: 
Connected to new peer: Peer: permanent_uui

[jira] [Assigned] (KUDU-3631) TSAN error in CountDownLatch destructor when using InternalMiniCluster

2024-12-18 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3631:


Assignee: Bakai Ádám

> TSAN error in CountDownLatch destructor when using InternalMiniCluster
> --
>
> Key: KUDU-3631
> URL: https://issues.apache.org/jira/browse/KUDU-3631
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (KUDU-3637) TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky

2025-01-10 Thread Jira
Bakai Ádám created KUDU-3637:


 Summary: 
TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky
 Key: KUDU-3637
 URL: https://issues.apache.org/jira/browse/KUDU-3637
 Project: Kudu
  Issue Type: Sub-task
Reporter: Bakai Ádám






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (KUDU-3637) TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky

2025-01-10 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám reassigned KUDU-3637:


Assignee: Bakai Ádám

> TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky
> ---
>
> Key: KUDU-3637
> URL: https://issues.apache.org/jira/browse/KUDU-3637
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (KUDU-3637) TestKuduTransaction is flaky

2025-01-14 Thread Jira


 [ 
https://issues.apache.org/jira/browse/KUDU-3637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bakai Ádám updated KUDU-3637:
-
Summary: TestKuduTransaction is flaky  (was: 
TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky)

> TestKuduTransaction is flaky
> 
>
> Key: KUDU-3637
> URL: https://issues.apache.org/jira/browse/KUDU-3637
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (KUDU-3637) TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky

2025-01-14 Thread Jira


[ 
https://issues.apache.org/jira/browse/KUDU-3637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17912931#comment-17912931
 ] 

Bakai Ádám commented on KUDU-3637:
--

In testTxnKeepaliveSwitchesToOtherTxnManager (keep in mind, this is the not 
rolling version)
According to the logs, for some reason the operation was aborted:
{code:java}
13:39:39.728 [INFO - cluster stderr printer] (MiniKuduCluster.java:751) 
W20250113 13:39:39.728489 26745 tablet_replica.cc:1307] Aborted: operation has 
been aborted: cancelling pending write operations {code}
I have not found yet the reason why. 

> TestKuduTransaction.testTxnKeepaliveRollingSwitchToOtherTxnManager is flaky
> ---
>
> Key: KUDU-3637
> URL: https://issues.apache.org/jira/browse/KUDU-3637
> Project: Kudu
>  Issue Type: Sub-task
>Reporter: Bakai Ádám
>Assignee: Bakai Ádám
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


  1   2   3   4   5   6   7   8   9   10   >