This is an automated email from the ASF dual-hosted git repository.
chrispeck pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/pinot.git
The following commit(s) were added to refs/heads/master by this push:
new a5bc956fb0 [multistage] Handle Column Less Tuples in BlockSplitter
(#16058)
a5bc956fb0 is described below
commit a5bc956fb08d2b9b9b91fb3861041a91d09ad0a4
Author: Ankit Sultana <[email protected]>
AuthorDate: Mon Jun 9 21:56:54 2025 -0500
[multistage] Handle Column Less Tuples in BlockSplitter (#16058)
---
.../main/java/org/apache/pinot/query/runtime/blocks/BlockSplitter.java | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git
a/pinot-query-runtime/src/main/java/org/apache/pinot/query/runtime/blocks/BlockSplitter.java
b/pinot-query-runtime/src/main/java/org/apache/pinot/query/runtime/blocks/BlockSplitter.java
index 28b673be55..9a4195ef32 100644
---
a/pinot-query-runtime/src/main/java/org/apache/pinot/query/runtime/blocks/BlockSplitter.java
+++
b/pinot-query-runtime/src/main/java/org/apache/pinot/query/runtime/blocks/BlockSplitter.java
@@ -67,7 +67,7 @@ public interface BlockSplitter {
// Use estimated row size, this estimate is not accurate and is used to
estimate numRowsPerChunk only.
DataSchema dataSchema = block.getDataSchema();
assert dataSchema != null;
- int estimatedRowSizeInBytes = dataSchema.getColumnNames().length *
MEDIAN_COLUMN_SIZE_BYTES;
+ int estimatedRowSizeInBytes = Math.max(1,
dataSchema.getColumnNames().length * MEDIAN_COLUMN_SIZE_BYTES);
int numRowsPerChunk = maxBlockSize / estimatedRowSizeInBytes;
Preconditions.checkState(numRowsPerChunk > 0, "row size too large for
query engine to handle, abort!");
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]