frankvicky commented on code in PR #19793:
URL: https://github.com/apache/kafka/pull/19793#discussion_r2110852699


##########
server/src/main/java/org/apache/kafka/server/DelayedProduce.java:
##########
@@ -0,0 +1,235 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ *    http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.kafka.server;
+
+import org.apache.kafka.common.TopicIdPartition;
+import org.apache.kafka.common.TopicPartition;
+import org.apache.kafka.common.protocol.Errors;
+import org.apache.kafka.common.requests.ProduceResponse.PartitionResponse;
+import org.apache.kafka.server.metrics.KafkaMetricsGroup;
+import org.apache.kafka.server.purgatory.DelayedOperation;
+
+import com.yammer.metrics.core.Meter;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import java.util.Map;
+import java.util.Objects;
+import java.util.concurrent.ConcurrentHashMap;
+import java.util.concurrent.TimeUnit;
+import java.util.function.BiConsumer;
+import java.util.function.Consumer;
+import java.util.stream.Collectors;
+
+/**
+ * A delayed produce operation that can be created by the replica manager and 
watched
+ * in the produce operation purgatory
+ */
+public class DelayedProduce extends DelayedOperation {
+    private static final Logger LOGGER = 
LoggerFactory.getLogger(DelayedProduce.class);
+
+    public static final class ProducePartitionStatus {
+        private final long requiredOffset;
+        private final PartitionResponse responseStatus;
+
+        private volatile boolean acksPending;
+
+        public ProducePartitionStatus(long requiredOffset, PartitionResponse 
responseStatus) {
+            this.requiredOffset = requiredOffset;
+            this.responseStatus = responseStatus;
+        }
+
+        public long requiredOffset() {
+            return requiredOffset;
+        }
+
+        public PartitionResponse responseStatus() {
+            return responseStatus;
+        }
+
+        public boolean acksPending() {
+            return acksPending;
+        }
+
+        public void setAcksPending(boolean acksPending) {
+            this.acksPending = acksPending;
+        }
+
+        @Override
+        public String toString() {
+            return String.format(
+                    "[acksPending: %s, error: %s, startOffset: %s, 
requiredOffset: %d]",
+                    acksPending,
+                    responseStatus.error.code(),
+                    responseStatus.baseOffset,
+                    requiredOffset
+            );
+        }
+
+        @Override
+        public boolean equals(Object o) {
+            if (o == null || getClass() != o.getClass()) {
+                return false;
+            }
+            final ProducePartitionStatus that = (ProducePartitionStatus) o;
+            return requiredOffset == that.requiredOffset && acksPending == 
that.acksPending && Objects.equals(responseStatus, that.responseStatus);
+        }
+
+        @Override
+        public int hashCode() {
+            return Objects.hash(requiredOffset, responseStatus, acksPending);
+        }
+    }
+
+    /**
+     * The produce metadata maintained by the delayed produce operation
+     */
+    public static final class ProduceMetadata {
+        private final short produceRequiredAcks;
+        private final Map<TopicIdPartition, ProducePartitionStatus> 
produceStatus;
+
+        public ProduceMetadata(short produceRequiredAcks,
+                               Map<TopicIdPartition, ProducePartitionStatus> 
produceStatus) {
+            this.produceRequiredAcks = produceRequiredAcks;
+            this.produceStatus = produceStatus;
+        }
+
+        @Override
+        public String toString() {
+            return String.format(
+                    "[requiredAcks: %d, partitionStatus: %s]",
+                    produceRequiredAcks,
+                    produceStatus
+            );
+        }
+
+        @Override
+        public boolean equals(Object o) {
+            if (o == null || getClass() != o.getClass()) {
+                return false;
+            }
+            final ProduceMetadata that = (ProduceMetadata) o;
+            return produceRequiredAcks == that.produceRequiredAcks && 
Objects.equals(produceStatus, that.produceStatus);
+        }
+
+        @Override
+        public int hashCode() {
+            return Objects.hash(produceRequiredAcks, produceStatus);
+        }
+    }
+
+    private static final class DelayedProduceMetrics {
+        private static final KafkaMetricsGroup METRICS_GROUP = new 
KafkaMetricsGroup(DelayedProduceMetrics.class);
+        private static final Meter AGGREGATE_EXPIRATION_METER = 
METRICS_GROUP.newMeter("ExpiresPerSec", "requests", TimeUnit.SECONDS);
+        private static final ConcurrentHashMap<TopicPartition, Meter> 
PARTITION_EXPIRATION_METERS = new ConcurrentHashMap<>();
+
+        private static void recordExpiration(TopicPartition partition) {
+            AGGREGATE_EXPIRATION_METER.mark();
+            PARTITION_EXPIRATION_METERS.computeIfAbsent(partition,
+                    key -> METRICS_GROUP.newMeter("ExpiresPerSec", "requests", 
TimeUnit.SECONDS,
+                                        Map.of("topic", key.topic(), 
"partition", String.valueOf(key.partition()))))
+                    .mark();
+        }
+    }
+
+    private final ProduceMetadata produceMetadata;
+    private final BiConsumer<TopicPartition, ProducePartitionStatus> 
updateProducePartitionStatusCallback;
+    private final Consumer<Map<TopicIdPartition, PartitionResponse>> 
responseCallback;
+
+    public DelayedProduce(long delayMs,
+                          ProduceMetadata produceMetadata,
+                          BiConsumer<TopicPartition, ProducePartitionStatus> 
updateProducePartitionStatusCallback,
+                          Consumer<Map<TopicIdPartition, PartitionResponse>> 
responseCallback) {
+        super(delayMs);
+
+        this.produceMetadata = produceMetadata;
+        this.updateProducePartitionStatusCallback = 
updateProducePartitionStatusCallback;
+        this.responseCallback = responseCallback;
+
+        // first update the acks pending variable according to the error code
+        produceMetadata.produceStatus.forEach((topicPartition, status) -> {
+            if (status.responseStatus.error == Errors.NONE) {
+                // Timeout error state will be cleared when required acks are 
received
+                status.acksPending = true;
+                status.responseStatus.error = Errors.REQUEST_TIMED_OUT;
+            } else {
+                status.acksPending = false;
+            }
+
+            LOGGER.trace("Initial partition status for {} is {}", 
topicPartition, status);
+        });
+    }
+
+    /**
+     * The delayed produce operation can be completed if every partition
+     * it produces to is satisfied by one of the following:
+     *
+     * Case A: Replica not assigned to partition
+     * Case B: Replica is no longer the leader of this partition
+     * Case C: This broker is the leader:
+     *   C.1 - If there was a local error thrown while checking if at least 
requiredAcks
+     *         replicas have caught up to this operation: set an error in 
response
+     *   C.2 - Otherwise, set the response with no error.
+     */
+    @Override
+    public boolean tryComplete() {
+        // check for each partition if it still has pending acks
+        produceMetadata.produceStatus.forEach((topicIdPartition, status) -> {
+            LOGGER.trace("Checking produce satisfaction for {}, current status 
{}", topicIdPartition, status);
+            // skip those partitions that have already been satisfied
+            if (status.acksPending) {
+                
updateProducePartitionStatusCallback.accept(topicIdPartition.topicPartition(), 
status);
+            }
+        });
+
+        // check if every partition has satisfied at least one of case A, B or 
C
+        boolean anyPending = produceMetadata.produceStatus
+                                            .values()
+                                            .stream()
+                                            
.anyMatch(ProducePartitionStatus::acksPending);
+        if (!anyPending) {
+            return forceComplete();
+        }
+
+        return false;

Review Comment:
   I'm okay with it.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: jira-unsubscr...@kafka.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org

Reply via email to