See <https://builds.apache.org/job/kafka-trunk-jdk8/642/changes>
Changes: [wangguoz] KAFKA-3735: Dispose all RocksObejcts upon completeness ------------------------------------------ [...truncated 4026 lines...] kafka.log.LogConfigTest > testKafkaConfigToProps PASSED kafka.log.LogConfigTest > testFromPropsInvalid PASSED kafka.log.CleanerTest > testBuildOffsetMap PASSED kafka.log.CleanerTest > testBuildOffsetMapFakeLarge PASSED kafka.log.CleanerTest > testSegmentGrouping PASSED kafka.log.CleanerTest > testCleanSegmentsWithAbort PASSED kafka.log.CleanerTest > testSegmentGroupingWithSparseOffsets PASSED kafka.log.CleanerTest > testRecoveryAfterCrash PASSED kafka.log.CleanerTest > testLogToClean PASSED kafka.log.CleanerTest > testCleaningWithDeletes PASSED kafka.log.CleanerTest > testCleanSegments PASSED kafka.log.CleanerTest > testCleaningWithUnkeyedMessages PASSED kafka.log.OffsetIndexTest > lookupExtremeCases PASSED kafka.log.OffsetIndexTest > appendTooMany PASSED kafka.log.OffsetIndexTest > randomLookupTest PASSED kafka.log.OffsetIndexTest > testReopen PASSED kafka.log.OffsetIndexTest > appendOutOfOrder PASSED kafka.log.OffsetIndexTest > truncate PASSED :test_core_2_11 Building project 'core' with Scala version 2.11.8 :kafka-trunk-jdk8:clients:compileJava UP-TO-DATE :kafka-trunk-jdk8:clients:processResources UP-TO-DATE :kafka-trunk-jdk8:clients:classes UP-TO-DATE :kafka-trunk-jdk8:clients:determineCommitId UP-TO-DATE :kafka-trunk-jdk8:clients:createVersionFile :kafka-trunk-jdk8:clients:jar UP-TO-DATE :kafka-trunk-jdk8:clients:compileTestJava UP-TO-DATE :kafka-trunk-jdk8:clients:processTestResources UP-TO-DATE :kafka-trunk-jdk8:clients:testClasses UP-TO-DATE :kafka-trunk-jdk8:core:compileJava UP-TO-DATE :kafka-trunk-jdk8:core:compileScalaJava HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=512m; support was removed in 8.0 <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/api/OffsetCommitRequest.scala>:79: value DEFAULT_TIMESTAMP in object OffsetCommitRequest is deprecated: see corresponding Javadoc for more information. org.apache.kafka.common.requests.OffsetCommitRequest.DEFAULT_TIMESTAMP ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/common/OffsetMetadataAndError.scala>:36: value DEFAULT_TIMESTAMP in object OffsetCommitRequest is deprecated: see corresponding Javadoc for more information. commitTimestamp: Long = org.apache.kafka.common.requests.OffsetCommitRequest.DEFAULT_TIMESTAMP, ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/common/OffsetMetadataAndError.scala>:37: value DEFAULT_TIMESTAMP in object OffsetCommitRequest is deprecated: see corresponding Javadoc for more information. expireTimestamp: Long = org.apache.kafka.common.requests.OffsetCommitRequest.DEFAULT_TIMESTAMP) { ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/coordinator/GroupMetadataManager.scala>:401: value DEFAULT_TIMESTAMP in object OffsetCommitRequest is deprecated: see corresponding Javadoc for more information. if (value.expireTimestamp == org.apache.kafka.common.requests.OffsetCommitRequest.DEFAULT_TIMESTAMP) ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/log/OffsetIndex.scala>:56: no valid targets for annotation on variable _file - it is discarded unused. You may specify targets with meta-annotations, e.g. @(volatile @param) class OffsetIndex(@volatile private[this] var _file: File, val baseOffset: Long, val maxIndexSize: Int = -1) extends Logging { ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/server/KafkaApis.scala>:301: value DEFAULT_TIMESTAMP in object OffsetCommitRequest is deprecated: see corresponding Javadoc for more information. if (partitionData.timestamp == OffsetCommitRequest.DEFAULT_TIMESTAMP) ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/admin/AclCommand.scala>:246: method readLine in class DeprecatedConsole is deprecated: Use the method in scala.io.StdIn Console.readLine().equalsIgnoreCase("y") ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/admin/TopicCommand.scala>:378: method readLine in class DeprecatedConsole is deprecated: Use the method in scala.io.StdIn if (!Console.readLine().equalsIgnoreCase("y")) { ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/client/ClientUtils.scala>:93: class ProducerConfig in package producer is deprecated: This class has been deprecated and will be removed in a future release. Please use org.apache.kafka.clients.producer.ProducerConfig instead. val producerConfig = new ProducerConfig(props) ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/client/ClientUtils.scala>:94: method fetchTopicMetadata in object ClientUtils is deprecated: This method has been deprecated and will be removed in a future release. fetchTopicMetadata(topics, brokers, producerConfig, correlationId) ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/controller/ControllerChannelManager.scala>:396: constructor UpdateMetadataRequest in class UpdateMetadataRequest is deprecated: see corresponding Javadoc for more information. new UpdateMetadataRequest(controllerId, controllerEpoch, liveBrokers.asJava, partitionStates.asJava) ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/metrics/KafkaMetricsGroup.scala>:191: object ProducerRequestStatsRegistry in package producer is deprecated: This object has been deprecated and will be removed in a future release. ProducerRequestStatsRegistry.removeProducerRequestStats(clientId) ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/network/BlockingChannel.scala>:129: method readFromReadableChannel in class NetworkReceive is deprecated: see corresponding Javadoc for more information. response.readFromReadableChannel(channel) ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/server/KafkaApis.scala>:301: value timestamp in class PartitionData is deprecated: see corresponding Javadoc for more information. if (partitionData.timestamp == OffsetCommitRequest.DEFAULT_TIMESTAMP) ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/server/KafkaApis.scala>:304: value timestamp in class PartitionData is deprecated: see corresponding Javadoc for more information. offsetRetention + partitionData.timestamp ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/tools/ConsoleProducer.scala>:43: class OldProducer in package producer is deprecated: This class has been deprecated and will be removed in a future release. Please use org.apache.kafka.clients.producer.KafkaProducer instead. new OldProducer(getOldProducerProps(config)) ^ <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/main/scala/kafka/tools/ConsoleProducer.scala>:45: class NewShinyProducer in package producer is deprecated: This class has been deprecated and will be removed in a future release. Please use org.apache.kafka.clients.producer.KafkaProducer instead. new NewShinyProducer(getNewProducerProps(config)) ^ 17 warnings found warning: [options] bootstrap class path not set in conjunction with -source 1.7 1 warning :kafka-trunk-jdk8:core:processResources UP-TO-DATE :kafka-trunk-jdk8:core:classes :kafka-trunk-jdk8:core:checkstyleMain :kafka-trunk-jdk8:core:compileTestJava UP-TO-DATE :kafka-trunk-jdk8:core:compileTestScala <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/src/test/scala/unit/kafka/metrics/MetricsTest.scala>:86: method createAndShutdownStep in class MetricsTest is deprecated: This test has been deprecated and it will be removed in a future release createAndShutdownStep("group0", "consumer0", "producer0") ^ one warning found :kafka-trunk-jdk8:core:processTestResources UP-TO-DATE :kafka-trunk-jdk8:core:testClasses :kafka-trunk-jdk8:core:checkstyleTest UP-TO-DATE :kafka-trunk-jdk8:core:test unit.kafka.admin.ReassignPartitionsClusterTest > shouldExpandCluster PASSED unit.kafka.admin.ReassignPartitionsClusterTest > shouldMoveSinglePartition PASSED unit.kafka.admin.ReassignPartitionsClusterTest > shouldShrinkCluster PASSED unit.kafka.server.ApiVersionsTest > testApiVersions PASSED kafka.tools.ConsoleProducerTest > testParseKeyProp PASSED kafka.tools.ConsoleProducerTest > testValidConfigsOldProducer PASSED kafka.tools.ConsoleProducerTest > testInvalidConfigs PASSED kafka.tools.ConsoleProducerTest > testValidConfigsNewProducer PASSED kafka.tools.ConsoleConsumerTest > shouldLimitReadsToMaxMessageLimit PASSED kafka.tools.ConsoleConsumerTest > shouldParseValidNewConsumerValidConfig PASSED kafka.tools.ConsoleConsumerTest > shouldParseConfigsFromFile PASSED kafka.tools.ConsoleConsumerTest > shouldParseValidOldConsumerValidConfig PASSED kafka.cluster.BrokerEndPointTest > testEndpointFromUri PASSED kafka.cluster.BrokerEndPointTest > testHashAndEquals PASSED kafka.cluster.BrokerEndPointTest > testFromJsonFutureVersion PASSED kafka.cluster.BrokerEndPointTest > testBrokerEndpointFromUri PASSED kafka.cluster.BrokerEndPointTest > testFromJsonV1 PASSED kafka.cluster.BrokerEndPointTest > testFromJsonV2 PASSED kafka.integration.UncleanLeaderElectionTest > testUncleanLeaderElectionEnabled PASSED kafka.integration.UncleanLeaderElectionTest > testCleanLeaderElectionDisabledByTopicOverride PASSED kafka.integration.UncleanLeaderElectionTest > testUncleanLeaderElectionDisabled PASSED kafka.integration.UncleanLeaderElectionTest > testUncleanLeaderElectionInvalidTopicOverride PASSED kafka.integration.UncleanLeaderElectionTest > testUncleanLeaderElectionEnabledByTopicOverride PASSED kafka.integration.AutoOffsetResetTest > testResetToLatestWhenOffsetTooLow PASSED kafka.integration.AutoOffsetResetTest > testResetToEarliestWhenOffsetTooLow PASSED kafka.integration.AutoOffsetResetTest > testResetToEarliestWhenOffsetTooHigh PASSED kafka.integration.AutoOffsetResetTest > testResetToLatestWhenOffsetTooHigh PASSED kafka.integration.MinIsrConfigTest > testDefaultKafkaConfig PASSED kafka.integration.FetcherTest > testFetcher PASSED kafka.integration.SslTopicMetadataTest > testIsrAfterBrokerShutDownAndJoinsBack PASSED kafka.integration.SslTopicMetadataTest > testAutoCreateTopicWithCollision PASSED kafka.integration.SslTopicMetadataTest > testAliveBrokerListWithNoTopics PASSED kafka.integration.SslTopicMetadataTest > testAutoCreateTopic PASSED kafka.integration.SslTopicMetadataTest > testGetAllTopicMetadata PASSED kafka.integration.SslTopicMetadataTest > testAliveBrokersListWithNoTopicsAfterNewBrokerStartup PASSED kafka.integration.SslTopicMetadataTest > testBasicTopicMetadata PASSED kafka.integration.SslTopicMetadataTest > testAliveBrokersListWithNoTopicsAfterABrokerShutdown PASSED kafka.integration.PlaintextTopicMetadataTest > testIsrAfterBrokerShutDownAndJoinsBack PASSED kafka.integration.PlaintextTopicMetadataTest > testAutoCreateTopicWithCollision PASSED kafka.integration.PlaintextTopicMetadataTest > testAliveBrokerListWithNoTopics PASSED kafka.integration.PlaintextTopicMetadataTest > testAutoCreateTopic PASSED kafka.integration.PlaintextTopicMetadataTest > testGetAllTopicMetadata PASSED kafka.integration.PlaintextTopicMetadataTest > testAliveBrokersListWithNoTopicsAfterNewBrokerStartup PASSED kafka.integration.PlaintextTopicMetadataTest > testBasicTopicMetadata PASSED kafka.integration.PlaintextTopicMetadataTest > testAliveBrokersListWithNoTopicsAfterABrokerShutdown PASSED kafka.integration.PrimitiveApiTest > testMultiProduce PASSED kafka.integration.PrimitiveApiTest > testDefaultEncoderProducerAndFetch PASSED kafka.integration.PrimitiveApiTest > testFetchRequestCanProperlySerialize PASSED kafka.integration.PrimitiveApiTest > testPipelinedProduceRequests PASSED kafka.integration.PrimitiveApiTest > testProduceAndMultiFetch PASSED kafka.integration.PrimitiveApiTest > testDefaultEncoderProducerAndFetchWithCompression PASSED kafka.integration.PrimitiveApiTest > testConsumerEmptyTopic PASSED kafka.integration.PrimitiveApiTest > testEmptyFetchRequest PASSED # Java HotSpot(TM) 64-Bit Server VM warning: INFO: os::commit_memory(0x00000000c7d00000, 34078720, 0) failed; error='Cannot allocate memory' (errno=12) # There is insufficient memory for the Java Runtime Environment to continue. # Native memory allocation (mmap) failed to map 34078720 bytes for committing reserved memory. # An error report file with more information is saved as: # <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/hs_err_pid4921.log> :kafka-trunk-jdk8:core:test FAILED :test_core_2_11 FAILED FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':core:test'. > Process 'Gradle Test Executor 2' finished with non-zero exit value 1 * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. BUILD FAILED Total time: 34 mins 43.07 secs Build step 'Execute shell' marked build as failure Recording test results Setting GRADLE_2_4_RC_2_HOME=/home/jenkins/jenkins-slave/tools/hudson.plugins.gradle.GradleInstallation/Gradle_2.4-rc-2 Setting JDK1_8_0_66_HOME=/home/jenkins/jenkins-slave/tools/hudson.model.JDK/jdk1.8.0_66 Setting GRADLE_2_4_RC_2_HOME=/home/jenkins/jenkins-slave/tools/hudson.plugins.gradle.GradleInstallation/Gradle_2.4-rc-2 Setting JDK1_8_0_66_HOME=/home/jenkins/jenkins-slave/tools/hudson.model.JDK/jdk1.8.0_66