From af719f880312bf02bae967de7414ffc97ebe454e Mon Sep 17 00:00:00 2001 From: m1a2st Date: Wed, 20 Nov 2024 23:00:33 +0800 Subject: [PATCH 01/10] wip --- checkstyle/checkstyle.xml | 4 +++- gradle/dependencies.gradle | 2 +- gradle/resources/dependencycheck-suppressions.xml | 11 ----------- 3 files changed, 4 insertions(+), 13 deletions(-) diff --git a/checkstyle/checkstyle.xml b/checkstyle/checkstyle.xml index 6b8aedb6e4700..7fdd583715c8e 100644 --- a/checkstyle/checkstyle.xml +++ b/checkstyle/checkstyle.xml @@ -39,7 +39,9 @@ - + + + diff --git a/gradle/dependencies.gradle b/gradle/dependencies.gradle index 27c69c90037a5..1b33e893e12ee 100644 --- a/gradle/dependencies.gradle +++ b/gradle/dependencies.gradle @@ -58,7 +58,7 @@ versions += [ // when updating checkstyle, check whether the exclusion of // CVE-2023-2976 and CVE-2020-8908 can be dropped from // gradle/resources/dependencycheck-suppressions.xml - checkstyle: "8.36.2", + checkstyle: "10.20.1", commonsCli: "1.4", commonsIo: "2.14.0", // ZooKeeper dependency. Do not use, this is going away. commonsValidator: "1.9.0", diff --git a/gradle/resources/dependencycheck-suppressions.xml b/gradle/resources/dependencycheck-suppressions.xml index 5ce34df1d2ef4..b28adc9002875 100644 --- a/gradle/resources/dependencycheck-suppressions.xml +++ b/gradle/resources/dependencycheck-suppressions.xml @@ -23,17 +23,6 @@ ]]> CVE-2023-35116 - - - CVE-2020-8908 - CVE-2023-2976 - Date: Tue, 26 Nov 2024 09:50:06 +0800 Subject: [PATCH 02/10] wip --- checkstyle/checkstyle.xml | 2 +- .../kafka/clients/admin/KafkaAdminClient.java | 6 +++--- .../common/requests/DescribeConfigsResponse.java | 1 - .../clients/admin/KafkaAdminClientTest.java | 12 ++++++------ .../ListConsumerGroupOffsetsHandlerTest.java | 16 ++++++++-------- .../consumer/internals/ConsumerProtocolTest.java | 4 ++-- .../internals/TransactionManagerTest.java | 8 ++++---- .../ConfigurationControlManagerTest.java | 8 ++++---- .../kafka/controller/QuorumControllerTest.java | 8 ++++---- .../ReplicationControlManagerTest.java | 8 ++++---- .../kafka/server/AssignmentsManagerTest.java | 10 +++++----- 11 files changed, 41 insertions(+), 42 deletions(-) diff --git a/checkstyle/checkstyle.xml b/checkstyle/checkstyle.xml index 7fdd583715c8e..fd3042376713d 100644 --- a/checkstyle/checkstyle.xml +++ b/checkstyle/checkstyle.xml @@ -40,7 +40,7 @@ - + diff --git a/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java b/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java index bfa15077c8ef7..e80ea7073228f 100644 --- a/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java +++ b/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java @@ -2896,8 +2896,8 @@ public AlterConfigsResult alterConfigs(Map configs, fina } else unifiedRequestResources.add(resource); } - if (!unifiedRequestResources.isEmpty()) - allFutures.putAll(alterConfigs(configs, options, unifiedRequestResources, new LeastLoadedBrokerOrActiveKController())); + if (!unifiedRequestResources.isEmpty()) + allFutures.putAll(alterConfigs(configs, options, unifiedRequestResources, new LeastLoadedBrokerOrActiveKController())); return new AlterConfigsResult(new HashMap<>(allFutures)); } @@ -5030,7 +5030,7 @@ AddRaftVoterRequest.Builder createRequest(int timeoutMs) { setHost(endpoint.host()). setPort(endpoint.port()))); return new AddRaftVoterRequest.Builder( - new AddRaftVoterRequestData(). + new AddRaftVoterRequestData(). setClusterId(options.clusterId().orElse(null)). setTimeoutMs(timeoutMs). setVoterId(voterId) . diff --git a/clients/src/main/java/org/apache/kafka/common/requests/DescribeConfigsResponse.java b/clients/src/main/java/org/apache/kafka/common/requests/DescribeConfigsResponse.java index cb128b8b42cf1..f352ba4eea9fe 100644 --- a/clients/src/main/java/org/apache/kafka/common/requests/DescribeConfigsResponse.java +++ b/clients/src/main/java/org/apache/kafka/common/requests/DescribeConfigsResponse.java @@ -17,7 +17,6 @@ package org.apache.kafka.common.requests; -import org.apache.kafka.clients.admin.ConfigEntry; import org.apache.kafka.common.config.ConfigResource; import org.apache.kafka.common.message.DescribeConfigsResponseData; import org.apache.kafka.common.protocol.ApiKeys; diff --git a/clients/src/test/java/org/apache/kafka/clients/admin/KafkaAdminClientTest.java b/clients/src/test/java/org/apache/kafka/clients/admin/KafkaAdminClientTest.java index a836e98214002..3f4f98c2a25ba 100644 --- a/clients/src/test/java/org/apache/kafka/clients/admin/KafkaAdminClientTest.java +++ b/clients/src/test/java/org/apache/kafka/clients/admin/KafkaAdminClientTest.java @@ -2046,9 +2046,9 @@ public void testDescribeBrokerAndLogConfigs() throws Exception { new DescribeConfigsResponseData().setResults(asList(new DescribeConfigsResponseData.DescribeConfigsResult() .setResourceName(brokerResource.name()).setResourceType(brokerResource.type().id()).setErrorCode(Errors.NONE.code()) .setConfigs(emptyList()), - new DescribeConfigsResponseData.DescribeConfigsResult() - .setResourceName(brokerLoggerResource.name()).setResourceType(brokerLoggerResource.type().id()).setErrorCode(Errors.NONE.code()) - .setConfigs(emptyList())))), env.cluster().nodeById(0)); + new DescribeConfigsResponseData.DescribeConfigsResult() + .setResourceName(brokerLoggerResource.name()).setResourceType(brokerLoggerResource.type().id()).setErrorCode(Errors.NONE.code()) + .setConfigs(emptyList())))), env.cluster().nodeById(0)); Map> result = env.adminClient().describeConfigs(asList( brokerResource, brokerLoggerResource)).values(); @@ -2087,9 +2087,9 @@ public void testDescribeConfigsUnrequested() throws Exception { new DescribeConfigsResponseData().setResults(asList(new DescribeConfigsResponseData.DescribeConfigsResult() .setResourceName(topic.name()).setResourceType(topic.type().id()).setErrorCode(Errors.NONE.code()) .setConfigs(emptyList()), - new DescribeConfigsResponseData.DescribeConfigsResult() - .setResourceName(unrequested.name()).setResourceType(unrequested.type().id()).setErrorCode(Errors.NONE.code()) - .setConfigs(emptyList()))))); + new DescribeConfigsResponseData.DescribeConfigsResult() + .setResourceName(unrequested.name()).setResourceType(unrequested.type().id()).setErrorCode(Errors.NONE.code()) + .setConfigs(emptyList()))))); Map> result = env.adminClient().describeConfigs(singletonList( topic)).values(); assertEquals(new HashSet<>(singletonList(topic)), result.keySet()); diff --git a/clients/src/test/java/org/apache/kafka/clients/admin/internals/ListConsumerGroupOffsetsHandlerTest.java b/clients/src/test/java/org/apache/kafka/clients/admin/internals/ListConsumerGroupOffsetsHandlerTest.java index d9935fbdca6e8..868603dd6e033 100644 --- a/clients/src/test/java/org/apache/kafka/clients/admin/internals/ListConsumerGroupOffsetsHandlerTest.java +++ b/clients/src/test/java/org/apache/kafka/clients/admin/internals/ListConsumerGroupOffsetsHandlerTest.java @@ -177,10 +177,10 @@ public void testSuccessfulHandleResponseWithOnePartitionErrorWithMultipleGroups( Map offsetAndMetadataMapTwo = Collections.singletonMap(t2p2, new OffsetAndMetadata(10L)); Map> expectedResult = - new HashMap>() {{ - put(groupZero, offsetAndMetadataMapZero); - put(groupOne, offsetAndMetadataMapOne); - put(groupTwo, offsetAndMetadataMapTwo); + new HashMap<>() {{ + put(groupZero, offsetAndMetadataMapZero); + put(groupOne, offsetAndMetadataMapOne); + put(groupTwo, offsetAndMetadataMapTwo); }}; assertCompletedForMultipleGroups( @@ -304,10 +304,10 @@ private OffsetFetchResponse buildResponseWithPartitionErrorWithMultipleGroups(Er responseDataTwo.put(t2p2, new OffsetFetchResponse.PartitionData(10, Optional.empty(), "", Errors.NONE)); Map> responseData = - new HashMap>() {{ - put(groupZero, responseDataZero); - put(groupOne, responseDataOne); - put(groupTwo, responseDataTwo); + new HashMap<>() {{ + put(groupZero, responseDataZero); + put(groupOne, responseDataOne); + put(groupTwo, responseDataTwo); }}; Map errorMap = errorMap(groups, Errors.NONE); diff --git a/clients/src/test/java/org/apache/kafka/clients/consumer/internals/ConsumerProtocolTest.java b/clients/src/test/java/org/apache/kafka/clients/consumer/internals/ConsumerProtocolTest.java index 07808f29806a5..b73576757e86e 100644 --- a/clients/src/test/java/org/apache/kafka/clients/consumer/internals/ConsumerProtocolTest.java +++ b/clients/src/test/java/org/apache/kafka/clients/consumer/internals/ConsumerProtocolTest.java @@ -298,8 +298,8 @@ public void serializeDeserializeConsumerProtocolSubscriptionAllVersions() { if (version >= 1) { assertEquals( Set.of( - new ConsumerProtocolSubscription.TopicPartition().setTopic("foo").setPartitions(Collections.singletonList(0)), - new ConsumerProtocolSubscription.TopicPartition().setTopic("bar").setPartitions(Collections.singletonList(0) + new ConsumerProtocolSubscription.TopicPartition().setTopic("foo").setPartitions(Collections.singletonList(0)), + new ConsumerProtocolSubscription.TopicPartition().setTopic("bar").setPartitions(Collections.singletonList(0) )), Set.copyOf(parsedSubscription.ownedPartitions()) ); diff --git a/clients/src/test/java/org/apache/kafka/clients/producer/internals/TransactionManagerTest.java b/clients/src/test/java/org/apache/kafka/clients/producer/internals/TransactionManagerTest.java index 02570b083ec1e..08bafd729a020 100644 --- a/clients/src/test/java/org/apache/kafka/clients/producer/internals/TransactionManagerTest.java +++ b/clients/src/test/java/org/apache/kafka/clients/producer/internals/TransactionManagerTest.java @@ -1800,10 +1800,10 @@ public void testMultipleAddPartitionsPerForOneProduce() throws InterruptedExcept @ParameterizedTest @EnumSource(names = { - "UNKNOWN_TOPIC_OR_PARTITION", - "REQUEST_TIMED_OUT", - "COORDINATOR_LOAD_IN_PROGRESS", - "CONCURRENT_TRANSACTIONS" + "UNKNOWN_TOPIC_OR_PARTITION", + "REQUEST_TIMED_OUT", + "COORDINATOR_LOAD_IN_PROGRESS", + "CONCURRENT_TRANSACTIONS" }) public void testRetriableErrors(Errors error) { // Ensure FindCoordinator retries. diff --git a/metadata/src/test/java/org/apache/kafka/controller/ConfigurationControlManagerTest.java b/metadata/src/test/java/org/apache/kafka/controller/ConfigurationControlManagerTest.java index 2ce417f63aa90..223913d20de98 100644 --- a/metadata/src/test/java/org/apache/kafka/controller/ConfigurationControlManagerTest.java +++ b/metadata/src/test/java/org/apache/kafka/controller/ConfigurationControlManagerTest.java @@ -316,10 +316,10 @@ public void testIncrementalAlterConfigsWithPolicy() { assertEquals(ControllerResult.atomicOf(asList(new ApiMessageAndVersion( new ConfigRecord().setResourceType(BROKER.id()).setResourceName("0"). setName("foo.bar").setValue("123"), CONFIG_RECORD.highestSupportedVersion()), new ApiMessageAndVersion( - new ConfigRecord().setResourceType(BROKER.id()).setResourceName("0"). - setName("quux").setValue("456"), CONFIG_RECORD.highestSupportedVersion()), new ApiMessageAndVersion( - new ConfigRecord().setResourceType(BROKER.id()).setResourceName("0"). - setName("broker.config.to.remove").setValue(null), CONFIG_RECORD.highestSupportedVersion()) + new ConfigRecord().setResourceType(BROKER.id()).setResourceName("0"). + setName("quux").setValue("456"), CONFIG_RECORD.highestSupportedVersion()), new ApiMessageAndVersion( + new ConfigRecord().setResourceType(BROKER.id()).setResourceName("0"). + setName("broker.config.to.remove").setValue(null), CONFIG_RECORD.highestSupportedVersion()) ), toMap(entry(MYTOPIC, new ApiError(Errors.POLICY_VIOLATION, "Expected: AlterConfigPolicy.RequestMetadata(resource=ConfigResource(" + diff --git a/metadata/src/test/java/org/apache/kafka/controller/QuorumControllerTest.java b/metadata/src/test/java/org/apache/kafka/controller/QuorumControllerTest.java index 867957513b4f6..874690ac7e99f 100644 --- a/metadata/src/test/java/org/apache/kafka/controller/QuorumControllerTest.java +++ b/metadata/src/test/java/org/apache/kafka/controller/QuorumControllerTest.java @@ -860,10 +860,10 @@ public void testSnapshotSaveAndLoad() throws Throwable { Arrays.asList(new CreatableReplicaAssignment(). setPartitionIndex(0). setBrokerIds(Arrays.asList(0, 1, 2)), - new CreatableReplicaAssignment(). - setPartitionIndex(1). - setBrokerIds(Arrays.asList(1, 2, 0))). - iterator()))).iterator())), + new CreatableReplicaAssignment(). + setPartitionIndex(1). + setBrokerIds(Arrays.asList(1, 2, 0))). + iterator()))).iterator())), Collections.singleton("foo")).get(); fooId = fooData.topics().find("foo").topicId(); active.allocateProducerIds(ANONYMOUS_CONTEXT, diff --git a/metadata/src/test/java/org/apache/kafka/controller/ReplicationControlManagerTest.java b/metadata/src/test/java/org/apache/kafka/controller/ReplicationControlManagerTest.java index c53f068e9d6e0..d5cc2f8af6304 100644 --- a/metadata/src/test/java/org/apache/kafka/controller/ReplicationControlManagerTest.java +++ b/metadata/src/test/java/org/apache/kafka/controller/ReplicationControlManagerTest.java @@ -1790,7 +1790,7 @@ public void testReassignPartitions(short version) { setReplicas(asList(0, 2, 1)), new ReassignablePartition().setPartitionIndex(2). setReplicas(asList(0, 2, 1)))), - new ReassignableTopic().setName("bar")))); + new ReassignableTopic().setName("bar")))); assertEquals(new AlterPartitionReassignmentsResponseData(). setErrorMessage(null).setResponses(asList( new ReassignableTopicResponse().setName("foo").setPartitions(asList( @@ -2141,12 +2141,12 @@ public void testCancelReassignPartitions() { setReplicas(asList(5, 6, 7)), new ReassignablePartition().setPartitionIndex(3). setReplicas(Collections.emptyList()))), - new ReassignableTopic().setName("bar").setPartitions(singletonList( + new ReassignableTopic().setName("bar").setPartitions(singletonList( new ReassignablePartition().setPartitionIndex(0). setReplicas(asList(1, 2, 3, 4, 0))))))); assertEquals(new AlterPartitionReassignmentsResponseData(). setErrorMessage(null).setResponses(asList( - new ReassignableTopicResponse().setName("foo").setPartitions(asList( + new ReassignableTopicResponse().setName("foo").setPartitions(asList( new ReassignablePartitionResponse().setPartitionIndex(0). setErrorMessage(null), new ReassignablePartitionResponse().setPartitionIndex(1). @@ -2159,7 +2159,7 @@ public void testCancelReassignPartitions() { setErrorCode(INVALID_REPLICA_ASSIGNMENT.code()). setErrorMessage("The manual partition assignment includes an empty " + "replica list."))), - new ReassignableTopicResponse().setName("bar").setPartitions(singletonList( + new ReassignableTopicResponse().setName("bar").setPartitions(singletonList( new ReassignablePartitionResponse().setPartitionIndex(0). setErrorMessage(null))))), alterResult.response()); diff --git a/server/src/test/java/org/apache/kafka/server/AssignmentsManagerTest.java b/server/src/test/java/org/apache/kafka/server/AssignmentsManagerTest.java index 543e67d03e21b..f426c355ceb50 100644 --- a/server/src/test/java/org/apache/kafka/server/AssignmentsManagerTest.java +++ b/server/src/test/java/org/apache/kafka/server/AssignmentsManagerTest.java @@ -493,11 +493,11 @@ void testBuildRequestData() { setPartitions(Collections.singletonList( new AssignReplicasToDirsRequestData.PartitionData(). setPartitionIndex(2))), - new AssignReplicasToDirsRequestData.TopicData(). - setTopicId(TOPIC_2). - setPartitions(Collections.singletonList( - new AssignReplicasToDirsRequestData.PartitionData(). - setPartitionIndex(5))))), + new AssignReplicasToDirsRequestData.TopicData(). + setTopicId(TOPIC_2). + setPartitions(Collections.singletonList( + new AssignReplicasToDirsRequestData.PartitionData(). + setPartitionIndex(5))))), new AssignReplicasToDirsRequestData.DirectoryData(). setId(DIR_3). setTopics(Collections.singletonList( From cbbbe39a700577ed8c3f8a9eb50f29340bfc50d5 Mon Sep 17 00:00:00 2001 From: m1a2st Date: Sun, 1 Dec 2024 12:34:09 +0800 Subject: [PATCH 03/10] pass all checkstyle rule --- checkstyle/checkstyle.xml | 2 +- .../internals/OffsetsForLeaderEpochUtils.java | 2 +- .../kafka/server/authorizer/Authorizer.java | 12 +++---- .../ListConsumerGroupOffsetsHandlerTest.java | 4 +-- .../kafka/common/network/SelectorTest.java | 4 +-- .../requests/OffsetFetchResponseTest.java | 20 ++++++------ .../common/requests/RequestResponseTest.java | 6 ++-- .../mirror/MirrorSourceConnectorTest.java | 4 +-- .../connect/runtime/AbstractHerderTest.java | 4 +-- .../transforms/predicates/HasHeaderKey.java | 5 ++- .../kafka/admin/AdminFenceProducersTest.java | 10 +++--- .../java/kafka/admin/ClientTelemetryTest.java | 2 +- .../java/kafka/admin/ConfigCommandTest.java | 4 +-- .../UserScramCredentialsCommandTest.java | 4 +-- .../BootstrapControllersIntegrationTest.java | 8 ++--- .../ConfigurationControlManagerTest.java | 8 ++--- .../ReplicationControlManagerTest.java | 28 +++++++--------- .../storage/internals/log/LogSegment.java | 4 +-- .../internals/log/LogValidatorTest.java | 26 +++++++-------- .../examples/pageview/PageViewTypedDemo.java | 12 +++---- ...ailabilityTaskAssignorIntegrationTest.java | 14 ++++---- .../QueryableStateIntegrationTest.java | 9 +++--- .../streams/integration/StoreQuerySuite.java | 16 +++++----- .../internals/StreamStreamJoinUtil.java | 2 +- .../internals/InternalTopologyBuilder.java | 1 - .../processor/internals/TaskSuite.java | 14 ++++---- .../state/internals/RocksDBStoreTest.java | 6 ++-- .../apache/kafka/streams/utils/TestUtils.java | 1 + .../test/api/ClusterTestExtensionsTest.java | 4 +-- .../org/apache/kafka/tools/ToolsUtils.java | 2 +- .../tools/BrokerApiVersionsCommandTest.java | 2 +- .../tools/MetadataQuorumCommandTest.java | 8 ++--- .../ReassignPartitionsCommandTest.java | 32 +++++++++---------- .../apache/kafka/trogdor/fault/Kibosh.java | 4 +-- .../apache/kafka/trogdor/rest/TaskState.java | 10 +++--- .../trogdor/workload/RecordProcessor.java | 2 +- 36 files changed, 146 insertions(+), 150 deletions(-) diff --git a/checkstyle/checkstyle.xml b/checkstyle/checkstyle.xml index fd3042376713d..fa1d5873a2cbc 100644 --- a/checkstyle/checkstyle.xml +++ b/checkstyle/checkstyle.xml @@ -40,7 +40,7 @@ - + diff --git a/clients/src/main/java/org/apache/kafka/clients/consumer/internals/OffsetsForLeaderEpochUtils.java b/clients/src/main/java/org/apache/kafka/clients/consumer/internals/OffsetsForLeaderEpochUtils.java index 033daa098761d..ae4b9f71f6c4b 100644 --- a/clients/src/main/java/org/apache/kafka/clients/consumer/internals/OffsetsForLeaderEpochUtils.java +++ b/clients/src/main/java/org/apache/kafka/clients/consumer/internals/OffsetsForLeaderEpochUtils.java @@ -44,7 +44,7 @@ public final class OffsetsForLeaderEpochUtils { private static final Logger LOG = LoggerFactory.getLogger(OffsetsForLeaderEpochUtils.class); - private OffsetsForLeaderEpochUtils(){} + private OffsetsForLeaderEpochUtils() {} static AbstractRequest.Builder prepareRequest( Map requestData) { diff --git a/clients/src/main/java/org/apache/kafka/server/authorizer/Authorizer.java b/clients/src/main/java/org/apache/kafka/server/authorizer/Authorizer.java index ca2a76b4716a9..58241c51ea085 100644 --- a/clients/src/main/java/org/apache/kafka/server/authorizer/Authorizer.java +++ b/clients/src/main/java/org/apache/kafka/server/authorizer/Authorizer.java @@ -202,14 +202,14 @@ op, new ResourcePattern(resourceType, "hardcode", PatternType.LITERAL), EnumMap> denyPatterns = new EnumMap<>(PatternType.class) {{ - put(PatternType.LITERAL, new HashSet<>()); - put(PatternType.PREFIXED, new HashSet<>()); - }}; + put(PatternType.LITERAL, new HashSet<>()); + put(PatternType.PREFIXED, new HashSet<>()); + }}; EnumMap> allowPatterns = new EnumMap<>(PatternType.class) {{ - put(PatternType.LITERAL, new HashSet<>()); - put(PatternType.PREFIXED, new HashSet<>()); - }}; + put(PatternType.LITERAL, new HashSet<>()); + put(PatternType.PREFIXED, new HashSet<>()); + }}; boolean hasWildCardAllow = false; diff --git a/clients/src/test/java/org/apache/kafka/clients/admin/internals/ListConsumerGroupOffsetsHandlerTest.java b/clients/src/test/java/org/apache/kafka/clients/admin/internals/ListConsumerGroupOffsetsHandlerTest.java index 868603dd6e033..e3bb56347a8ae 100644 --- a/clients/src/test/java/org/apache/kafka/clients/admin/internals/ListConsumerGroupOffsetsHandlerTest.java +++ b/clients/src/test/java/org/apache/kafka/clients/admin/internals/ListConsumerGroupOffsetsHandlerTest.java @@ -181,7 +181,7 @@ public void testSuccessfulHandleResponseWithOnePartitionErrorWithMultipleGroups( put(groupZero, offsetAndMetadataMapZero); put(groupOne, offsetAndMetadataMapOne); put(groupTwo, offsetAndMetadataMapTwo); - }}; + }}; assertCompletedForMultipleGroups( handleWithPartitionErrorMultipleGroups(Errors.UNKNOWN_TOPIC_OR_PARTITION), expectedResult); @@ -308,7 +308,7 @@ private OffsetFetchResponse buildResponseWithPartitionErrorWithMultipleGroups(Er put(groupZero, responseDataZero); put(groupOne, responseDataOne); put(groupTwo, responseDataTwo); - }}; + }}; Map errorMap = errorMap(groups, Errors.NONE); return new OffsetFetchResponse(0, errorMap, responseData); diff --git a/clients/src/test/java/org/apache/kafka/common/network/SelectorTest.java b/clients/src/test/java/org/apache/kafka/common/network/SelectorTest.java index b4c73d64d38c4..d955c7939ae8f 100644 --- a/clients/src/test/java/org/apache/kafka/common/network/SelectorTest.java +++ b/clients/src/test/java/org/apache/kafka/common/network/SelectorTest.java @@ -907,8 +907,8 @@ public void testLowestPriorityChannel() throws Exception { } assertNotNull(selector.lowestPriorityChannel()); for (int i = conns - 1; i >= 0; i--) { - if (i != 2) - assertEquals("", blockingRequest(String.valueOf(i), "")); + if (i != 2) + assertEquals("", blockingRequest(String.valueOf(i), "")); time.sleep(10); } assertEquals("2", selector.lowestPriorityChannel().id()); diff --git a/clients/src/test/java/org/apache/kafka/common/requests/OffsetFetchResponseTest.java b/clients/src/test/java/org/apache/kafka/common/requests/OffsetFetchResponseTest.java index c85d26dac5756..d0ef79b4479e8 100644 --- a/clients/src/test/java/org/apache/kafka/common/requests/OffsetFetchResponseTest.java +++ b/clients/src/test/java/org/apache/kafka/common/requests/OffsetFetchResponseTest.java @@ -390,16 +390,16 @@ public void testUseDefaultLeaderEpochV0ToV7() { .setErrorCode(Errors.NOT_COORDINATOR.code()) .setThrottleTimeMs(throttleTimeMs) .setTopics(Collections.singletonList( - new OffsetFetchResponseTopic() - .setName(topicOne) - .setPartitions(Collections.singletonList( - new OffsetFetchResponsePartition() - .setPartitionIndex(partitionOne) - .setCommittedOffset(offset) - .setCommittedLeaderEpoch(RecordBatch.NO_PARTITION_LEADER_EPOCH) - .setErrorCode(Errors.UNKNOWN_TOPIC_OR_PARTITION.code()) - .setMetadata(metadata)) - )) + new OffsetFetchResponseTopic() + .setName(topicOne) + .setPartitions(Collections.singletonList( + new OffsetFetchResponsePartition() + .setPartitionIndex(partitionOne) + .setCommittedOffset(offset) + .setCommittedLeaderEpoch(RecordBatch.NO_PARTITION_LEADER_EPOCH) + .setErrorCode(Errors.UNKNOWN_TOPIC_OR_PARTITION.code()) + .setMetadata(metadata)) + )) ); assertEquals(expectedData, response.data()); } diff --git a/clients/src/test/java/org/apache/kafka/common/requests/RequestResponseTest.java b/clients/src/test/java/org/apache/kafka/common/requests/RequestResponseTest.java index cb6a2458261ea..0d32fd442af88 100644 --- a/clients/src/test/java/org/apache/kafka/common/requests/RequestResponseTest.java +++ b/clients/src/test/java/org/apache/kafka/common/requests/RequestResponseTest.java @@ -2187,9 +2187,9 @@ private JoinGroupRequest createJoinGroupRequest(short version) { JoinGroupRequestData.JoinGroupRequestProtocolCollection protocols = new JoinGroupRequestData.JoinGroupRequestProtocolCollection( Collections.singleton( - new JoinGroupRequestData.JoinGroupRequestProtocol() - .setName("consumer-range") - .setMetadata(new byte[0])).iterator() + new JoinGroupRequestData.JoinGroupRequestProtocol() + .setName("consumer-range") + .setMetadata(new byte[0])).iterator() ); JoinGroupRequestData data = new JoinGroupRequestData() diff --git a/connect/mirror/src/test/java/org/apache/kafka/connect/mirror/MirrorSourceConnectorTest.java b/connect/mirror/src/test/java/org/apache/kafka/connect/mirror/MirrorSourceConnectorTest.java index a410adde94458..588baf8c09047 100644 --- a/connect/mirror/src/test/java/org/apache/kafka/connect/mirror/MirrorSourceConnectorTest.java +++ b/connect/mirror/src/test/java/org/apache/kafka/connect/mirror/MirrorSourceConnectorTest.java @@ -167,10 +167,10 @@ public void testAclFiltering() { new DefaultReplicationPolicy(), x -> true, getConfigPropertyFilter()); assertFalse(connector.shouldReplicateAcl( new AclBinding(new ResourcePattern(ResourceType.TOPIC, "test_topic", PatternType.LITERAL), - new AccessControlEntry("kafka", "", AclOperation.WRITE, AclPermissionType.ALLOW))), "should not replicate ALLOW WRITE"); + new AccessControlEntry("kafka", "", AclOperation.WRITE, AclPermissionType.ALLOW))), "should not replicate ALLOW WRITE"); assertTrue(connector.shouldReplicateAcl( new AclBinding(new ResourcePattern(ResourceType.TOPIC, "test_topic", PatternType.LITERAL), - new AccessControlEntry("kafka", "", AclOperation.ALL, AclPermissionType.ALLOW))), "should replicate ALLOW ALL"); + new AccessControlEntry("kafka", "", AclOperation.ALL, AclPermissionType.ALLOW))), "should replicate ALLOW ALL"); } @Test diff --git a/connect/runtime/src/test/java/org/apache/kafka/connect/runtime/AbstractHerderTest.java b/connect/runtime/src/test/java/org/apache/kafka/connect/runtime/AbstractHerderTest.java index 01f847d94f667..aa715667d24c4 100644 --- a/connect/runtime/src/test/java/org/apache/kafka/connect/runtime/AbstractHerderTest.java +++ b/connect/runtime/src/test/java/org/apache/kafka/connect/runtime/AbstractHerderTest.java @@ -1196,7 +1196,7 @@ protected void addConfigKey(Map keys, String name, keys.putAll(configDef.configKeys()); } - protected void addValue(List values, String name, String value, String...errors) { + protected void addValue(List values, String name, String value, String... errors) { values.add(new ConfigValue(name, value, new ArrayList<>(), Arrays.asList(errors))); } @@ -1211,7 +1211,7 @@ protected void assertNoInfoKey(ConfigInfos infos, String name) { assertNull(info.configKey()); } - protected void assertInfoValue(ConfigInfos infos, String name, String value, String...errors) { + protected void assertInfoValue(ConfigInfos infos, String name, String value, String... errors) { ConfigValueInfo info = findInfo(infos, name).configValue(); assertEquals(name, info.name()); assertEquals(value, info.value()); diff --git a/connect/transforms/src/main/java/org/apache/kafka/connect/transforms/predicates/HasHeaderKey.java b/connect/transforms/src/main/java/org/apache/kafka/connect/transforms/predicates/HasHeaderKey.java index fe0a99c4abe72..566bdfab23832 100644 --- a/connect/transforms/src/main/java/org/apache/kafka/connect/transforms/predicates/HasHeaderKey.java +++ b/connect/transforms/src/main/java/org/apache/kafka/connect/transforms/predicates/HasHeaderKey.java @@ -35,9 +35,8 @@ public class HasHeaderKey> implements Predicate, V private static final String NAME_CONFIG = "name"; public static final String OVERVIEW_DOC = "A predicate which is true for records with at least one header with the configured name."; public static final ConfigDef CONFIG_DEF = new ConfigDef() - .define(NAME_CONFIG, ConfigDef.Type.STRING, ConfigDef.NO_DEFAULT_VALUE, - new ConfigDef.NonEmptyString(), ConfigDef.Importance.MEDIUM, - "The header name."); + .define(NAME_CONFIG, ConfigDef.Type.STRING, ConfigDef.NO_DEFAULT_VALUE, + new ConfigDef.NonEmptyString(), ConfigDef.Importance.MEDIUM, "The header name."); private String name; @Override diff --git a/core/src/test/java/kafka/admin/AdminFenceProducersTest.java b/core/src/test/java/kafka/admin/AdminFenceProducersTest.java index 4b8178032e06f..d4b87ef8ed572 100644 --- a/core/src/test/java/kafka/admin/AdminFenceProducersTest.java +++ b/core/src/test/java/kafka/admin/AdminFenceProducersTest.java @@ -50,11 +50,11 @@ import static org.junit.jupiter.api.Assertions.assertTrue; @ClusterTestDefaults(serverProperties = { - @ClusterConfigProperty(key = ServerLogConfigs.AUTO_CREATE_TOPICS_ENABLE_CONFIG, value = "false"), - @ClusterConfigProperty(key = TransactionLogConfig.TRANSACTIONS_TOPIC_PARTITIONS_CONFIG, value = "1"), - @ClusterConfigProperty(key = TransactionLogConfig.TRANSACTIONS_TOPIC_REPLICATION_FACTOR_CONFIG, value = "1"), - @ClusterConfigProperty(key = TransactionLogConfig.TRANSACTIONS_TOPIC_MIN_ISR_CONFIG, value = "1"), - @ClusterConfigProperty(key = TransactionStateManagerConfig.TRANSACTIONS_ABORT_TIMED_OUT_TRANSACTION_CLEANUP_INTERVAL_MS_CONFIG, value = "2000") + @ClusterConfigProperty(key = ServerLogConfigs.AUTO_CREATE_TOPICS_ENABLE_CONFIG, value = "false"), + @ClusterConfigProperty(key = TransactionLogConfig.TRANSACTIONS_TOPIC_PARTITIONS_CONFIG, value = "1"), + @ClusterConfigProperty(key = TransactionLogConfig.TRANSACTIONS_TOPIC_REPLICATION_FACTOR_CONFIG, value = "1"), + @ClusterConfigProperty(key = TransactionLogConfig.TRANSACTIONS_TOPIC_MIN_ISR_CONFIG, value = "1"), + @ClusterConfigProperty(key = TransactionStateManagerConfig.TRANSACTIONS_ABORT_TIMED_OUT_TRANSACTION_CLEANUP_INTERVAL_MS_CONFIG, value = "2000") }) @ExtendWith(ClusterTestExtensions.class) public class AdminFenceProducersTest { diff --git a/core/src/test/java/kafka/admin/ClientTelemetryTest.java b/core/src/test/java/kafka/admin/ClientTelemetryTest.java index 56d466bd27b5c..f701d3f38450e 100644 --- a/core/src/test/java/kafka/admin/ClientTelemetryTest.java +++ b/core/src/test/java/kafka/admin/ClientTelemetryTest.java @@ -75,7 +75,7 @@ public class ClientTelemetryTest { types = Type.KRAFT, brokers = 3, serverProperties = { - @ClusterConfigProperty(key = METRIC_REPORTER_CLASSES_CONFIG, value = "kafka.admin.ClientTelemetryTest$GetIdClientTelemetry"), + @ClusterConfigProperty(key = METRIC_REPORTER_CLASSES_CONFIG, value = "kafka.admin.ClientTelemetryTest$GetIdClientTelemetry"), }) public void testClientInstanceId(ClusterInstance clusterInstance) throws InterruptedException, ExecutionException { Map configs = new HashMap<>(); diff --git a/core/src/test/java/kafka/admin/ConfigCommandTest.java b/core/src/test/java/kafka/admin/ConfigCommandTest.java index 74fe3271f23e2..fc08fae23e64f 100644 --- a/core/src/test/java/kafka/admin/ConfigCommandTest.java +++ b/core/src/test/java/kafka/admin/ConfigCommandTest.java @@ -420,7 +420,7 @@ public void testParseConfigsToBeAddedForAddConfigFile() throws IOException { assertEquals("[[1, 2], [3, 4]]", addedProps.getProperty("nested")); } - public void testExpectedEntityTypeNames(List expectedTypes, List expectedNames, List connectOpts, String...args) { + public void testExpectedEntityTypeNames(List expectedTypes, List expectedNames, List connectOpts, String... args) { ConfigCommand.ConfigCommandOptions createOpts = new ConfigCommand.ConfigCommandOptions(toArray(Arrays.asList(connectOpts.get(0), connectOpts.get(1), "--describe"), Arrays.asList(args))); createOpts.checkArgs(); assertEquals(createOpts.entityTypes().toSeq(), seq(expectedTypes)); @@ -1435,7 +1435,7 @@ public static List concat(List... lists) { } @SafeVarargs - public static Map concat(Map...maps) { + public static Map concat(Map... maps) { Map res = new HashMap<>(); Stream.of(maps) .map(Map::entrySet) diff --git a/core/src/test/java/kafka/admin/UserScramCredentialsCommandTest.java b/core/src/test/java/kafka/admin/UserScramCredentialsCommandTest.java index 55db18893f8f7..697dda07363a7 100644 --- a/core/src/test/java/kafka/admin/UserScramCredentialsCommandTest.java +++ b/core/src/test/java/kafka/admin/UserScramCredentialsCommandTest.java @@ -62,10 +62,10 @@ public ConfigCommandResult(String stdout, OptionalInt exitStatus) { } } - private ConfigCommandResult runConfigCommandViaBroker(String...args) { + private ConfigCommandResult runConfigCommandViaBroker(String... args) { AtomicReference exitStatus = new AtomicReference<>(OptionalInt.empty()); Exit.setExitProcedure((status, __) -> { - exitStatus.set(OptionalInt.of((Integer) status)); + exitStatus.set(OptionalInt.of(status)); throw new RuntimeException(); }); diff --git a/core/src/test/java/kafka/server/BootstrapControllersIntegrationTest.java b/core/src/test/java/kafka/server/BootstrapControllersIntegrationTest.java index 1462e1c7f64a7..75cd070b93d30 100644 --- a/core/src/test/java/kafka/server/BootstrapControllersIntegrationTest.java +++ b/core/src/test/java/kafka/server/BootstrapControllersIntegrationTest.java @@ -311,16 +311,16 @@ private static List> translatePartitionInfoToNodeIdList(List deserializer() { @SuppressWarnings("DefaultAnnotationParam") // being explicit for the example @JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "_t") @JsonSubTypes({ - @JsonSubTypes.Type(value = PageView.class, name = "pv"), - @JsonSubTypes.Type(value = UserProfile.class, name = "up"), - @JsonSubTypes.Type(value = PageViewByRegion.class, name = "pvbr"), - @JsonSubTypes.Type(value = WindowedPageViewByRegion.class, name = "wpvbr"), - @JsonSubTypes.Type(value = RegionCount.class, name = "rc") - }) + @JsonSubTypes.Type(value = PageView.class, name = "pv"), + @JsonSubTypes.Type(value = UserProfile.class, name = "up"), + @JsonSubTypes.Type(value = PageViewByRegion.class, name = "pvbr"), + @JsonSubTypes.Type(value = WindowedPageViewByRegion.class, name = "wpvbr"), + @JsonSubTypes.Type(value = RegionCount.class, name = "rc") + }) public interface JSONSerdeCompatible { } diff --git a/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/HighAvailabilityTaskAssignorIntegrationTest.java b/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/HighAvailabilityTaskAssignorIntegrationTest.java index df4d981fa7175..ce05eabb56606 100644 --- a/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/HighAvailabilityTaskAssignorIntegrationTest.java +++ b/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/HighAvailabilityTaskAssignorIntegrationTest.java @@ -97,9 +97,10 @@ public static void closeCluster() { @ParameterizedTest @ValueSource(strings = { - StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_NONE, - StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_MIN_TRAFFIC, - StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_BALANCE_SUBTOPOLOGY}) + StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_NONE, + StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_MIN_TRAFFIC, + StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_BALANCE_SUBTOPOLOGY + }) public void shouldScaleOutWithWarmupTasksAndInMemoryStores(final String rackAwareStrategy, final TestInfo testInfo) throws InterruptedException { // NB: this test takes at least a minute to run, because it needs a probing rebalance, and the minimum // value is one minute @@ -108,9 +109,10 @@ public void shouldScaleOutWithWarmupTasksAndInMemoryStores(final String rackAwar @ParameterizedTest @ValueSource(strings = { - StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_NONE, - StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_MIN_TRAFFIC, - StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_BALANCE_SUBTOPOLOGY}) + StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_NONE, + StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_MIN_TRAFFIC, + StreamsConfig.RACK_AWARE_ASSIGNMENT_STRATEGY_BALANCE_SUBTOPOLOGY + }) public void shouldScaleOutWithWarmupTasksAndPersistentStores(final String rackAwareStrategy, final TestInfo testInfo) throws InterruptedException { // NB: this test takes at least a minute to run, because it needs a probing rebalance, and the minimum // value is one minute diff --git a/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/QueryableStateIntegrationTest.java b/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/QueryableStateIntegrationTest.java index 6b67af943fcb3..ca3936633fc19 100644 --- a/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/QueryableStateIntegrationTest.java +++ b/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/QueryableStateIntegrationTest.java @@ -982,10 +982,11 @@ private void verifyCanQueryState(final int cacheSizeBytes) throws Exception { streamOne, batch1, TestUtils.producerConfig( - CLUSTER.bootstrapServers(), - StringSerializer.class, - StringSerializer.class, - new Properties()), + CLUSTER.bootstrapServers(), + StringSerializer.class, + StringSerializer.class, + new Properties() + ), mockTime); final KStream s1 = builder.stream(streamOne); diff --git a/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/StoreQuerySuite.java b/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/StoreQuerySuite.java index 08a163de59ef5..e6b212e7d1144 100644 --- a/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/StoreQuerySuite.java +++ b/streams/integration-tests/src/test/java/org/apache/kafka/streams/integration/StoreQuerySuite.java @@ -36,13 +36,13 @@ */ @Suite @SelectClasses({ - CompositeReadOnlyKeyValueStoreTest.class, - CompositeReadOnlyWindowStoreTest.class, - CompositeReadOnlySessionStoreTest.class, - GlobalStateStoreProviderTest.class, - StreamThreadStateStoreProviderTest.class, - WrappingStoreProviderTest.class, - QueryableStateIntegrationTest.class, - }) + CompositeReadOnlyKeyValueStoreTest.class, + CompositeReadOnlyWindowStoreTest.class, + CompositeReadOnlySessionStoreTest.class, + GlobalStateStoreProviderTest.class, + StreamThreadStateStoreProviderTest.class, + WrappingStoreProviderTest.class, + QueryableStateIntegrationTest.class, +}) public class StoreQuerySuite { } diff --git a/streams/src/main/java/org/apache/kafka/streams/kstream/internals/StreamStreamJoinUtil.java b/streams/src/main/java/org/apache/kafka/streams/kstream/internals/StreamStreamJoinUtil.java index ab23663bbc732..3e2c97f6e24a5 100644 --- a/streams/src/main/java/org/apache/kafka/streams/kstream/internals/StreamStreamJoinUtil.java +++ b/streams/src/main/java/org/apache/kafka/streams/kstream/internals/StreamStreamJoinUtil.java @@ -26,7 +26,7 @@ public final class StreamStreamJoinUtil { - private StreamStreamJoinUtil(){ + private StreamStreamJoinUtil() { } public static boolean skipRecord( diff --git a/streams/src/main/java/org/apache/kafka/streams/processor/internals/InternalTopologyBuilder.java b/streams/src/main/java/org/apache/kafka/streams/processor/internals/InternalTopologyBuilder.java index 9f65a415d9516..3dc46831b1eb6 100644 --- a/streams/src/main/java/org/apache/kafka/streams/processor/internals/InternalTopologyBuilder.java +++ b/streams/src/main/java/org/apache/kafka/streams/processor/internals/InternalTopologyBuilder.java @@ -24,7 +24,6 @@ import org.apache.kafka.streams.StreamsConfig; import org.apache.kafka.streams.Topology; import org.apache.kafka.streams.TopologyConfig; -import org.apache.kafka.streams.TopologyDescription; import org.apache.kafka.streams.errors.TopologyException; import org.apache.kafka.streams.internals.ApiUtils; import org.apache.kafka.streams.processor.StateStore; diff --git a/streams/src/test/java/org/apache/kafka/streams/processor/internals/TaskSuite.java b/streams/src/test/java/org/apache/kafka/streams/processor/internals/TaskSuite.java index 172321fab11e8..06a994a5aa8d3 100644 --- a/streams/src/test/java/org/apache/kafka/streams/processor/internals/TaskSuite.java +++ b/streams/src/test/java/org/apache/kafka/streams/processor/internals/TaskSuite.java @@ -31,13 +31,13 @@ */ @Suite @SelectClasses({ - StreamTaskTest.class, - StandbyTaskTest.class, - GlobalStateTaskTest.class, - TaskManagerTest.class, - TaskMetricsTest.class, - LegacyStickyTaskAssignorTest.class, - StreamsPartitionAssignorTest.class, + StreamTaskTest.class, + StandbyTaskTest.class, + GlobalStateTaskTest.class, + TaskManagerTest.class, + TaskMetricsTest.class, + LegacyStickyTaskAssignorTest.class, + StreamsPartitionAssignorTest.class, }) public class TaskSuite { } diff --git a/streams/src/test/java/org/apache/kafka/streams/state/internals/RocksDBStoreTest.java b/streams/src/test/java/org/apache/kafka/streams/state/internals/RocksDBStoreTest.java index 2c98c3427d0d3..f5dd717f5b568 100644 --- a/streams/src/test/java/org/apache/kafka/streams/state/internals/RocksDBStoreTest.java +++ b/streams/src/test/java/org/apache/kafka/streams/state/internals/RocksDBStoreTest.java @@ -235,7 +235,7 @@ public void shouldRemoveValueProvidersFromInjectedMetricsRecorderOnClose() { } public static class RocksDBConfigSetterWithUserProvidedStatistics implements RocksDBConfigSetter { - public RocksDBConfigSetterWithUserProvidedStatistics(){} + public RocksDBConfigSetterWithUserProvidedStatistics() {} public void setConfig(final String storeName, final Options options, final Map configs) { lastStatistics = new Statistics(); @@ -306,7 +306,7 @@ public void shouldCloseStatisticsWhenUserProvidesNoStatistics() throws Exception public static class RocksDBConfigSetterWithUserProvidedNewBlockBasedTableFormatConfig implements RocksDBConfigSetter { - public RocksDBConfigSetterWithUserProvidedNewBlockBasedTableFormatConfig(){} + public RocksDBConfigSetterWithUserProvidedNewBlockBasedTableFormatConfig() {} public void setConfig(final String storeName, final Options options, final Map configs) { options.setTableFormatConfig(new BlockBasedTableConfig()); @@ -335,7 +335,7 @@ public void shouldThrowWhenUserProvidesNewBlockBasedTableFormatConfig() { } public static class RocksDBConfigSetterWithUserProvidedNewPlainTableFormatConfig implements RocksDBConfigSetter { - public RocksDBConfigSetterWithUserProvidedNewPlainTableFormatConfig(){} + public RocksDBConfigSetterWithUserProvidedNewPlainTableFormatConfig() {} public void setConfig(final String storeName, final Options options, final Map configs) { options.setTableFormatConfig(new PlainTableConfig()); diff --git a/streams/src/test/java/org/apache/kafka/streams/utils/TestUtils.java b/streams/src/test/java/org/apache/kafka/streams/utils/TestUtils.java index 24ac2f2306dbe..9bb8c6fe45a89 100644 --- a/streams/src/test/java/org/apache/kafka/streams/utils/TestUtils.java +++ b/streams/src/test/java/org/apache/kafka/streams/utils/TestUtils.java @@ -121,6 +121,7 @@ public static class RecordingProcessorWrapper implements ProcessorWrapper { private Set wrappedProcessorNames; + @SuppressWarnings("unchecked") @Override public void configure(final Map configs) { if (configs.containsKey(PROCESSOR_WRAPPER_COUNTER_CONFIG)) { diff --git a/test-common/test-common-api/src/test/java/org/apache/kafka/common/test/api/ClusterTestExtensionsTest.java b/test-common/test-common-api/src/test/java/org/apache/kafka/common/test/api/ClusterTestExtensionsTest.java index bc20aa1427399..af7a39a92be50 100644 --- a/test-common/test-common-api/src/test/java/org/apache/kafka/common/test/api/ClusterTestExtensionsTest.java +++ b/test-common/test-common-api/src/test/java/org/apache/kafka/common/test/api/ClusterTestExtensionsTest.java @@ -116,7 +116,7 @@ public void testClusterTemplate() { @ClusterConfigProperty(key = "spam", value = "eggs"), @ClusterConfigProperty(key = "default.key", value = "overwrite.value") }, tags = { - "default.display.key1", "default.display.key2" + "default.display.key1", "default.display.key2" }), @ClusterTest(types = {Type.CO_KRAFT}, serverProperties = { @ClusterConfigProperty(key = "foo", value = "baz"), @@ -126,7 +126,7 @@ public void testClusterTemplate() { @ClusterConfigProperty(key = "spam", value = "eggs"), @ClusterConfigProperty(key = "default.key", value = "overwrite.value") }, tags = { - "default.display.key1", "default.display.key2" + "default.display.key1", "default.display.key2" }) }) public void testClusterTests() throws ExecutionException, InterruptedException { diff --git a/tools/src/main/java/org/apache/kafka/tools/ToolsUtils.java b/tools/src/main/java/org/apache/kafka/tools/ToolsUtils.java index a2ec83be81eb4..1f1914df26823 100644 --- a/tools/src/main/java/org/apache/kafka/tools/ToolsUtils.java +++ b/tools/src/main/java/org/apache/kafka/tools/ToolsUtils.java @@ -148,7 +148,7 @@ public static Set duplicates(List s) { * @param Element type. */ @SuppressWarnings("unchecked") - public static Set minus(Set set, T...toRemove) { + public static Set minus(Set set, T... toRemove) { Set res = new HashSet<>(set); for (T t : toRemove) res.remove(t); diff --git a/tools/src/test/java/org/apache/kafka/tools/BrokerApiVersionsCommandTest.java b/tools/src/test/java/org/apache/kafka/tools/BrokerApiVersionsCommandTest.java index f83024837ba57..74919740170d0 100644 --- a/tools/src/test/java/org/apache/kafka/tools/BrokerApiVersionsCommandTest.java +++ b/tools/src/test/java/org/apache/kafka/tools/BrokerApiVersionsCommandTest.java @@ -45,7 +45,7 @@ @ExtendWith(ClusterTestExtensions.class) @ClusterTestDefaults(serverProperties = { - @ClusterConfigProperty(key = ServerConfigs.UNSTABLE_API_VERSIONS_ENABLE_CONFIG, value = "true"), + @ClusterConfigProperty(key = ServerConfigs.UNSTABLE_API_VERSIONS_ENABLE_CONFIG, value = "true"), }) public class BrokerApiVersionsCommandTest { @ClusterTest diff --git a/tools/src/test/java/org/apache/kafka/tools/MetadataQuorumCommandTest.java b/tools/src/test/java/org/apache/kafka/tools/MetadataQuorumCommandTest.java index fa3803a6eaff2..484f09ec5ef00 100644 --- a/tools/src/test/java/org/apache/kafka/tools/MetadataQuorumCommandTest.java +++ b/tools/src/test/java/org/apache/kafka/tools/MetadataQuorumCommandTest.java @@ -63,10 +63,10 @@ public void testDescribeQuorumReplicationSuccessful(ClusterInstance cluster) thr assertTrue(header.matches("NodeId\\s+DirectoryId\\s+LogEndOffset\\s+Lag\\s+LastFetchTimestamp\\s+LastCaughtUpTimestamp\\s+Status\\s+")); - if (cluster.type() == Type.CO_KRAFT) - assertEquals(Math.max(cluster.config().numControllers(), cluster.config().numBrokers()), data.size()); - else - assertEquals(cluster.config().numBrokers() + cluster.config().numControllers(), data.size()); + if (cluster.type() == Type.CO_KRAFT) + assertEquals(Math.max(cluster.config().numControllers(), cluster.config().numBrokers()), data.size()); + else + assertEquals(cluster.config().numBrokers() + cluster.config().numControllers(), data.size()); Pattern leaderPattern = Pattern.compile("\\d+\\s+\\S+\\s+\\d+\\s+\\d+\\s+-?\\d+\\s+-?\\d+\\s+Leader\\s*"); assertTrue(leaderPattern.matcher(data.get(0)).find()); diff --git a/tools/src/test/java/org/apache/kafka/tools/reassign/ReassignPartitionsCommandTest.java b/tools/src/test/java/org/apache/kafka/tools/reassign/ReassignPartitionsCommandTest.java index 389b1584c99f3..ae174d63c6ae5 100644 --- a/tools/src/test/java/org/apache/kafka/tools/reassign/ReassignPartitionsCommandTest.java +++ b/tools/src/test/java/org/apache/kafka/tools/reassign/ReassignPartitionsCommandTest.java @@ -101,16 +101,16 @@ import static org.junit.jupiter.api.Assertions.assertTrue; @ClusterTestDefaults(brokers = 5, disksPerBroker = 3, serverProperties = { - // shorter backoff to reduce test durations when no active partitions are eligible for fetching due to throttling - @ClusterConfigProperty(key = REPLICA_FETCH_BACKOFF_MS_CONFIG, value = "100"), - // Don't move partition leaders automatically. - @ClusterConfigProperty(key = AUTO_LEADER_REBALANCE_ENABLE_CONFIG, value = "false"), - @ClusterConfigProperty(key = REPLICA_LAG_TIME_MAX_MS_CONFIG, value = "1000"), - @ClusterConfigProperty(id = 0, key = "broker.rack", value = "rack0"), - @ClusterConfigProperty(id = 1, key = "broker.rack", value = "rack0"), - @ClusterConfigProperty(id = 2, key = "broker.rack", value = "rack1"), - @ClusterConfigProperty(id = 3, key = "broker.rack", value = "rack1"), - @ClusterConfigProperty(id = 4, key = "broker.rack", value = "rack1"), + // shorter backoff to reduce test durations when no active partitions are eligible for fetching due to throttling + @ClusterConfigProperty(key = REPLICA_FETCH_BACKOFF_MS_CONFIG, value = "100"), + // Don't move partition leaders automatically. + @ClusterConfigProperty(key = AUTO_LEADER_REBALANCE_ENABLE_CONFIG, value = "false"), + @ClusterConfigProperty(key = REPLICA_LAG_TIME_MAX_MS_CONFIG, value = "1000"), + @ClusterConfigProperty(id = 0, key = "broker.rack", value = "rack0"), + @ClusterConfigProperty(id = 1, key = "broker.rack", value = "rack0"), + @ClusterConfigProperty(id = 2, key = "broker.rack", value = "rack1"), + @ClusterConfigProperty(id = 3, key = "broker.rack", value = "rack1"), + @ClusterConfigProperty(id = 4, key = "broker.rack", value = "rack1"), }) @ExtendWith(ClusterTestExtensions.class) public class ReassignPartitionsCommandTest { @@ -133,7 +133,7 @@ public void testReassignment() throws Exception { } @ClusterTests({ - @ClusterTest(types = {Type.KRAFT, Type.CO_KRAFT}, metadataVersion = IBP_3_3_IV0) + @ClusterTest(types = {Type.KRAFT, Type.CO_KRAFT}, metadataVersion = IBP_3_3_IV0) }) public void testReassignmentWithAlterPartitionDisabled() throws Exception { // Test reassignment when the IBP is on an older version which does not use @@ -145,11 +145,11 @@ public void testReassignmentWithAlterPartitionDisabled() throws Exception { } @ClusterTests({ - @ClusterTest(types = {Type.KRAFT, Type.CO_KRAFT}, serverProperties = { - @ClusterConfigProperty(id = 1, key = INTER_BROKER_PROTOCOL_VERSION_CONFIG, value = "3.3-IV0"), - @ClusterConfigProperty(id = 2, key = INTER_BROKER_PROTOCOL_VERSION_CONFIG, value = "3.3-IV0"), - @ClusterConfigProperty(id = 3, key = INTER_BROKER_PROTOCOL_VERSION_CONFIG, value = "3.3-IV0"), - }) + @ClusterTest(types = {Type.KRAFT, Type.CO_KRAFT}, serverProperties = { + @ClusterConfigProperty(id = 1, key = INTER_BROKER_PROTOCOL_VERSION_CONFIG, value = "3.3-IV0"), + @ClusterConfigProperty(id = 2, key = INTER_BROKER_PROTOCOL_VERSION_CONFIG, value = "3.3-IV0"), + @ClusterConfigProperty(id = 3, key = INTER_BROKER_PROTOCOL_VERSION_CONFIG, value = "3.3-IV0"), + }) }) public void testReassignmentCompletionDuringPartialUpgrade() throws Exception { // Test reassignment during a partial upgrade when some brokers are relying on diff --git a/trogdor/src/main/java/org/apache/kafka/trogdor/fault/Kibosh.java b/trogdor/src/main/java/org/apache/kafka/trogdor/fault/Kibosh.java index b6d2801bdff42..eb3de14722833 100644 --- a/trogdor/src/main/java/org/apache/kafka/trogdor/fault/Kibosh.java +++ b/trogdor/src/main/java/org/apache/kafka/trogdor/fault/Kibosh.java @@ -41,8 +41,8 @@ public final class Kibosh { @JsonTypeInfo(use = JsonTypeInfo.Id.NAME, property = "type") @JsonSubTypes({ - @JsonSubTypes.Type(value = KiboshFilesUnreadableFaultSpec.class, name = "unreadable"), - }) + @JsonSubTypes.Type(value = KiboshFilesUnreadableFaultSpec.class, name = "unreadable"), + }) public abstract static class KiboshFaultSpec { @Override public final boolean equals(Object o) { diff --git a/trogdor/src/main/java/org/apache/kafka/trogdor/rest/TaskState.java b/trogdor/src/main/java/org/apache/kafka/trogdor/rest/TaskState.java index 6bb9e7f4f1f26..1f7fa30e6c4b9 100644 --- a/trogdor/src/main/java/org/apache/kafka/trogdor/rest/TaskState.java +++ b/trogdor/src/main/java/org/apache/kafka/trogdor/rest/TaskState.java @@ -30,11 +30,11 @@ */ @JsonTypeInfo(use = JsonTypeInfo.Id.NAME, property = "state") @JsonSubTypes({ - @JsonSubTypes.Type(value = TaskPending.class, name = TaskStateType.Constants.PENDING_VALUE), - @JsonSubTypes.Type(value = TaskRunning.class, name = TaskStateType.Constants.RUNNING_VALUE), - @JsonSubTypes.Type(value = TaskStopping.class, name = TaskStateType.Constants.STOPPING_VALUE), - @JsonSubTypes.Type(value = TaskDone.class, name = TaskStateType.Constants.DONE_VALUE) - }) + @JsonSubTypes.Type(value = TaskPending.class, name = TaskStateType.Constants.PENDING_VALUE), + @JsonSubTypes.Type(value = TaskRunning.class, name = TaskStateType.Constants.RUNNING_VALUE), + @JsonSubTypes.Type(value = TaskStopping.class, name = TaskStateType.Constants.STOPPING_VALUE), + @JsonSubTypes.Type(value = TaskDone.class, name = TaskStateType.Constants.DONE_VALUE) +}) public abstract class TaskState extends Message { private final TaskSpec spec; diff --git a/trogdor/src/main/java/org/apache/kafka/trogdor/workload/RecordProcessor.java b/trogdor/src/main/java/org/apache/kafka/trogdor/workload/RecordProcessor.java index 39a2f8b56b623..821c3a486c49d 100644 --- a/trogdor/src/main/java/org/apache/kafka/trogdor/workload/RecordProcessor.java +++ b/trogdor/src/main/java/org/apache/kafka/trogdor/workload/RecordProcessor.java @@ -28,7 +28,7 @@ */ @JsonTypeInfo(use = JsonTypeInfo.Id.NAME, property = "type") @JsonSubTypes(value = { - @JsonSubTypes.Type(value = TimestampRecordProcessor.class, name = "timestamp"), + @JsonSubTypes.Type(value = TimestampRecordProcessor.class, name = "timestamp"), }) public interface RecordProcessor { void processRecords(ConsumerRecords consumerRecords); From 86b563963dcc26b061436a36371ec95bf1d2f67a Mon Sep 17 00:00:00 2001 From: m1a2st Date: Mon, 2 Dec 2024 23:32:58 +0800 Subject: [PATCH 04/10] update to 10.20.2 --- gradle/dependencies.gradle | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/gradle/dependencies.gradle b/gradle/dependencies.gradle index 985548450a343..cc468644ddcfa 100644 --- a/gradle/dependencies.gradle +++ b/gradle/dependencies.gradle @@ -61,7 +61,7 @@ versions += [ // when updating checkstyle, check whether the exclusion of // CVE-2023-2976 and CVE-2020-8908 can be dropped from // gradle/resources/dependencycheck-suppressions.xml - checkstyle: project.hasProperty('checkstyleVersion') ? checkstyleVersion : "10.20.1", + checkstyle: project.hasProperty('checkstyleVersion') ? checkstyleVersion : "10.20.2", commonsCli: "1.4", commonsIo: "2.14.0", // ZooKeeper dependency. Do not use, this is going away. commonsValidator: "1.9.0", From 326bad434897c0de0aa40c86b4d36cbe3cac1b01 Mon Sep 17 00:00:00 2001 From: m1a2st Date: Mon, 2 Dec 2024 23:46:04 +0800 Subject: [PATCH 05/10] update KafkaAdminClient --- .../kafka/clients/admin/KafkaAdminClient.java | 17 +++++++++-------- 1 file changed, 9 insertions(+), 8 deletions(-) diff --git a/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java b/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java index 4b83a8f304f5a..8f49043db4020 100644 --- a/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java +++ b/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java @@ -182,6 +182,8 @@ import org.apache.kafka.common.requests.AddRaftVoterResponse; import org.apache.kafka.common.requests.AlterClientQuotasRequest; import org.apache.kafka.common.requests.AlterClientQuotasResponse; +import org.apache.kafka.common.requests.AlterConfigsRequest; +import org.apache.kafka.common.requests.AlterConfigsResponse; import org.apache.kafka.common.requests.AlterPartitionReassignmentsRequest; import org.apache.kafka.common.requests.AlterPartitionReassignmentsResponse; import org.apache.kafka.common.requests.AlterReplicaLogDirsRequest; @@ -2866,7 +2868,6 @@ private ConfigEntry.ConfigSource configSource(DescribeConfigsResponse.ConfigSour return configSource; } - @Override @Deprecated public AlterConfigsResult alterConfigs(Map configs, final AlterConfigsOptions options) { final Map> allFutures = new HashMap<>(); @@ -2884,7 +2885,7 @@ public AlterConfigsResult alterConfigs(Map configs, fina unifiedRequestResources.add(resource); } if (!unifiedRequestResources.isEmpty()) - allFutures.putAll(alterConfigs(configs, options, unifiedRequestResources, new LeastLoadedBrokerOrActiveKController())); + allFutures.putAll(alterConfigs(configs, options, unifiedRequestResources, new LeastLoadedBrokerOrActiveKController())); return new AlterConfigsResult(new HashMap<>(allFutures)); } @@ -4888,12 +4889,12 @@ AddRaftVoterRequest.Builder createRequest(int timeoutMs) { setHost(endpoint.host()). setPort(endpoint.port()))); return new AddRaftVoterRequest.Builder( - new AddRaftVoterRequestData(). - setClusterId(options.clusterId().orElse(null)). - setTimeoutMs(timeoutMs). - setVoterId(voterId) . - setVoterDirectoryId(voterDirectoryId). - setListeners(listeners)); + new AddRaftVoterRequestData(). + setClusterId(options.clusterId().orElse(null)). + setTimeoutMs(timeoutMs). + setVoterId(voterId) . + setVoterDirectoryId(voterDirectoryId). + setListeners(listeners)); } @Override From f9a8f5fed15a69b3e90283f7fdb71b79896b9044 Mon Sep 17 00:00:00 2001 From: m1a2st Date: Mon, 2 Dec 2024 23:47:37 +0800 Subject: [PATCH 06/10] update KafkaAdminClient --- .../kafka/clients/admin/KafkaAdminClient.java | 69 +------------------ 1 file changed, 1 insertion(+), 68 deletions(-) diff --git a/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java b/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java index 8f49043db4020..1a58bc651c089 100644 --- a/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java +++ b/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java @@ -182,8 +182,6 @@ import org.apache.kafka.common.requests.AddRaftVoterResponse; import org.apache.kafka.common.requests.AlterClientQuotasRequest; import org.apache.kafka.common.requests.AlterClientQuotasResponse; -import org.apache.kafka.common.requests.AlterConfigsRequest; -import org.apache.kafka.common.requests.AlterConfigsResponse; import org.apache.kafka.common.requests.AlterPartitionReassignmentsRequest; import org.apache.kafka.common.requests.AlterPartitionReassignmentsResponse; import org.apache.kafka.common.requests.AlterReplicaLogDirsRequest; @@ -2867,72 +2865,7 @@ private ConfigEntry.ConfigSource configSource(DescribeConfigsResponse.ConfigSour } return configSource; } - - @Deprecated - public AlterConfigsResult alterConfigs(Map configs, final AlterConfigsOptions options) { - final Map> allFutures = new HashMap<>(); - // We must make a separate AlterConfigs request for every BROKER resource we want to alter - // and send the request to that specific node. Other resources are grouped together into - // a single request that may be sent to any node. - final Collection unifiedRequestResources = new ArrayList<>(); - - for (ConfigResource resource : configs.keySet()) { - Integer node = nodeFor(resource); - if (node != null) { - NodeProvider nodeProvider = new ConstantBrokerOrActiveKController(node); - allFutures.putAll(alterConfigs(configs, options, Collections.singleton(resource), nodeProvider)); - } else - unifiedRequestResources.add(resource); - } - if (!unifiedRequestResources.isEmpty()) - allFutures.putAll(alterConfigs(configs, options, unifiedRequestResources, new LeastLoadedBrokerOrActiveKController())); - return new AlterConfigsResult(new HashMap<>(allFutures)); - } - - private Map> alterConfigs(Map configs, - final AlterConfigsOptions options, - Collection resources, - NodeProvider nodeProvider) { - final Map> futures = new HashMap<>(); - final Map requestMap = new HashMap<>(resources.size()); - for (ConfigResource resource : resources) { - List configEntries = new ArrayList<>(); - for (ConfigEntry configEntry: configs.get(resource).entries()) - configEntries.add(new AlterConfigsRequest.ConfigEntry(configEntry.name(), configEntry.value())); - requestMap.put(resource, new AlterConfigsRequest.Config(configEntries)); - futures.put(resource, new KafkaFutureImpl<>()); - } - - final long now = time.milliseconds(); - runnable.call(new Call("alterConfigs", calcDeadlineMs(now, options.timeoutMs()), nodeProvider) { - - @Override - public AlterConfigsRequest.Builder createRequest(int timeoutMs) { - return new AlterConfigsRequest.Builder(requestMap, options.shouldValidateOnly()); - } - - @Override - public void handleResponse(AbstractResponse abstractResponse) { - AlterConfigsResponse response = (AlterConfigsResponse) abstractResponse; - for (Map.Entry> entry : futures.entrySet()) { - KafkaFutureImpl future = entry.getValue(); - ApiException exception = response.errors().get(entry.getKey()).exception(); - if (exception != null) { - future.completeExceptionally(exception); - } else { - future.complete(null); - } - } - } - - @Override - void handleFailure(Throwable throwable) { - completeAllExceptionally(futures.values(), throwable); - } - }, now); - return futures; - } - + @Override public AlterConfigsResult incrementalAlterConfigs(Map> configs, final AlterConfigsOptions options) { From b1e938aeb8ab7e60a6d9a9b1e4d25d911d1d18ac Mon Sep 17 00:00:00 2001 From: m1a2st Date: Mon, 2 Dec 2024 23:48:10 +0800 Subject: [PATCH 07/10] revert unused change --- .../java/org/apache/kafka/clients/admin/KafkaAdminClient.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java b/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java index 1a58bc651c089..bd135acc74b30 100644 --- a/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java +++ b/clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java @@ -2865,7 +2865,7 @@ private ConfigEntry.ConfigSource configSource(DescribeConfigsResponse.ConfigSour } return configSource; } - + @Override public AlterConfigsResult incrementalAlterConfigs(Map> configs, final AlterConfigsOptions options) { From 84bb5f61888cd22b597c82f11658717efb8dcaf2 Mon Sep 17 00:00:00 2001 From: m1a2st Date: Tue, 3 Dec 2024 00:09:01 +0800 Subject: [PATCH 08/10] revert to 10.20.1 --- gradle/dependencies.gradle | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/gradle/dependencies.gradle b/gradle/dependencies.gradle index cc468644ddcfa..985548450a343 100644 --- a/gradle/dependencies.gradle +++ b/gradle/dependencies.gradle @@ -61,7 +61,7 @@ versions += [ // when updating checkstyle, check whether the exclusion of // CVE-2023-2976 and CVE-2020-8908 can be dropped from // gradle/resources/dependencycheck-suppressions.xml - checkstyle: project.hasProperty('checkstyleVersion') ? checkstyleVersion : "10.20.2", + checkstyle: project.hasProperty('checkstyleVersion') ? checkstyleVersion : "10.20.1", commonsCli: "1.4", commonsIo: "2.14.0", // ZooKeeper dependency. Do not use, this is going away. commonsValidator: "1.9.0", From b8b5b82186b70a72a9d887732c5e8ed5a2c0470a Mon Sep 17 00:00:00 2001 From: Chia-Ping Tsai Date: Tue, 3 Dec 2024 06:55:57 +0800 Subject: [PATCH 09/10] use 10.20.2 --- gradle/dependencies.gradle | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/gradle/dependencies.gradle b/gradle/dependencies.gradle index 985548450a343..cc468644ddcfa 100644 --- a/gradle/dependencies.gradle +++ b/gradle/dependencies.gradle @@ -61,7 +61,7 @@ versions += [ // when updating checkstyle, check whether the exclusion of // CVE-2023-2976 and CVE-2020-8908 can be dropped from // gradle/resources/dependencycheck-suppressions.xml - checkstyle: project.hasProperty('checkstyleVersion') ? checkstyleVersion : "10.20.1", + checkstyle: project.hasProperty('checkstyleVersion') ? checkstyleVersion : "10.20.2", commonsCli: "1.4", commonsIo: "2.14.0", // ZooKeeper dependency. Do not use, this is going away. commonsValidator: "1.9.0", From 6e4e391b63f0019ad35fa3c8c22fee08ebed9c72 Mon Sep 17 00:00:00 2001 From: m1a2st Date: Wed, 4 Dec 2024 05:17:38 +0800 Subject: [PATCH 10/10] remove comments --- gradle/dependencies.gradle | 3 --- 1 file changed, 3 deletions(-) diff --git a/gradle/dependencies.gradle b/gradle/dependencies.gradle index cc468644ddcfa..bc4d6fcea3437 100644 --- a/gradle/dependencies.gradle +++ b/gradle/dependencies.gradle @@ -58,9 +58,6 @@ versions += [ // but currently, tests are failing in >=3.1.2. Therefore, we are temporarily using version 3.1.1. // The failing tests should be fixed under KAFKA-18089, allowing us to upgrade to >=3.1.2. caffeine: "3.1.1", - // when updating checkstyle, check whether the exclusion of - // CVE-2023-2976 and CVE-2020-8908 can be dropped from - // gradle/resources/dependencycheck-suppressions.xml checkstyle: project.hasProperty('checkstyleVersion') ? checkstyleVersion : "10.20.2", commonsCli: "1.4", commonsIo: "2.14.0", // ZooKeeper dependency. Do not use, this is going away.