diff --git a/clients/src/test/java/org/apache/kafka/clients/consumer/ConsumerConfigTest.java b/clients/src/test/java/org/apache/kafka/clients/consumer/ConsumerConfigTest.java index 48580c2657d12..c72c2f8698d31 100644 --- a/clients/src/test/java/org/apache/kafka/clients/consumer/ConsumerConfigTest.java +++ b/clients/src/test/java/org/apache/kafka/clients/consumer/ConsumerConfigTest.java @@ -29,8 +29,8 @@ public class ConsumerConfigTest { - private final Deserializer keyDeserializer = new ByteArrayDeserializer(); - private final Deserializer valueDeserializer = new StringDeserializer(); + private final Deserializer keyDeserializer = new ByteArrayDeserializer(); + private final Deserializer valueDeserializer = new StringDeserializer(); private final String keyDeserializerClassName = keyDeserializer.getClass().getName(); private final String valueDeserializerClassName = valueDeserializer.getClass().getName(); private final Object keyDeserializerClass = keyDeserializer.getClass(); diff --git a/clients/src/test/java/org/apache/kafka/clients/consumer/internals/ConsumerCoordinatorTest.java b/clients/src/test/java/org/apache/kafka/clients/consumer/internals/ConsumerCoordinatorTest.java index a9ecdad203f23..6ea23ba9ec85c 100644 --- a/clients/src/test/java/org/apache/kafka/clients/consumer/internals/ConsumerCoordinatorTest.java +++ b/clients/src/test/java/org/apache/kafka/clients/consumer/internals/ConsumerCoordinatorTest.java @@ -53,7 +53,6 @@ import org.apache.kafka.common.message.SyncGroupResponseData; import org.apache.kafka.common.metrics.KafkaMetric; import org.apache.kafka.common.metrics.Metrics; -import org.apache.kafka.common.metrics.Sensor; import org.apache.kafka.common.protocol.Errors; import org.apache.kafka.common.record.RecordBatch; import org.apache.kafka.common.requests.AbstractRequest; @@ -252,10 +251,6 @@ private KafkaMetric getMetric(final String name) { return metrics.metrics().get(metrics.metricName(name, "consumer" + groupId + "-coordinator-metrics")); } - private Sensor getSensor(final String name) { - return metrics.sensor(name); - } - @Test public void testSelectRebalanceProtcol() { List assignors = new ArrayList<>(); diff --git a/clients/src/test/java/org/apache/kafka/clients/consumer/internals/FetcherTest.java b/clients/src/test/java/org/apache/kafka/clients/consumer/internals/FetcherTest.java index d0eba44bedea8..7cbc1fc3bb401 100644 --- a/clients/src/test/java/org/apache/kafka/clients/consumer/internals/FetcherTest.java +++ b/clients/src/test/java/org/apache/kafka/clients/consumer/internals/FetcherTest.java @@ -135,7 +135,6 @@ import static org.mockito.Mockito.doAnswer; import static org.mockito.Mockito.spy; -@SuppressWarnings("deprecation") public class FetcherTest { private static final double EPSILON = 0.0001; @@ -2015,7 +2014,7 @@ public void testQuotaMetrics() { ClientRequest request = client.newClientRequest(node.idString(), builder, time.milliseconds(), true); client.send(request, time.milliseconds()); client.poll(1, time.milliseconds()); - FetchResponse response = fullFetchResponse(tp0, nextRecords, Errors.NONE, i, throttleTimeMs); + FetchResponse response = fullFetchResponse(tp0, nextRecords, Errors.NONE, i, throttleTimeMs); buffer = response.serialize(ApiKeys.FETCH, ApiKeys.FETCH.latestVersion(), request.correlationId()); @@ -3032,7 +3031,7 @@ public void testConsumingViaIncrementalFetchRequests() { 2, 0L, null, this.records)); partitions1.put(tp1, new FetchResponse.PartitionData<>(Errors.NONE, 100L, FetchResponse.INVALID_LAST_STABLE_OFFSET, 0L, null, emptyRecords)); - FetchResponse resp1 = new FetchResponse<>(Errors.NONE, partitions1, 0, 123); + FetchResponse resp1 = new FetchResponse<>(Errors.NONE, partitions1, 0, 123); client.prepareResponse(resp1); assertEquals(1, fetcher.sendFetches()); assertFalse(fetcher.hasCompletedFetches()); @@ -3058,7 +3057,7 @@ public void testConsumingViaIncrementalFetchRequests() { // The second response contains no new records. LinkedHashMap> partitions2 = new LinkedHashMap<>(); - FetchResponse resp2 = new FetchResponse<>(Errors.NONE, partitions2, 0, 123); + FetchResponse resp2 = new FetchResponse<>(Errors.NONE, partitions2, 0, 123); client.prepareResponse(resp2); assertEquals(1, fetcher.sendFetches()); consumerClient.poll(time.timer(0)); @@ -3071,7 +3070,7 @@ public void testConsumingViaIncrementalFetchRequests() { LinkedHashMap> partitions3 = new LinkedHashMap<>(); partitions3.put(tp0, new FetchResponse.PartitionData<>(Errors.NONE, 100L, 4, 0L, null, this.nextRecords)); - FetchResponse resp3 = new FetchResponse<>(Errors.NONE, partitions3, 0, 123); + FetchResponse resp3 = new FetchResponse<>(Errors.NONE, partitions3, 0, 123); client.prepareResponse(resp3); assertEquals(1, fetcher.sendFetches()); consumerClient.poll(time.timer(0)); @@ -3130,7 +3129,7 @@ public Builder newBuilder() { } @Override - public boolean handleResponse(FetchResponse response) { + public boolean handleResponse(FetchResponse response) { verifySessionPartitions(); return handler.handleResponse(response); } diff --git a/clients/src/test/java/org/apache/kafka/clients/consumer/internals/PartitionAssignorAdapterTest.java b/clients/src/test/java/org/apache/kafka/clients/consumer/internals/PartitionAssignorAdapterTest.java index 5a5ef66f7b711..ba16a52a7d79e 100644 --- a/clients/src/test/java/org/apache/kafka/clients/consumer/internals/PartitionAssignorAdapterTest.java +++ b/clients/src/test/java/org/apache/kafka/clients/consumer/internals/PartitionAssignorAdapterTest.java @@ -101,7 +101,6 @@ public void shouldThrowKafkaExceptionOnListWithNonAssignorClassType() { } @Test - @SuppressWarnings("deprecation") public void testOnAssignment() { OldPartitionAssignor oldAssignor = new OldPartitionAssignor(); ConsumerPartitionAssignor adaptedAssignor = new PartitionAssignorAdapter(oldAssignor); diff --git a/clients/src/test/java/org/apache/kafka/clients/producer/KafkaProducerTest.java b/clients/src/test/java/org/apache/kafka/clients/producer/KafkaProducerTest.java index fe3e9d6481f64..d11ab84c8f7b6 100644 --- a/clients/src/test/java/org/apache/kafka/clients/producer/KafkaProducerTest.java +++ b/clients/src/test/java/org/apache/kafka/clients/producer/KafkaProducerTest.java @@ -132,8 +132,8 @@ public void testOverwriteAcksAndRetriesForIdempotentProducers() { Properties props = new Properties(); props.setProperty(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, "localhost:9999"); props.setProperty(ProducerConfig.TRANSACTIONAL_ID_CONFIG, "transactionalId"); - props.setProperty(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG, new StringSerializer().getClass().getName()); - props.setProperty(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG, new StringSerializer().getClass().getName()); + props.setProperty(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG, StringSerializer.class.getName()); + props.setProperty(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG, StringSerializer.class.getName()); ProducerConfig config = new ProducerConfig(props); assertTrue(config.getBoolean(ProducerConfig.ENABLE_IDEMPOTENCE_CONFIG)); @@ -456,7 +456,7 @@ Sender newSender(LogContext logContext, KafkaClient kafkaClient, ProducerMetadat // Four request updates where the topic isn't present, at which point the timeout expires and a // TimeoutException is thrown - Future future = producer.send(record); + Future future = producer.send(record); verify(metadata, times(4)).requestUpdateForTopic(topic); verify(metadata, times(4)).awaitUpdate(anyInt(), anyLong()); verify(metadata, times(5)).fetch(); @@ -532,7 +532,7 @@ Sender newSender(LogContext logContext, KafkaClient kafkaClient, ProducerMetadat // Four request updates where the requested partition is out of range, at which point the timeout expires // and a TimeoutException is thrown - Future future = producer.send(record); + Future future = producer.send(record); verify(metadata, times(4)).requestUpdateForTopic(topic); verify(metadata, times(4)).awaitUpdate(anyInt(), anyLong()); verify(metadata, times(5)).fetch(); @@ -643,9 +643,7 @@ public void testHeaders() { private > void doTestHeaders(Class serializerClassToMock) { Map configs = new HashMap<>(); configs.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, "localhost:9999"); - @SuppressWarnings("unchecked") // it is safe to suppress, since this is a mock class Serializer keySerializer = mock(serializerClassToMock); - @SuppressWarnings("unchecked") Serializer valueSerializer = mock(serializerClassToMock); long nowMs = Time.SYSTEM.milliseconds(); @@ -688,7 +686,7 @@ private > void doTestHeaders(Class serializerCla public void closeShouldBeIdempotent() { Properties producerProps = new Properties(); producerProps.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, "localhost:9000"); - Producer producer = new KafkaProducer<>(producerProps, new ByteArraySerializer(), new ByteArraySerializer()); + Producer producer = new KafkaProducer<>(producerProps, new ByteArraySerializer(), new ByteArraySerializer()); producer.close(); producer.close(); } @@ -697,12 +695,12 @@ public void closeShouldBeIdempotent() { public void testMetricConfigRecordingLevel() { Properties props = new Properties(); props.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, "localhost:9000"); - try (KafkaProducer producer = new KafkaProducer<>(props, new ByteArraySerializer(), new ByteArraySerializer())) { + try (KafkaProducer producer = new KafkaProducer<>(props, new ByteArraySerializer(), new ByteArraySerializer())) { assertEquals(Sensor.RecordingLevel.INFO, producer.metrics.config().recordLevel()); } props.put(ProducerConfig.METRICS_RECORDING_LEVEL_CONFIG, "DEBUG"); - try (KafkaProducer producer = new KafkaProducer<>(props, new ByteArraySerializer(), new ByteArraySerializer())) { + try (KafkaProducer producer = new KafkaProducer<>(props, new ByteArraySerializer(), new ByteArraySerializer())) { assertEquals(Sensor.RecordingLevel.DEBUG, producer.metrics.config().recordLevel()); } } diff --git a/clients/src/test/java/org/apache/kafka/common/config/AbstractConfigTest.java b/clients/src/test/java/org/apache/kafka/common/config/AbstractConfigTest.java index b5fff6f78d74d..22ce576db1f76 100644 --- a/clients/src/test/java/org/apache/kafka/common/config/AbstractConfigTest.java +++ b/clients/src/test/java/org/apache/kafka/common/config/AbstractConfigTest.java @@ -407,7 +407,7 @@ public void testAutoConfigResolutionWithInvalidConfigProviderClass() { "org.apache.kafka.common.config.provider.InvalidConfigProvider"); props.put("testKey", "${test:/foo/bar/testpath:testKey}"); try { - TestIndirectConfigResolution config = new TestIndirectConfigResolution(props); + new TestIndirectConfigResolution(props); fail("Expected a config exception due to invalid props :" + props); } catch (KafkaException e) { // this is good diff --git a/clients/src/test/java/org/apache/kafka/common/metrics/stats/FrequenciesTest.java b/clients/src/test/java/org/apache/kafka/common/metrics/stats/FrequenciesTest.java index bf7647a9afb40..061f60c18fda3 100644 --- a/clients/src/test/java/org/apache/kafka/common/metrics/stats/FrequenciesTest.java +++ b/clients/src/test/java/org/apache/kafka/common/metrics/stats/FrequenciesTest.java @@ -101,7 +101,6 @@ public void testBooleanFrequencies() { } @Test - @SuppressWarnings("deprecation") public void testUseWithMetrics() { MetricName name1 = name("1"); MetricName name2 = name("2"); diff --git a/clients/src/test/java/org/apache/kafka/common/network/SslTransportLayerTest.java b/clients/src/test/java/org/apache/kafka/common/network/SslTransportLayerTest.java index 126af6ee2fe57..a494d500b4cef 100644 --- a/clients/src/test/java/org/apache/kafka/common/network/SslTransportLayerTest.java +++ b/clients/src/test/java/org/apache/kafka/common/network/SslTransportLayerTest.java @@ -277,7 +277,6 @@ public void testEndpointIdentificationDisabled() throws Exception { sslServerConfigs = getTrustingConfig(serverCertStores, clientCertStores); sslClientConfigs = getTrustingConfig(clientCertStores, serverCertStores); - SecurityProtocol securityProtocol = SecurityProtocol.SSL; server = createEchoServer(SecurityProtocol.SSL); InetSocketAddress addr = new InetSocketAddress("localhost", server.port()); diff --git a/clients/src/test/java/org/apache/kafka/common/protocol/ErrorsTest.java b/clients/src/test/java/org/apache/kafka/common/protocol/ErrorsTest.java index e4243842b8bd7..2c68b86fb8ae4 100644 --- a/clients/src/test/java/org/apache/kafka/common/protocol/ErrorsTest.java +++ b/clients/src/test/java/org/apache/kafka/common/protocol/ErrorsTest.java @@ -40,7 +40,7 @@ public void testUniqueErrorCodes() { @Test public void testUniqueExceptions() { - Set exceptionSet = new HashSet<>(); + Set> exceptionSet = new HashSet<>(); for (Errors error : Errors.values()) { if (error != Errors.NONE) exceptionSet.add(error.exception().getClass()); diff --git a/clients/src/test/java/org/apache/kafka/common/requests/RequestResponseTest.java b/clients/src/test/java/org/apache/kafka/common/requests/RequestResponseTest.java index aedb94ecd4464..babc61fe99b5b 100644 --- a/clients/src/test/java/org/apache/kafka/common/requests/RequestResponseTest.java +++ b/clients/src/test/java/org/apache/kafka/common/requests/RequestResponseTest.java @@ -642,7 +642,7 @@ public void testFetchResponseV4() { 6, FetchResponse.INVALID_LOG_START_OFFSET, Optional.empty(), emptyList(), records)); FetchResponse response = new FetchResponse<>(Errors.NONE, responseData, 10, INVALID_SESSION_ID); - FetchResponse deserialized = FetchResponse.parse(toBuffer(response.toStruct((short) 4)), (short) 4); + FetchResponse deserialized = FetchResponse.parse(toBuffer(response.toStruct((short) 4)), (short) 4); assertEquals(responseData, deserialized.responseData()); } @@ -656,7 +656,7 @@ public void verifyFetchResponseFullWrites() throws Exception { } } - private void verifyFetchResponseFullWrite(short apiVersion, FetchResponse fetchResponse) throws Exception { + private void verifyFetchResponseFullWrite(short apiVersion, FetchResponse fetchResponse) throws Exception { int correlationId = 15; short responseHeaderVersion = FETCH.responseHeaderVersion(apiVersion); @@ -1038,14 +1038,6 @@ private SyncGroupRequest createSyncGroupRequest(int version) { .setProtocolName("range") // Added in v5 but ignorable .setAssignments(assignments); - JoinGroupRequestData.JoinGroupRequestProtocolCollection protocols = - new JoinGroupRequestData.JoinGroupRequestProtocolCollection( - Collections.singleton( - new JoinGroupRequestData.JoinGroupRequestProtocol() - .setName("consumer-range") - .setMetadata(new byte[0])).iterator() - ); - // v3 and above could set group instance id if (version >= 3) data.setGroupInstanceId("groupInstanceId"); diff --git a/clients/src/test/java/org/apache/kafka/common/security/authenticator/SaslServerAuthenticatorTest.java b/clients/src/test/java/org/apache/kafka/common/security/authenticator/SaslServerAuthenticatorTest.java index 2778d73e3e431..0b4c63bc42cb1 100644 --- a/clients/src/test/java/org/apache/kafka/common/security/authenticator/SaslServerAuthenticatorTest.java +++ b/clients/src/test/java/org/apache/kafka/common/security/authenticator/SaslServerAuthenticatorTest.java @@ -31,7 +31,6 @@ import org.apache.kafka.common.security.auth.SecurityProtocol; import org.apache.kafka.common.protocol.types.Struct; import org.apache.kafka.common.requests.RequestHeader; -import org.apache.kafka.common.security.JaasContext; import org.apache.kafka.common.security.plain.PlainLoginModule; import org.apache.kafka.common.utils.AppInfoParser; import org.apache.kafka.common.utils.Time; @@ -152,8 +151,6 @@ private SaslServerAuthenticator setupAuthenticator(Map configs, Trans String mechanism, ChannelMetadataRegistry metadataRegistry) throws IOException { TestJaasConfig jaasConfig = new TestJaasConfig(); jaasConfig.addEntry("jaasContext", PlainLoginModule.class.getName(), new HashMap()); - Map jaasContexts = Collections.singletonMap(mechanism, - new JaasContext("jaasContext", JaasContext.Type.SERVER, jaasConfig, null)); Map subjects = Collections.singletonMap(mechanism, new Subject()); Map callbackHandlers = Collections.singletonMap( mechanism, new SaslServerCallbackHandler()); diff --git a/clients/src/test/java/org/apache/kafka/common/security/ssl/SslFactoryTest.java b/clients/src/test/java/org/apache/kafka/common/security/ssl/SslFactoryTest.java index fab0bf0eca554..6be9e9ba0c1af 100644 --- a/clients/src/test/java/org/apache/kafka/common/security/ssl/SslFactoryTest.java +++ b/clients/src/test/java/org/apache/kafka/common/security/ssl/SslFactoryTest.java @@ -376,15 +376,6 @@ private SslEngineBuilder.SecurityStore sslKeyStore(Map sslConfig ); } - private SslEngineBuilder.SecurityStore sslTrustStore(Map sslConfig) { - return new SslEngineBuilder.SecurityStore( - (String) sslConfig.get(SslConfigs.SSL_TRUSTSTORE_TYPE_CONFIG), - (String) sslConfig.get(SslConfigs.SSL_TRUSTSTORE_LOCATION_CONFIG), - (Password) sslConfig.get(SslConfigs.SSL_TRUSTSTORE_PASSWORD_CONFIG), - null - ); - } - private TestSslUtils.SslConfigsBuilder sslConfigsBuilder(Mode mode) { return new TestSslUtils.SslConfigsBuilder(mode).tlsProtocol(tlsProtocol); } diff --git a/clients/src/test/java/org/apache/kafka/common/utils/UtilsTest.java b/clients/src/test/java/org/apache/kafka/common/utils/UtilsTest.java index 983232a6bc37a..c0e5fc8f454f7 100755 --- a/clients/src/test/java/org/apache/kafka/common/utils/UtilsTest.java +++ b/clients/src/test/java/org/apache/kafka/common/utils/UtilsTest.java @@ -72,8 +72,8 @@ public void testMurmur2() { cases.put("lkjh234lh9fiuh90y23oiuhsafujhadof229phr9h19h89h8".getBytes(), -58897971); cases.put(new byte[]{'a', 'b', 'c'}, 479470107); - for (Map.Entry c : cases.entrySet()) { - assertEquals((int) c.getValue(), murmur2((byte[]) c.getKey())); + for (Map.Entry c : cases.entrySet()) { + assertEquals(c.getValue().intValue(), murmur2(c.getKey())); } } diff --git a/clients/src/test/java/org/apache/kafka/test/TestUtils.java b/clients/src/test/java/org/apache/kafka/test/TestUtils.java index ad2ad9950344e..1e9c76eb0fdf3 100644 --- a/clients/src/test/java/org/apache/kafka/test/TestUtils.java +++ b/clients/src/test/java/org/apache/kafka/test/TestUtils.java @@ -276,8 +276,8 @@ public static File tempDirectory(final Path parent, String prefix) { } public static Properties producerConfig(final String bootstrapServers, - final Class keySerializer, - final Class valueSerializer, + final Class keySerializer, + final Class valueSerializer, final Properties additional) { final Properties properties = new Properties(); properties.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, bootstrapServers); @@ -289,14 +289,14 @@ public static Properties producerConfig(final String bootstrapServers, return properties; } - public static Properties producerConfig(final String bootstrapServers, final Class keySerializer, final Class valueSerializer) { + public static Properties producerConfig(final String bootstrapServers, final Class keySerializer, final Class valueSerializer) { return producerConfig(bootstrapServers, keySerializer, valueSerializer, new Properties()); } public static Properties consumerConfig(final String bootstrapServers, final String groupId, - final Class keyDeserializer, - final Class valueDeserializer, + final Class keyDeserializer, + final Class valueDeserializer, final Properties additional) { final Properties consumerConfig = new Properties(); @@ -311,8 +311,8 @@ public static Properties consumerConfig(final String bootstrapServers, public static Properties consumerConfig(final String bootstrapServers, final String groupId, - final Class keyDeserializer, - final Class valueDeserializer) { + final Class keyDeserializer, + final Class valueDeserializer) { return consumerConfig(bootstrapServers, groupId, keyDeserializer, @@ -323,7 +323,7 @@ public static Properties consumerConfig(final String bootstrapServers, /** * returns consumer config with random UUID for the Group ID */ - public static Properties consumerConfig(final String bootstrapServers, final Class keyDeserializer, final Class valueDeserializer) { + public static Properties consumerConfig(final String bootstrapServers, final Class keyDeserializer, final Class valueDeserializer) { return consumerConfig(bootstrapServers, UUID.randomUUID().toString(), keyDeserializer,