mirror of https://github.com/apache/kafka.git
MINOR remove some flaky annotations (#18357)
Remove the flaky annotation from the following tests * RemoteLogManagerTest#testFetchOffsetByTimestampWithTieredStorageDoesNotFetchIndexWhenExistsLocally * All the children of BaseConsumerTest#testCoordinatorFailover * TransactionsTest#testFailureToFenceEpoch * TransactionsTest#testReadCommittedConsumerShouldNotSeeUndecidedData * MetricsDuringTopicCreationDeletionTest#testMetricsDuringTopicCreateDelete * ProduceRequestTest#testProduceWithInvalidTimestamp Reviewers: Chia-Ping Tsai <chia7712@gmail.com>
This commit is contained in:
parent
dd0fd55316
commit
a2a8d87153
|
@ -36,7 +36,6 @@ import org.apache.kafka.common.record.RemoteLogInputStream;
|
||||||
import org.apache.kafka.common.record.SimpleRecord;
|
import org.apache.kafka.common.record.SimpleRecord;
|
||||||
import org.apache.kafka.common.requests.FetchRequest;
|
import org.apache.kafka.common.requests.FetchRequest;
|
||||||
import org.apache.kafka.common.security.auth.SecurityProtocol;
|
import org.apache.kafka.common.security.auth.SecurityProtocol;
|
||||||
import org.apache.kafka.common.test.api.Flaky;
|
|
||||||
import org.apache.kafka.common.utils.MockTime;
|
import org.apache.kafka.common.utils.MockTime;
|
||||||
import org.apache.kafka.common.utils.Time;
|
import org.apache.kafka.common.utils.Time;
|
||||||
import org.apache.kafka.server.common.OffsetAndEpoch;
|
import org.apache.kafka.server.common.OffsetAndEpoch;
|
||||||
|
@ -1683,7 +1682,6 @@ public class RemoteLogManagerTest {
|
||||||
remoteLogManager.onLeadershipChange(Collections.singleton(mockPartition(leaderTopicIdPartition)), Collections.emptySet(), topicIds);
|
remoteLogManager.onLeadershipChange(Collections.singleton(mockPartition(leaderTopicIdPartition)), Collections.emptySet(), topicIds);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Flaky("KAFKA-17779")
|
|
||||||
@Test
|
@Test
|
||||||
void testFetchOffsetByTimestampWithTieredStorageDoesNotFetchIndexWhenExistsLocally() throws Exception {
|
void testFetchOffsetByTimestampWithTieredStorageDoesNotFetchIndexWhenExistsLocally() throws Exception {
|
||||||
TopicPartition tp = new TopicPartition("sample", 0);
|
TopicPartition tp = new TopicPartition("sample", 0);
|
||||||
|
|
|
@ -23,7 +23,6 @@ import org.apache.kafka.common.header.Headers
|
||||||
import org.apache.kafka.common.{ClusterResource, ClusterResourceListener, PartitionInfo}
|
import org.apache.kafka.common.{ClusterResource, ClusterResourceListener, PartitionInfo}
|
||||||
import org.apache.kafka.common.internals.Topic
|
import org.apache.kafka.common.internals.Topic
|
||||||
import org.apache.kafka.common.serialization.{ByteArrayDeserializer, ByteArraySerializer, Deserializer, Serializer}
|
import org.apache.kafka.common.serialization.{ByteArrayDeserializer, ByteArraySerializer, Deserializer, Serializer}
|
||||||
import org.apache.kafka.common.test.api.Flaky
|
|
||||||
import org.junit.jupiter.api.Assertions._
|
import org.junit.jupiter.api.Assertions._
|
||||||
import org.junit.jupiter.params.ParameterizedTest
|
import org.junit.jupiter.params.ParameterizedTest
|
||||||
import org.junit.jupiter.params.provider.MethodSource
|
import org.junit.jupiter.params.provider.MethodSource
|
||||||
|
@ -80,7 +79,6 @@ abstract class BaseConsumerTest extends AbstractConsumerTest {
|
||||||
assertNotEquals(0, BaseConsumerTest.updateConsumerCount.get())
|
assertNotEquals(0, BaseConsumerTest.updateConsumerCount.get())
|
||||||
}
|
}
|
||||||
|
|
||||||
@Flaky("KAFKA-15920")
|
|
||||||
@ParameterizedTest(name = TestInfoUtils.TestWithParameterizedQuorumAndGroupProtocolNames)
|
@ParameterizedTest(name = TestInfoUtils.TestWithParameterizedQuorumAndGroupProtocolNames)
|
||||||
@MethodSource(Array("getTestQuorumAndGroupProtocolParametersAll"))
|
@MethodSource(Array("getTestQuorumAndGroupProtocolParametersAll"))
|
||||||
def testCoordinatorFailover(quorum: String, groupProtocol: String): Unit = {
|
def testCoordinatorFailover(quorum: String, groupProtocol: String): Unit = {
|
||||||
|
|
|
@ -23,7 +23,6 @@ import org.apache.kafka.clients.consumer._
|
||||||
import org.apache.kafka.clients.producer.{KafkaProducer, ProducerRecord}
|
import org.apache.kafka.clients.producer.{KafkaProducer, ProducerRecord}
|
||||||
import org.apache.kafka.common.TopicPartition
|
import org.apache.kafka.common.TopicPartition
|
||||||
import org.apache.kafka.common.errors.{InvalidProducerEpochException, ProducerFencedException, TimeoutException}
|
import org.apache.kafka.common.errors.{InvalidProducerEpochException, ProducerFencedException, TimeoutException}
|
||||||
import org.apache.kafka.common.test.api.Flaky
|
|
||||||
import org.apache.kafka.coordinator.group.GroupCoordinatorConfig
|
import org.apache.kafka.coordinator.group.GroupCoordinatorConfig
|
||||||
import org.apache.kafka.coordinator.transaction.{TransactionLogConfig, TransactionStateManagerConfig}
|
import org.apache.kafka.coordinator.transaction.{TransactionLogConfig, TransactionStateManagerConfig}
|
||||||
import org.apache.kafka.server.config.{ReplicationConfigs, ServerConfigs, ServerLogConfigs}
|
import org.apache.kafka.server.config.{ReplicationConfigs, ServerConfigs, ServerLogConfigs}
|
||||||
|
@ -172,7 +171,6 @@ class TransactionsTest extends IntegrationTestHarness {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Flaky("KAFKA-18036")
|
|
||||||
@ParameterizedTest(name = TestInfoUtils.TestWithParameterizedQuorumAndGroupProtocolNames)
|
@ParameterizedTest(name = TestInfoUtils.TestWithParameterizedQuorumAndGroupProtocolNames)
|
||||||
@MethodSource(Array("getTestQuorumAndGroupProtocolParametersAll"))
|
@MethodSource(Array("getTestQuorumAndGroupProtocolParametersAll"))
|
||||||
def testReadCommittedConsumerShouldNotSeeUndecidedData(quorum: String, groupProtocol: String): Unit = {
|
def testReadCommittedConsumerShouldNotSeeUndecidedData(quorum: String, groupProtocol: String): Unit = {
|
||||||
|
@ -850,7 +848,6 @@ class TransactionsTest extends IntegrationTestHarness {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Flaky("KAFKA-18306")
|
|
||||||
@ParameterizedTest(name = "{displayName}.quorum={0}.groupProtocol={1}.isTV2Enabled={2}")
|
@ParameterizedTest(name = "{displayName}.quorum={0}.groupProtocol={1}.isTV2Enabled={2}")
|
||||||
@CsvSource(Array(
|
@CsvSource(Array(
|
||||||
"kraft, classic, false",
|
"kraft, classic, false",
|
||||||
|
|
|
@ -24,7 +24,6 @@ import kafka.utils.{Logging, TestUtils}
|
||||||
import scala.jdk.CollectionConverters._
|
import scala.jdk.CollectionConverters._
|
||||||
import org.junit.jupiter.api.{BeforeEach, TestInfo}
|
import org.junit.jupiter.api.{BeforeEach, TestInfo}
|
||||||
import com.yammer.metrics.core.Gauge
|
import com.yammer.metrics.core.Gauge
|
||||||
import org.apache.kafka.common.test.api.Flaky
|
|
||||||
import org.apache.kafka.server.config.{ReplicationConfigs, ServerConfigs, ServerLogConfigs}
|
import org.apache.kafka.server.config.{ReplicationConfigs, ServerConfigs, ServerLogConfigs}
|
||||||
import org.apache.kafka.server.metrics.KafkaYammerMetrics
|
import org.apache.kafka.server.metrics.KafkaYammerMetrics
|
||||||
import org.junit.jupiter.params.ParameterizedTest
|
import org.junit.jupiter.params.ParameterizedTest
|
||||||
|
@ -72,7 +71,6 @@ class MetricsDuringTopicCreationDeletionTest extends KafkaServerTestHarness with
|
||||||
*/
|
*/
|
||||||
@ParameterizedTest
|
@ParameterizedTest
|
||||||
@ValueSource(strings = Array("kraft"))
|
@ValueSource(strings = Array("kraft"))
|
||||||
@Flaky("KAFKA-18245")
|
|
||||||
def testMetricsDuringTopicCreateDelete(quorum: String): Unit = {
|
def testMetricsDuringTopicCreateDelete(quorum: String): Unit = {
|
||||||
|
|
||||||
// For UnderReplicatedPartitions, because of https://issues.apache.org/jira/browse/KAFKA-4605
|
// For UnderReplicatedPartitions, because of https://issues.apache.org/jira/browse/KAFKA-4605
|
||||||
|
|
|
@ -103,7 +103,7 @@ class ProduceRequestTest extends BaseRequestTest {
|
||||||
}).toMap
|
}).toMap
|
||||||
}
|
}
|
||||||
|
|
||||||
@ParameterizedTest(name = "quorum=kraft")
|
@ParameterizedTest
|
||||||
@MethodSource(Array("timestampConfigProvider"))
|
@MethodSource(Array("timestampConfigProvider"))
|
||||||
def testProduceWithInvalidTimestamp(messageTimeStampConfig: String, recordTimestamp: Long): Unit = {
|
def testProduceWithInvalidTimestamp(messageTimeStampConfig: String, recordTimestamp: Long): Unit = {
|
||||||
val topic = "topic"
|
val topic = "topic"
|
||||||
|
|
Loading…
Reference in New Issue