Go to file
Colin Patrick McCabe 62ea4c46a9
KAFKA-13749: CreateTopics in KRaft must return configs (#11941)
Previously, when in KRaft mode, CreateTopics did not return the active configurations for the
topic(s) it had just created. This PR addresses that gap. We will now return these topic
configuration(s) when the user has DESCRIBE_CONFIGS permission. (In the case where the user does
not have this permission, we will omit the configurations and set TopicErrorCode. We will also omit
the number of partitions and replication factor data as well.)

For historical reasons, we use different names to refer to each topic configuration when it is set
in the broker context, as opposed to the topic context. For example, the topic configuration
"segment.ms" corresponds to the broker configuration "log.roll.ms". Additionally, some broker
configurations have synonyms. For example, the broker configuration "log.roll.hours" can be used to
set the log roll time instead of "log.roll.ms". In order to track all of this, this PR adds a
table in LogConfig.scala which maps each topic configuration to an ordered list of ConfigSynonym
classes. (This table is then passed to KafkaConfigSchema as a constructor argument.)

Some synonyms require transformations. For example, in order to convert from "log.roll.hours" to
"segment.ms", we must convert hours to milliseconds. (Note that our assumption right now is that
topic configurations do not have synonyms, only broker configurations. If this changes, we will
need to add some logic to handle it.)

This PR makes the 8-argument constructor for ConfigEntry public. We need this in order to make full
use of ConfigEntry outside of the admin namespace. This change is probably inevitable in general
since otherwise we cannot easily test the output from various admin APIs in junit tests outside the
admin package.

Testing:

This PR adds PlaintextAdminIntegrationTest#testCreateTopicsReturnsConfigs. This test validates
some of the configurations that it gets back from the call to CreateTopics, rather than just checking
if it got back a non-empty map like some of the existing tests. In order to test the
configuration override logic, testCreateDeleteTopics now sets up some custom static and dynamic
configurations.

In QuorumTestHarness, we now allow tests to configure what the ID of the controller should be. This
allows us to set dynamic configurations for the controller in testCreateDeleteTopics. We will have
a more complete fix for setting dynamic configuations on the controller later.

This PR changes ConfigurationControlManager so that it is created via a Builder. This will make it
easier to add more parameters to its constructor without having to update every piece of test code
that uses it. It will also make the test code easier to read.

Reviewers: David Arthur <mumrah@gmail.com>
2022-04-01 10:50:25 -07:00
bin KAFKA-13748: Do not include file stream connectors in Connect's CLASSPATH and plugin.path by default (#11908) 2022-03-30 13:15:42 -07:00
checkstyle KAFKA-13749: CreateTopics in KRaft must return configs (#11941) 2022-04-01 10:50:25 -07:00
clients KAFKA-13749: CreateTopics in KRaft must return configs (#11941) 2022-04-01 10:50:25 -07:00
config MINOR: Fix and clarify kraft README and example configuration files (#11616) 2022-02-15 10:27:38 -08:00
connect KAFKA-13719: Fix connector restart cause duplicate tasks (#11869) 2022-03-30 14:58:58 +02:00
core KAFKA-13749: CreateTopics in KRaft must return configs (#11941) 2022-04-01 10:50:25 -07:00
docs KAFKA-13748: Do not include file stream connectors in Connect's CLASSPATH and plugin.path by default (#11908) 2022-03-30 13:15:42 -07:00
examples KAFKA-9922: Update demo instructions in examples README (#8559) 2020-04-29 19:31:26 -07:00
generator/src KAFKA-13277; Fix size calculation for tagged string fields in message generator (#11308) 2021-09-07 13:02:45 -07:00
gradle KAFKA-13660: Switch log4j12 to reload4j (#11743) 2022-03-30 20:54:01 +02:00
jmh-benchmarks MINOR: jmh.sh swallows compile errors (#11870) 2022-03-10 18:18:41 -05:00
licenses MINOR: Add missing licenses and update versions in LICENSE-binary for 3.0 (#11232) 2021-08-18 21:58:36 -07:00
log4j-appender/src KAFKA-13761: KafkaLog4jAppender deadlocks when idempotence is enabled (#11939) 2022-03-24 06:36:56 -07:00
metadata KAFKA-13749: CreateTopics in KRaft must return configs (#11941) 2022-04-01 10:50:25 -07:00
raft Polish Javadoc for EpochState (#11897) 2022-03-15 19:58:47 -04:00
server-common/src MINOR: Move `KafkaYammerMetrics` to server-common (#11970) 2022-03-30 13:59:22 -07:00
shell/src MINOR: show LogRecoveryState in MetadataShell and fix log message 2022-03-21 14:33:51 -07:00
storage MINOR: Cleanup admin creation logic in integration tests (#11790) 2022-02-24 07:37:28 -08:00
streams KAFKA-13785: add processor metadata to be committed with offset (#11829) 2022-03-31 09:48:21 -07:00
tests KAFKA-13748: Do not include file stream connectors in Connect's CLASSPATH and plugin.path by default (#11908) 2022-03-30 13:15:42 -07:00
tools KAFKA-13728: fix PushHttpMetricsReporter no longer pushes metrics when network failure is recovered. (#11879) 2022-03-19 21:09:28 -07:00
trogdor MINOR: MiniTrogdorCluster mutates objects from other threads (#11710) 2022-02-03 10:57:21 +01:00
vagrant MINOR: Bump latest 3.0 version to 3.0.1 (#11885) 2022-03-16 11:43:37 +01:00
.asf.yaml MINOR: Add ableegoldman and cadonna to asf whitelist (#9171) 2020-08-12 10:32:03 -07:00
.gitignore MINOR: exclude all `src/generated` and `src/generated-test` (#10671) 2021-05-12 15:27:49 +08:00
CONTRIBUTING.md
HEADER
Jenkinsfile KAFKA-13671: Add ppc64le build stage (#11833) 2022-03-07 10:18:54 +01:00
LICENSE KAFKA-12602: Fix LICENSE file (#10474) 2021-04-06 15:08:24 -05:00
LICENSE-binary KAFKA-13660: Switch log4j12 to reload4j (#11743) 2022-03-30 20:54:01 +02:00
NOTICE MINOR: Update year in NOTICE (#11670) 2022-01-12 09:46:50 +01:00
NOTICE-binary KAFKA-12625: Fix the NOTICE file (#10693) 2021-07-29 13:53:29 -05:00
PULL_REQUEST_TEMPLATE.md
README.md KAFKA-13714: Fix cache flush position (#11926) 2022-03-23 22:09:05 -05:00
TROGDOR.md
Vagrantfile
build.gradle MINOR: Move `KafkaYammerMetrics` to server-common (#11970) 2022-03-30 13:59:22 -07:00
doap_Kafka.rdf MINOR: Fix category in doap_Kafka.rdf (#11423) 2021-10-27 10:53:48 +02:00
gradle.properties MINOR: Bump trunk to 3.3.0-SNAPSHOT (#11925) 2022-03-21 21:37:05 +01:00
gradlew MINOR: Upgrade to Gradle 7.3.3 (#11518) 2022-02-02 05:47:05 -08:00
gradlewAll
kafka-merge-pr.py MINOR: Bump trunk to 3.3.0-SNAPSHOT (#11925) 2022-03-21 21:37:05 +01:00
release.py KAFKA-10712; Update release scripts to Python3 (#11538) 2021-11-29 10:42:14 +01:00
release_notes.py KAFKA-10712; Update release scripts to Python3 (#11538) 2021-11-29 10:42:14 +01:00
settings.gradle MINOR: Add 3.0 and 3.1 to streams system tests (#11716) 2022-01-28 10:06:31 +01:00
wrapper.gradle MINOR: Upgrade gradle to 6.8 and test retry plugin to 1.2.0 (#9849) 2021-01-18 09:04:20 -08:00

README.md

Apache Kafka

See our web site for details on the project.

You need to have Java installed.

We build and test Apache Kafka with Java 8, 11 and 17. We set the release parameter in javac and scalac to 8 to ensure the generated binaries are compatible with Java 8 or higher (independently of the Java version used for compilation). Java 8 support has been deprecated since Apache Kafka 3.0 and will be removed in Apache Kafka 4.0 (see KIP-750 for more details).

Scala 2.12 and 2.13 are supported and 2.13 is used by default. Scala 2.12 support has been deprecated since Apache Kafka 3.0 and will be removed in Apache Kafka 4.0 (see KIP-751 for more details). See below for how to use a specific Scala version or all of the supported Scala versions.

Build a jar and run it

./gradlew jar

Follow instructions in https://kafka.apache.org/quickstart

Build source jar

./gradlew srcJar

Build aggregated javadoc

./gradlew aggregatedJavadoc

Build javadoc and scaladoc

./gradlew javadoc
./gradlew javadocJar # builds a javadoc jar for each module
./gradlew scaladoc
./gradlew scaladocJar # builds a scaladoc jar for each module
./gradlew docsJar # builds both (if applicable) javadoc and scaladoc jars for each module

Run unit/integration tests

./gradlew test # runs both unit and integration tests
./gradlew unitTest
./gradlew integrationTest

Force re-running tests without code change

./gradlew -Prerun-tests test
./gradlew -Prerun-tests unitTest
./gradlew -Prerun-tests integrationTest

Running a particular unit/integration test

./gradlew clients:test --tests RequestResponseTest

Repeatedly running a particular unit/integration test

I=0; while ./gradlew clients:test -Prerun-tests --tests RequestResponseTest --fail-fast; do (( I=$I+1 )); echo "Completed run: $I"; sleep 1; done

Running a particular test method within a unit/integration test

./gradlew core:test --tests kafka.api.ProducerFailureHandlingTest.testCannotSendToInternalTopic
./gradlew clients:test --tests org.apache.kafka.clients.MetadataTest.testTimeToNextUpdate

Running a particular unit/integration test with log4j output

Change the log4j setting in either clients/src/test/resources/log4j.properties or core/src/test/resources/log4j.properties

./gradlew clients:test --tests RequestResponseTest

Specifying test retries

By default, each failed test is retried once up to a maximum of five retries per test run. Tests are retried at the end of the test task. Adjust these parameters in the following way:

./gradlew test -PmaxTestRetries=1 -PmaxTestRetryFailures=5

See Test Retry Gradle Plugin for more details.

Generating test coverage reports

Generate coverage reports for the whole project:

./gradlew reportCoverage -PenableTestCoverage=true -Dorg.gradle.parallel=false

Generate coverage for a single module, i.e.:

./gradlew clients:reportCoverage -PenableTestCoverage=true -Dorg.gradle.parallel=false

Building a binary release gzipped tar ball

./gradlew clean releaseTarGz

The release file can be found inside ./core/build/distributions/.

Building auto generated messages

Sometimes it is only necessary to rebuild the RPC auto-generated message data when switching between branches, as they could fail due to code changes. You can just run:

./gradlew processMessages processTestMessages

Running a Kafka broker in ZooKeeper mode

./bin/zookeeper-server-start.sh config/zookeeper.properties
./bin/kafka-server-start.sh config/server.properties

Running a Kafka broker in KRaft (Kafka Raft metadata) mode

See config/kraft/README.md.

Cleaning the build

./gradlew clean

Running a task with one of the Scala versions available (2.12.x or 2.13.x)

Note that if building the jars with a version other than 2.13.x, you need to set the SCALA_VERSION variable or change it in bin/kafka-run-class.sh to run the quick start.

You can pass either the major version (eg 2.12) or the full version (eg 2.12.7):

./gradlew -PscalaVersion=2.12 jar
./gradlew -PscalaVersion=2.12 test
./gradlew -PscalaVersion=2.12 releaseTarGz

Running a task with all the scala versions enabled by default

Invoke the gradlewAll script followed by the task(s):

./gradlewAll test
./gradlewAll jar
./gradlewAll releaseTarGz

Running a task for a specific project

This is for core, examples and clients

./gradlew core:jar
./gradlew core:test

Streams has multiple sub-projects, but you can run all the tests:

./gradlew :streams:testAll

Listing all gradle tasks

./gradlew tasks

Building IDE project

Note that this is not strictly necessary (IntelliJ IDEA has good built-in support for Gradle projects, for example).

./gradlew eclipse
./gradlew idea

The eclipse task has been configured to use ${project_dir}/build_eclipse as Eclipse's build directory. Eclipse's default build directory (${project_dir}/bin) clashes with Kafka's scripts directory and we don't use Gradle's build directory to avoid known issues with this configuration.

Publishing the jar for all versions of Scala and for all projects to maven

The recommended command is:

./gradlewAll publish

For backwards compatibility, the following also works:

./gradlewAll uploadArchives

Please note for this to work you should create/update ${GRADLE_USER_HOME}/gradle.properties (typically, ~/.gradle/gradle.properties) and assign the following variables

mavenUrl=
mavenUsername=
mavenPassword=
signing.keyId=
signing.password=
signing.secretKeyRingFile=

Publishing the streams quickstart archetype artifact to maven

For the Streams archetype project, one cannot use gradle to upload to maven; instead the mvn deploy command needs to be called at the quickstart folder:

cd streams/quickstart
mvn deploy

Please note for this to work you should create/update user maven settings (typically, ${USER_HOME}/.m2/settings.xml) to assign the following variables

<settings xmlns="http://maven.apache.org/SETTINGS/1.0.0"
   xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
   xsi:schemaLocation="http://maven.apache.org/SETTINGS/1.0.0
                       https://maven.apache.org/xsd/settings-1.0.0.xsd">
...                           
<servers>
   ...
   <server>
      <id>apache.snapshots.https</id>
      <username>${maven_username}</username>
      <password>${maven_password}</password>
   </server>
   <server>
      <id>apache.releases.https</id>
      <username>${maven_username}</username>
      <password>${maven_password}</password>
    </server>
    ...
 </servers>
 ...

Installing ALL the jars to the local Maven repository

The recommended command to build for both Scala 2.12 and 2.13 is:

./gradlewAll publishToMavenLocal

For backwards compatibility, the following also works:

./gradlewAll install

Installing specific projects to the local Maven repository

./gradlew -PskipSigning :streams:publishToMavenLocal

If needed, you can specify the Scala version with -PscalaVersion=2.13.

Building the test jar

./gradlew testJar

Determining how transitive dependencies are added

./gradlew core:dependencies --configuration runtime

Determining if any dependencies could be updated

./gradlew dependencyUpdates

Running code quality checks

There are two code quality analysis tools that we regularly run, spotbugs and checkstyle.

Checkstyle

Checkstyle enforces a consistent coding style in Kafka. You can run checkstyle using:

./gradlew checkstyleMain checkstyleTest

The checkstyle warnings will be found in reports/checkstyle/reports/main.html and reports/checkstyle/reports/test.html files in the subproject build directories. They are also printed to the console. The build will fail if Checkstyle fails.

Spotbugs

Spotbugs uses static analysis to look for bugs in the code. You can run spotbugs using:

./gradlew spotbugsMain spotbugsTest -x test

The spotbugs warnings will be found in reports/spotbugs/main.html and reports/spotbugs/test.html files in the subproject build directories. Use -PxmlSpotBugsReport=true to generate an XML report instead of an HTML one.

JMH microbenchmarks

We use JMH to write microbenchmarks that produce reliable results in the JVM.

See jmh-benchmarks/README.md for details on how to run the microbenchmarks.

Common build options

The following options should be set with a -P switch, for example ./gradlew -PmaxParallelForks=1 test.

  • commitId: sets the build commit ID as .git/HEAD might not be correct if there are local commits added for build purposes.
  • mavenUrl: sets the URL of the maven deployment repository (file://path/to/repo can be used to point to a local repository).
  • maxParallelForks: maximum number of test processes to start in parallel. Defaults to the number of processors available to the JVM.
  • maxScalacThreads: maximum number of worker threads for the scalac backend. Defaults to the lowest of 8 and the number of processors available to the JVM. The value must be between 1 and 16 (inclusive).
  • ignoreFailures: ignore test failures from junit
  • showStandardStreams: shows standard out and standard error of the test JVM(s) on the console.
  • skipSigning: skips signing of artifacts.
  • testLoggingEvents: unit test events to be logged, separated by comma. For example ./gradlew -PtestLoggingEvents=started,passed,skipped,failed test.
  • xmlSpotBugsReport: enable XML reports for spotBugs. This also disables HTML reports as only one can be enabled at a time.
  • maxTestRetries: maximum number of retries for a failing test case.
  • maxTestRetryFailures: maximum number of test failures before retrying is disabled for subsequent tests.
  • enableTestCoverage: enables test coverage plugins and tasks, including bytecode enhancement of classes required to track said coverage. Note that this introduces some overhead when running tests and hence why it's disabled by default (the overhead varies, but 15-20% is a reasonable estimate).
  • scalaOptimizerMode: configures the optimizing behavior of the scala compiler, the value should be one of none, method, inline-kafka or inline-scala (the default is inline-kafka). none is the scala compiler default, which only eliminates unreachable code. method also includes method-local optimizations. inline-kafka adds inlining of methods within the kafka packages. Finally, inline-scala also includes inlining of methods within the scala library (which avoids lambda allocations for methods like Option.exists). inline-scala is only safe if the Scala library version is the same at compile time and runtime. Since we cannot guarantee this for all cases (for example, users may depend on the kafka jar for integration tests where they may include a scala library with a different version), we don't enable it by default. See https://www.lightbend.com/blog/scala-inliner-optimizer for more details.

Dependency Analysis

The gradle dependency debugging documentation mentions using the dependencies or dependencyInsight tasks to debug dependencies for the root project or individual subprojects.

Alternatively, use the allDeps or allDepInsight tasks for recursively iterating through all subprojects:

./gradlew allDeps

./gradlew allDepInsight --configuration runtimeClasspath --dependency com.fasterxml.jackson.core:jackson-databind

These take the same arguments as the builtin variants.

Running system tests

See tests/README.md.

Running in Vagrant

See vagrant/README.md.

Contribution

Apache Kafka is interested in building the community; we would welcome any thoughts or patches. You can reach us on the Apache mailing lists.

To contribute follow the instructions here: