From 9d88e2d8ea697eb09ccc255e6f6090f61b68a0b5 Mon Sep 17 00:00:00 2001 From: Yi Hu Date: Tue, 7 Oct 2025 10:34:54 -0400 Subject: [PATCH] Fix execute gradle task for examples --- examples/java/build.gradle | 28 --------------------------- examples/java/common.gradle | 31 ++++++++++++++++++++++++++++++ examples/java/iceberg/build.gradle | 28 --------------------------- examples/java/sql/build.gradle | 26 +------------------------ 4 files changed, 32 insertions(+), 81 deletions(-) diff --git a/examples/java/build.gradle b/examples/java/build.gradle index cdbcb5ce8bf9..08aad3dbaff5 100644 --- a/examples/java/build.gradle +++ b/examples/java/build.gradle @@ -38,26 +38,6 @@ artifact includes all Apache Beam Java SDK examples.""" apply from: "$projectDir/common.gradle" -/** Define the list of runners which execute a precommit test. - * Some runners are run from separate projects, see the preCommit task below - * for details. - */ -def preCommitRunners = ["directRunner", "flinkRunner", "sparkRunner"] -// The following runners have configuration created but not added to preCommit -def nonPreCommitRunners = ["dataflowRunner", "prismRunner"] -for (String runner : preCommitRunners) { - configurations.create(runner + "PreCommit") -} -for (String runner: nonPreCommitRunners) { - configurations.create(runner + "PreCommit") -} -configurations.sparkRunnerPreCommit { - // Ban certain dependencies to prevent a StackOverflow within Spark - // because JUL -> SLF4J -> JUL, and similarly JDK14 -> SLF4J -> JDK14 - exclude group: "org.slf4j", module: "jul-to-slf4j" - exclude group: "org.slf4j", module: "slf4j-jdk14" -} - dependencies { implementation enforcedPlatform(library.java.google_cloud_platform_libraries_bom) implementation library.java.vendored_guava_32_1_2_jre @@ -124,14 +104,6 @@ dependencies { for (String runner : preCommitRunners) { delegate.add(runner + "PreCommit", project(path: ":examples:java", configuration: "testRuntimeMigration")) } - directRunnerPreCommit project(path: ":runners:direct-java", configuration: "shadow") - flinkRunnerPreCommit project(":runners:flink:${project.ext.latestFlinkVersion}") - sparkRunnerPreCommit project(":runners:spark:3") - sparkRunnerPreCommit project(":sdks:java:io:hadoop-file-system") - dataflowRunnerPreCommit project(":runners:google-cloud-dataflow-java") - dataflowRunnerPreCommit project(":runners:google-cloud-dataflow-java:worker") // v2 worker - dataflowRunnerPreCommit project(":sdks:java:harness") // v2 worker - prismRunnerPreCommit project(":runners:prism:java") // Add dependency if requested on command line for runner if (project.hasProperty("runnerDependency")) { diff --git a/examples/java/common.gradle b/examples/java/common.gradle index b8a3ef27f9a8..10ea43628bc8 100644 --- a/examples/java/common.gradle +++ b/examples/java/common.gradle @@ -16,6 +16,37 @@ * limitations under the License. */ +/** Define the list of runners which execute a precommit test. + * Some runners are run from separate projects, see the preCommit task below + * for details. + */ +project.ext.preCommitRunners = ["directRunner", "flinkRunner", "sparkRunner"] +// The following runners have configuration created but not added to preCommit +project.ext.nonPreCommitRunners = ["dataflowRunner", "prismRunner"] +for (String runner : ext.preCommitRunners) { + configurations.create(runner + "PreCommit") +} +for (String runner: ext.nonPreCommitRunners) { + configurations.create(runner + "PreCommit") +} +configurations.sparkRunnerPreCommit { + // Ban certain dependencies to prevent a StackOverflow within Spark + // because JUL -> SLF4J -> JUL, and similarly JDK14 -> SLF4J -> JDK14 + exclude group: "org.slf4j", module: "jul-to-slf4j" + exclude group: "org.slf4j", module: "slf4j-jdk14" +} + +dependencies { + directRunnerPreCommit project(path: ":runners:direct-java", configuration: "shadow") + flinkRunnerPreCommit project(":runners:flink:${project.ext.latestFlinkVersion}") + sparkRunnerPreCommit project(":runners:spark:3") + sparkRunnerPreCommit project(":sdks:java:io:hadoop-file-system") + dataflowRunnerPreCommit project(":runners:google-cloud-dataflow-java") + dataflowRunnerPreCommit project(":runners:google-cloud-dataflow-java:worker") // v1 worker + dataflowRunnerPreCommit project(":sdks:java:harness") // v2 worker + prismRunnerPreCommit project(":runners:prism:java") +} + /* * A convenient task to run individual example directly on Beam repo. * diff --git a/examples/java/iceberg/build.gradle b/examples/java/iceberg/build.gradle index 4d258e9be5ac..4d4a1fb44413 100644 --- a/examples/java/iceberg/build.gradle +++ b/examples/java/iceberg/build.gradle @@ -35,26 +35,6 @@ ext.summary = """Apache Beam Java SDK examples using IcebergIO.""" apply from: "$project.rootDir/examples/java/common.gradle" -/** Define the list of runners which execute a precommit test. - * Some runners are run from separate projects, see the preCommit task below - * for details. - */ -def preCommitRunners = ["directRunner", "flinkRunner", "sparkRunner"] -// The following runners have configuration created but not added to preCommit -def nonPreCommitRunners = ["dataflowRunner", "prismRunner"] -for (String runner : preCommitRunners) { - configurations.create(runner + "PreCommit") -} -for (String runner: nonPreCommitRunners) { - configurations.create(runner + "PreCommit") -} -configurations.sparkRunnerPreCommit { - // Ban certain dependencies to prevent a StackOverflow within Spark - // because JUL -> SLF4J -> JUL, and similarly JDK14 -> SLF4J -> JDK14 - exclude group: "org.slf4j", module: "jul-to-slf4j" - exclude group: "org.slf4j", module: "slf4j-jdk14" -} - dependencies { implementation enforcedPlatform(library.java.google_cloud_platform_libraries_bom) runtimeOnly project(":sdks:java:io:iceberg") @@ -75,14 +55,6 @@ dependencies { for (String runner : preCommitRunners) { delegate.add(runner + "PreCommit", project(path: ":examples:java", configuration: "testRuntimeMigration")) } - directRunnerPreCommit project(path: ":runners:direct-java", configuration: "shadow") - flinkRunnerPreCommit project(":runners:flink:${project.ext.latestFlinkVersion}") - sparkRunnerPreCommit project(":runners:spark:3") - sparkRunnerPreCommit project(":sdks:java:io:hadoop-file-system") - dataflowRunnerPreCommit project(":runners:google-cloud-dataflow-java") - dataflowRunnerPreCommit project(":runners:google-cloud-dataflow-java:worker") // v2 worker - dataflowRunnerPreCommit project(":sdks:java:harness") // v2 worker - prismRunnerPreCommit project(":runners:prism:java") // Add dependency if requested on command line for runner if (project.hasProperty("runnerDependency")) { diff --git a/examples/java/sql/build.gradle b/examples/java/sql/build.gradle index af61989f11c5..730b6a5620aa 100644 --- a/examples/java/sql/build.gradle +++ b/examples/java/sql/build.gradle @@ -36,20 +36,7 @@ ext.summary = """Apache Beam SDK provides a simple, Java-based interface for processing virtually any size data. This artifact includes all Apache Beam Java SDK examples.""" -/** Define the list of runners which execute a precommit test. - * Some runners are run from separate projects, see the preCommit task below - * for details. - */ -def preCommitRunners = ["directRunner", "flinkRunner", "sparkRunner"] -for (String runner : preCommitRunners) { - configurations.create(runner + "PreCommit") -} -configurations.sparkRunnerPreCommit { - // Ban certain dependencies to prevent a StackOverflow within Spark - // because JUL -> SLF4J -> JUL, and similarly JDK14 -> SLF4J -> JDK14 - exclude group: "org.slf4j", module: "jul-to-slf4j" - exclude group: "org.slf4j", module: "slf4j-jdk14" -} +apply from: "$project.rootDir/examples/java/common.gradle" dependencies { implementation enforcedPlatform(library.java.google_cloud_platform_libraries_bom) @@ -70,10 +57,6 @@ dependencies { for (String runner : preCommitRunners) { delegate.add(runner + "PreCommit", project(path: ":examples:java", configuration: "testRuntimeMigration")) } - directRunnerPreCommit project(path: ":runners:direct-java", configuration: "shadow") - flinkRunnerPreCommit project(":runners:flink:${project.ext.latestFlinkVersion}") - sparkRunnerPreCommit project(":runners:spark:3") - sparkRunnerPreCommit project(":sdks:java:io:hadoop-file-system") // Add dependency if requested on command line for runner if (project.hasProperty("runnerDependency")) { @@ -113,10 +96,3 @@ task preCommit() { dependsOn runner + "PreCommit" } } - -tasks.create(name:"execute", type:JavaExec) { - main = project.hasProperty("mainClass") ? project.getProperty("mainClass") : "NONE" - classpath = sourceSets.main.runtimeClasspath - systemProperties System.getProperties() - args project.hasProperty("exec.args") ? project.getProperty("exec.args").split() : [] -}