From ddd1ef573a5318748763fafc974e4f7d8876fd6f Mon Sep 17 00:00:00 2001 From: Sean Owen Date: Wed, 10 Apr 2024 10:52:01 -0500 Subject: [PATCH] Update for 0.18.0, move CICD configs to supported Spark versions (#680) --- ...est_spark_3_2_java_8.yml => test_spark_3_3_java_8.yml} | 4 ++-- .github/workflows/test_spark_3_4_java_11.yml | 4 ++-- ...t_spark_3_3_java_11.yml => test_spark_3_5_java_11.yml} | 4 ++-- README.md | 8 ++++---- build.sbt | 6 +++--- 5 files changed, 13 insertions(+), 13 deletions(-) rename .github/workflows/{test_spark_3_2_java_8.yml => test_spark_3_3_java_8.yml} (76%) rename .github/workflows/{test_spark_3_3_java_11.yml => test_spark_3_5_java_11.yml} (67%) diff --git a/.github/workflows/test_spark_3_2_java_8.yml b/.github/workflows/test_spark_3_3_java_8.yml similarity index 76% rename from .github/workflows/test_spark_3_2_java_8.yml rename to .github/workflows/test_spark_3_3_java_8.yml index 7e1e5450..29d9a873 100644 --- a/.github/workflows/test_spark_3_2_java_8.yml +++ b/.github/workflows/test_spark_3_3_java_8.yml @@ -1,4 +1,4 @@ -name: Spark 3.2 / Java 8 / Scala 2.12 +name: Spark 3.3 / Java 8 / Scala 2.12 on: push: branches: [master] @@ -14,4 +14,4 @@ jobs: with: java-version: 'adopt@1.8' - name: Build and test - run: sbt -Dspark.testVersion=3.2.4 ++2.12.15 clean mimaReportBinaryIssues test + run: sbt -Dspark.testVersion=3.3.4 ++2.12.15 clean mimaReportBinaryIssues test diff --git a/.github/workflows/test_spark_3_4_java_11.yml b/.github/workflows/test_spark_3_4_java_11.yml index bf907a15..46b52131 100644 --- a/.github/workflows/test_spark_3_4_java_11.yml +++ b/.github/workflows/test_spark_3_4_java_11.yml @@ -1,4 +1,4 @@ -name: Spark 3.4 / Java 11 / Scala 2.13 +name: Spark 3.4 / Java 11 / Scala 2.12 on: push: branches: [master] @@ -14,4 +14,4 @@ jobs: with: java-version: 'adopt@1.11' - name: Build and test - run: sbt -Dspark.testVersion=3.4.1 ++2.13.8 clean scalastyle test:scalastyle mimaReportBinaryIssues test + run: sbt -Dspark.testVersion=3.4.1 ++2.12.15 clean mimaReportBinaryIssues test diff --git a/.github/workflows/test_spark_3_3_java_11.yml b/.github/workflows/test_spark_3_5_java_11.yml similarity index 67% rename from .github/workflows/test_spark_3_3_java_11.yml rename to .github/workflows/test_spark_3_5_java_11.yml index bbcb1e29..2479e644 100644 --- a/.github/workflows/test_spark_3_3_java_11.yml +++ b/.github/workflows/test_spark_3_5_java_11.yml @@ -1,4 +1,4 @@ -name: Spark 3.3 / Java 11 / Scala 2.13 +name: Spark 3.5 / Java 11 / Scala 2.13 on: push: branches: [master] @@ -14,4 +14,4 @@ jobs: with: java-version: 'adopt@1.11' - name: Build and test - run: sbt -Dspark.testVersion=3.3.3 ++2.13.8 clean mimaReportBinaryIssues test + run: sbt -Dspark.testVersion=3.5.1 ++2.13.8 clean scalastyle test:scalastyle mimaReportBinaryIssues test diff --git a/README.md b/README.md index 95cd2851..a39f8d9e 100644 --- a/README.md +++ b/README.md @@ -16,7 +16,7 @@ You can link against this library in your program at the following coordinates: ``` groupId: com.databricks artifactId: spark-xml_2.12 -version: 0.17.0 +version: 0.18.0 ``` ## Using with Spark shell @@ -24,7 +24,7 @@ version: 0.17.0 This package can be added to Spark using the `--packages` command line option. For example, to include it when starting the spark shell: ``` -$SPARK_HOME/bin/spark-shell --packages com.databricks:spark-xml_2.12:0.17.0 +$SPARK_HOME/bin/spark-shell --packages com.databricks:spark-xml_2.12:0.18.0 ``` ## Features @@ -410,7 +410,7 @@ Automatically infer schema (data types) ```R library(SparkR) -sparkR.session("local[4]", sparkPackages = c("com.databricks:spark-xml_2.12:0.17.0")) +sparkR.session("local[4]", sparkPackages = c("com.databricks:spark-xml_2.12:0.18.0")) df <- read.df("books.xml", source = "xml", rowTag = "book") @@ -422,7 +422,7 @@ You can manually specify schema: ```R library(SparkR) -sparkR.session("local[4]", sparkPackages = c("com.databricks:spark-xml_2.12:0.17.0")) +sparkR.session("local[4]", sparkPackages = c("com.databricks:spark-xml_2.12:0.18.0")) customSchema <- structType( structField("_id", "string"), structField("author", "string"), diff --git a/build.sbt b/build.sbt index 3c3292e2..5f6cf78c 100755 --- a/build.sbt +++ b/build.sbt @@ -2,7 +2,7 @@ import com.typesafe.tools.mima.core.MissingClassProblem name := "spark-xml" -version := "0.17.0" +version := "0.18.0" organization := "com.databricks" @@ -12,7 +12,7 @@ crossScalaVersions := Seq("2.12.15", "2.13.8") scalacOptions := Seq("-unchecked", "-deprecation") -val sparkVersion = sys.props.get("spark.testVersion").getOrElse("3.4.1") +val sparkVersion = sys.props.get("spark.testVersion").getOrElse("3.5.1") // To avoid packaging it, it's Provided below autoScalaLibrary := false @@ -81,7 +81,7 @@ fork := true // Prints JUnit tests in output Test / testOptions := Seq(Tests.Argument(TestFrameworks.JUnit, "-v")) -mimaPreviousArtifacts := Set("com.databricks" %% "spark-xml" % "0.16.0") +mimaPreviousArtifacts := Set("com.databricks" %% "spark-xml" % "0.17.0") mimaBinaryIssueFilters ++= { Seq()