|
1 | 1 | name := "spark-xml" |
2 | 2 |
|
3 | | -version := "0.14.0" |
| 3 | +version := "0.15.0" |
4 | 4 |
|
5 | 5 | organization := "com.databricks" |
6 | 6 |
|
7 | | -scalaVersion := "2.12.10" |
| 7 | +scalaVersion := "2.12.15" |
8 | 8 |
|
9 | | -crossScalaVersions := Seq("2.12.10", "2.13.5") |
| 9 | +crossScalaVersions := Seq("2.12.15", "2.13.8") |
10 | 10 |
|
11 | 11 | scalacOptions := Seq("-unchecked", "-deprecation") |
12 | 12 |
|
13 | | -val sparkVersion = sys.props.get("spark.testVersion").getOrElse("3.2.0") |
| 13 | +val sparkVersion = sys.props.get("spark.testVersion").getOrElse("3.2.1") |
14 | 14 |
|
15 | 15 | // To avoid packaging it, it's Provided below |
16 | 16 | autoScalaLibrary := false |
17 | 17 |
|
18 | 18 | libraryDependencies ++= Seq( |
19 | | - "commons-io" % "commons-io" % "2.8.0", |
20 | | - "org.glassfish.jaxb" % "txw2" % "2.3.4", |
21 | | - "org.apache.ws.xmlschema" % "xmlschema-core" % "2.2.5", |
22 | | - "org.slf4j" % "slf4j-api" % "1.7.30" % Provided, |
23 | | - "org.scalatest" %% "scalatest" % "3.2.9" % Test, |
| 19 | + "commons-io" % "commons-io" % "2.11.0", |
| 20 | + "org.glassfish.jaxb" % "txw2" % "3.0.2", |
| 21 | + "org.apache.ws.xmlschema" % "xmlschema-core" % "2.3.0", |
| 22 | + "org.slf4j" % "slf4j-api" % "1.7.36" % Provided, |
| 23 | + "org.scalatest" %% "scalatest" % "3.2.12" % Test, |
24 | 24 | "com.novocode" % "junit-interface" % "0.11" % Test, |
25 | 25 | "org.apache.spark" %% "spark-core" % sparkVersion % Provided, |
26 | 26 | "org.apache.spark" %% "spark-sql" % sparkVersion % Provided, |
@@ -78,35 +78,11 @@ fork := true |
78 | 78 | // Prints JUnit tests in output |
79 | 79 | testOptions in Test := Seq(Tests.Argument(TestFrameworks.JUnit, "-v")) |
80 | 80 |
|
81 | | -mimaPreviousArtifacts := Set("com.databricks" %% "spark-xml" % "0.12.0") |
| 81 | +mimaPreviousArtifacts := Set("com.databricks" %% "spark-xml" % "0.14.0") |
82 | 82 |
|
83 | 83 | mimaBinaryIssueFilters ++= { |
84 | 84 | import com.typesafe.tools.mima.core.ProblemFilters.exclude |
85 | 85 | import com.typesafe.tools.mima.core.DirectMissingMethodProblem |
86 | 86 | Seq( |
87 | | - exclude[DirectMissingMethodProblem]( |
88 | | - "com.databricks.spark.xml.parsers.StaxXmlParser.convertField"), |
89 | | - exclude[DirectMissingMethodProblem]( |
90 | | - "com.databricks.spark.xml.util.TypeCast.parseXmlTimestamp"), |
91 | | - exclude[DirectMissingMethodProblem]( |
92 | | - "com.databricks.spark.xml.util.TypeCast.supportedXmlTimestampFormatters"), |
93 | | - exclude[DirectMissingMethodProblem]( |
94 | | - "com.databricks.spark.xml.util.TypeCast.parseXmlDate"), |
95 | | - exclude[DirectMissingMethodProblem]( |
96 | | - "com.databricks.spark.xml.util.TypeCast.supportedXmlDateFormatters"), |
97 | | - exclude[DirectMissingMethodProblem]( |
98 | | - "com.databricks.spark.xml.util.TypeCast.supportedXmlDateFormatters"), |
99 | | - exclude[DirectMissingMethodProblem]( |
100 | | - "com.databricks.spark.xml.util.TypeCast.parseXmlDate"), |
101 | | - exclude[DirectMissingMethodProblem]( |
102 | | - "com.databricks.spark.xml.util.TypeCast.supportedXmlTimestampFormatters"), |
103 | | - exclude[DirectMissingMethodProblem]( |
104 | | - "com.databricks.spark.xml.util.TypeCast.parseXmlTimestamp"), |
105 | | - exclude[DirectMissingMethodProblem]( |
106 | | - "com.databricks.spark.xml.util.TypeCast.isTimestamp"), |
107 | | - exclude[DirectMissingMethodProblem]( |
108 | | - "com.databricks.spark.xml.util.TypeCast.castTo"), |
109 | | - exclude[DirectMissingMethodProblem]( |
110 | | - "com.databricks.spark.xml.util.TypeCast.castTo$default$4") |
111 | 87 | ) |
112 | 88 | } |
0 commit comments