Skip to content

Commit 8fc7199

Browse files
committed
Revert "[SPARK-49700][CONNECT][SQL] Unified Scala Interface for Connect and Classic"
### Why are the changes needed? This reverts commit 86efd59 from branch-4.0. ### Does this PR introduce _any_ user-facing change? This commit causes SBT and Maven test failure. #48818 (comment) ### How was this patch tested? Pass the CIs and manual test. **BEFORE** ``` $ build/sbt clean "catalyst/testOnly org.apache.spark.sql.catalyst.encoders.EncoderResolutionSuite" ... [info] *** 1 SUITE ABORTED *** [error] Error during tests: [error] org.apache.spark.sql.catalyst.encoders.EncoderResolutionSuite [error] (catalyst / Test / testOnly) sbt.TestsFailedException: Tests unsuccessful [error] Total time: 66 s (01:06), completed Jan 27, 2025, 9:52:35 AM ``` **AFTER** ``` $ build/sbt clean "catalyst/testOnly org.apache.spark.sql.catalyst.encoders.EncoderResolutionSuite" ... [info] Run completed in 2 seconds, 236 milliseconds. [info] Total number of tests run: 28 [info] Suites: completed 1, aborted 0 [info] Tests: succeeded 28, failed 0, canceled 0, ignored 0, pending 0 [info] All tests passed. [success] Total time: 68 s (01:08), completed Jan 27, 2025, 9:51:08 AM ``` ### Was this patch authored or co-authored using generative AI tooling? No. Closes #49696 from dongjoon-hyun/SPARK-49700. Authored-by: Dongjoon Hyun <dongjoon@apache.org> Signed-off-by: Dongjoon Hyun <dongjoon@apache.org>
1 parent 808b3ac commit 8fc7199

File tree

400 files changed

+1752
-1872
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

400 files changed

+1752
-1872
lines changed

common/utils/src/main/resources/error/error-conditions.json

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -5283,11 +5283,6 @@
52835283
"Resilient Distributed Datasets (RDDs)."
52845284
]
52855285
},
5286-
"REGISTER_UDAF" : {
5287-
"message" : [
5288-
"Registering User Defined Aggregate Functions (UDAFs)."
5289-
]
5290-
},
52915286
"SESSION_BASE_RELATION_TO_DATAFRAME" : {
52925287
"message" : [
52935288
"Invoking SparkSession 'baseRelationToDataFrame'. This is server side developer API"

sql/connect/common/src/main/scala/org/apache/spark/sql/connect/DataFrameNaFunctions.scala renamed to connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/DataFrameNaFunctions.scala

Lines changed: 4 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -15,30 +15,28 @@
1515
* limitations under the License.
1616
*/
1717

18-
package org.apache.spark.sql.connect
18+
package org.apache.spark.sql
1919

2020
import scala.jdk.CollectionConverters._
2121

2222
import org.apache.spark.connect.proto.{NAReplace, Relation}
2323
import org.apache.spark.connect.proto.Expression.{Literal => GLiteral}
2424
import org.apache.spark.connect.proto.NAReplace.Replacement
25-
import org.apache.spark.sql
2625
import org.apache.spark.sql.connect.ConnectConversions._
27-
import org.apache.spark.sql.functions
2826

2927
/**
3028
* Functionality for working with missing data in `DataFrame`s.
3129
*
3230
* @since 3.4.0
3331
*/
3432
final class DataFrameNaFunctions private[sql] (sparkSession: SparkSession, root: Relation)
35-
extends sql.DataFrameNaFunctions {
33+
extends api.DataFrameNaFunctions {
3634
import sparkSession.RichColumn
3735

38-
override protected def drop(minNonNulls: Option[Int]): DataFrame =
36+
override protected def drop(minNonNulls: Option[Int]): Dataset[Row] =
3937
buildDropDataFrame(None, minNonNulls)
4038

41-
override protected def drop(minNonNulls: Option[Int], cols: Seq[String]): DataFrame =
39+
override protected def drop(minNonNulls: Option[Int], cols: Seq[String]): Dataset[Row] =
4240
buildDropDataFrame(Option(cols), minNonNulls)
4341

4442
private def buildDropDataFrame(

sql/connect/common/src/main/scala/org/apache/spark/sql/connect/DataFrameReader.scala renamed to connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/DataFrameReader.scala

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
* limitations under the License.
1616
*/
1717

18-
package org.apache.spark.sql.connect
18+
package org.apache.spark.sql
1919

2020
import java.util.Properties
2121

@@ -25,7 +25,6 @@ import org.apache.spark.annotation.Stable
2525
import org.apache.spark.api.java.JavaRDD
2626
import org.apache.spark.connect.proto.Parse.ParseFormat
2727
import org.apache.spark.rdd.RDD
28-
import org.apache.spark.sql
2928
import org.apache.spark.sql.connect.ConnectClientUnsupportedErrors
3029
import org.apache.spark.sql.connect.ConnectConversions._
3130
import org.apache.spark.sql.connect.common.DataTypeProtoConverter
@@ -38,7 +37,8 @@ import org.apache.spark.sql.types.StructType
3837
* @since 3.4.0
3938
*/
4039
@Stable
41-
class DataFrameReader private[sql] (sparkSession: SparkSession) extends sql.DataFrameReader {
40+
class DataFrameReader private[sql] (sparkSession: SparkSession) extends api.DataFrameReader {
41+
type DS[U] = Dataset[U]
4242

4343
/** @inheritdoc */
4444
override def format(source: String): this.type = super.format(source)
@@ -140,15 +140,15 @@ class DataFrameReader private[sql] (sparkSession: SparkSession) extends sql.Data
140140
override def json(paths: String*): DataFrame = super.json(paths: _*)
141141

142142
/** @inheritdoc */
143-
def json(jsonDataset: sql.Dataset[String]): DataFrame =
143+
def json(jsonDataset: Dataset[String]): DataFrame =
144144
parse(jsonDataset, ParseFormat.PARSE_FORMAT_JSON)
145145

146146
/** @inheritdoc */
147-
override def json(jsonRDD: JavaRDD[String]): DataFrame =
147+
override def json(jsonRDD: JavaRDD[String]): Dataset[Row] =
148148
throw ConnectClientUnsupportedErrors.rdd()
149149

150150
/** @inheritdoc */
151-
override def json(jsonRDD: RDD[String]): DataFrame =
151+
override def json(jsonRDD: RDD[String]): Dataset[Row] =
152152
throw ConnectClientUnsupportedErrors.rdd()
153153

154154
/** @inheritdoc */
@@ -159,7 +159,7 @@ class DataFrameReader private[sql] (sparkSession: SparkSession) extends sql.Data
159159
override def csv(paths: String*): DataFrame = super.csv(paths: _*)
160160

161161
/** @inheritdoc */
162-
def csv(csvDataset: sql.Dataset[String]): DataFrame =
162+
def csv(csvDataset: Dataset[String]): DataFrame =
163163
parse(csvDataset, ParseFormat.PARSE_FORMAT_CSV)
164164

165165
/** @inheritdoc */
@@ -170,7 +170,7 @@ class DataFrameReader private[sql] (sparkSession: SparkSession) extends sql.Data
170170
override def xml(paths: String*): DataFrame = super.xml(paths: _*)
171171

172172
/** @inheritdoc */
173-
def xml(xmlDataset: sql.Dataset[String]): DataFrame =
173+
def xml(xmlDataset: Dataset[String]): DataFrame =
174174
parse(xmlDataset, ParseFormat.PARSE_FORMAT_UNSPECIFIED)
175175

176176
/** @inheritdoc */

sql/connect/common/src/main/scala/org/apache/spark/sql/connect/DataFrameStatFunctions.scala renamed to connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/DataFrameStatFunctions.scala

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -15,16 +15,14 @@
1515
* limitations under the License.
1616
*/
1717

18-
package org.apache.spark.sql.connect
18+
package org.apache.spark.sql
1919

2020
import java.{lang => jl, util => ju}
2121

2222
import org.apache.spark.connect.proto.{Relation, StatSampleBy}
23-
import org.apache.spark.sql
24-
import org.apache.spark.sql.Column
23+
import org.apache.spark.sql.DataFrameStatFunctions.approxQuantileResultEncoder
2524
import org.apache.spark.sql.catalyst.encoders.AgnosticEncoders.{ArrayEncoder, PrimitiveDoubleEncoder}
2625
import org.apache.spark.sql.connect.ConnectConversions._
27-
import org.apache.spark.sql.connect.DataFrameStatFunctions.approxQuantileResultEncoder
2826
import org.apache.spark.sql.functions.lit
2927

3028
/**
@@ -33,7 +31,7 @@ import org.apache.spark.sql.functions.lit
3331
* @since 3.4.0
3432
*/
3533
final class DataFrameStatFunctions private[sql] (protected val df: DataFrame)
36-
extends sql.DataFrameStatFunctions {
34+
extends api.DataFrameStatFunctions {
3735
private def root: Relation = df.plan.getRoot
3836
private val sparkSession: SparkSession = df.sparkSession
3937

0 commit comments

Comments
 (0)