diff --git a/connector/connect/src/main/scala/org/apache/spark/sql/connect/SparkConnectPlugin.scala b/connector/connect/src/main/scala/org/apache/spark/sql/connect/SparkConnectPlugin.scala index d262947015cb1..7ac33fa9324ac 100644 --- a/connector/connect/src/main/scala/org/apache/spark/sql/connect/SparkConnectPlugin.scala +++ b/connector/connect/src/main/scala/org/apache/spark/sql/connect/SparkConnectPlugin.scala @@ -44,8 +44,8 @@ class SparkConnectPlugin extends SparkPlugin { override def driverPlugin(): DriverPlugin = new DriverPlugin { override def init( - sc: SparkContext, - pluginContext: PluginContext): util.Map[String, String] = { + sc: SparkContext, + pluginContext: PluginContext): util.Map[String, String] = { SparkConnectService.start() Map.empty[String, String].asJava } diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala index 6ffc1d880c5d1..1de8bc6a47ded 100644 --- a/project/SparkBuild.scala +++ b/project/SparkBuild.scala @@ -624,17 +624,17 @@ object SparkConnect { // For some reason the resolution from the imported Maven build does not work for some // of these dependendencies that we need to shade later on. libraryDependencies ++= Seq( - "io.grpc" % "protoc-gen-grpc-java" % BuildCommons.gprcVersion asProtocPlugin(), - "org.scala-lang" % "scala-library" % "2.12.16" % "provided", - "com.google.guava" % "guava" % "31.0.1-jre", - "com.google.guava" % "failureaccess" % "1.0.1", - "com.google.protobuf" % "protobuf-java" % protoVersion % "protobuf" + "io.grpc" % "protoc-gen-grpc-java" % BuildCommons.gprcVersion asProtocPlugin(), + "org.scala-lang" % "scala-library" % "2.12.16" % "provided", + "com.google.guava" % "guava" % "31.0.1-jre", + "com.google.guava" % "failureaccess" % "1.0.1", + "com.google.protobuf" % "protobuf-java" % protoVersion % "protobuf" ), dependencyOverrides ++= Seq( - "com.google.guava" % "guava" % "31.0.1-jre", - "com.google.guava" % "failureaccess" % "1.0.1", - "com.google.protobuf" % "protobuf-java" % protoVersion + "com.google.guava" % "guava" % "31.0.1-jre", + "com.google.guava" % "failureaccess" % "1.0.1", + "com.google.protobuf" % "protobuf-java" % protoVersion ), (Compile / PB.targets) := Seq( @@ -642,7 +642,7 @@ object SparkConnect { PB.gens.plugin("grpc-java") -> (Compile / sourceManaged).value ), - (assembly / test) := false, + (assembly / test) := { }, (assembly / logLevel) := Level.Info, @@ -1195,7 +1195,7 @@ object CopyDependencies { // For the SparkConnect build, we manually call the assembly target to // produce the shaded Jar which happens automatically in the case of Maven. // Later, when the dependencies are copied, we manually copy the shaded Jar only. - val fid = (LocalProject("connect")/assembly).value + val fid = (LocalProject("connect") / assembly).value (Compile / dependencyClasspath).value.map(_.data) .filter { jar => jar.isFile() } diff --git a/python/pyspark/sql/connect/__init__.py b/python/pyspark/sql/connect/__init__.py index c748f8f6590e7..3df96963f929c 100644 --- a/python/pyspark/sql/connect/__init__.py +++ b/python/pyspark/sql/connect/__init__.py @@ -19,4 +19,4 @@ Spark through this API are can be changed at any time without warning.""" -from pyspark.sql.connect.data_frame import DataFrame # noqa: F401 +from pyspark.sql.connect.dataframe import DataFrame # noqa: F401 diff --git a/python/pyspark/sql/connect/client.py b/python/pyspark/sql/connect/client.py index 3aa335d35c051..cf825f88e857e 100644 --- a/python/pyspark/sql/connect/client.py +++ b/python/pyspark/sql/connect/client.py @@ -30,7 +30,7 @@ import pyspark.sql.connect.proto.base_pb2_grpc as grpc_lib import pyspark.sql.types from pyspark import cloudpickle -from pyspark.sql.connect.data_frame import DataFrame +from pyspark.sql.connect.dataframe import DataFrame from pyspark.sql.connect.readwriter import DataFrameReader from pyspark.sql.connect.plan import SQL diff --git a/python/pyspark/sql/connect/data_frame.py b/python/pyspark/sql/connect/dataframe.py similarity index 100% rename from python/pyspark/sql/connect/data_frame.py rename to python/pyspark/sql/connect/dataframe.py diff --git a/python/pyspark/sql/connect/readwriter.py b/python/pyspark/sql/connect/readwriter.py index fb1e6da36f58f..285e78e59ae96 100644 --- a/python/pyspark/sql/connect/readwriter.py +++ b/python/pyspark/sql/connect/readwriter.py @@ -15,7 +15,7 @@ # limitations under the License. # -from pyspark.sql.connect.data_frame import DataFrame +from pyspark.sql.connect.dataframe import DataFrame from pyspark.sql.connect.plan import Read from typing import TYPE_CHECKING