-
Notifications
You must be signed in to change notification settings - Fork 357
Spark 2.0 Branch / Support [Enhancement] #78
Comments
It would be great if you can make this backwards compatible |
To be backwards compatible we would need to retain the SparkContext as the Now if you really want a non-breaking change I can add it in a separate 2016-06-08 11:39 GMT-07:00 mriduljain notifications@github.com:
|
fyi I have created said branch and looking to see if can make this happen. update l8r today. |
Backwards compatibility is in place. Two methods are provided: using SparkSession or using SparkContext. The updates are in the same branch to simplify our discussion. Feel free to take look at the latest commit on the previously provided PR. |
Thanks will go through and comment soon On Wed, Jun 8, 2016 at 1:21 PM, StephenBoesch notifications@github.com
|
I should clarify: the backwards compatibility applies only to the consuming source code: they do not need to be changed. The SparkSession class is included and thus this change will not run against Spark 1.X. To truly achieve the backwards compatibility then we would need to add some shell scripts to manipulate the source files. I do not believe that were worth it. Instead maintain in separate Spark 2.X branch. At some point you decide to merge it into main and then the Spark 1.X becomes maintenance mode. |
A new PR has been opened that simplifies the approach and provides full backwards compatibility via a maven profile. #79
to use Spark 2.x. The actual versions of the following items are specified in that profile
|
I found a bug in #79 and am looking into it. java.lang.NoSuchMethodError: org.apache.spark.sql.SQLContext.createDataFrame(Lorg/apache/spark/rdd/RDD;Lorg/apache/spark/sql/types/StructType;)Lorg/apache/spark/sql/DataFrame; |
False alarm! The reason is: If you first try the spark 1.X via
and then try spark 2.X via
it WILL fail. We need to do clean to get all the stuff recompiled. So the following is required:
The tests are passing both under spark 1.x / scala 2.10 and spark 2.x / scala 2.11 |
I guess this pull request has been merged. Closing this. Thanks so much |
I added a PR for Spark 2.0 using the SparkSession instead of SparkContext. In addition the libraries were moved to scala 2.11 and hadoop 2.7.1 to be more in line with spark 2.X direction. The tests were run and pass.
#77
The text was updated successfully, but these errors were encountered: