From b1ae5125f65e0d8a59a4006a9777ed5185a758c9 Mon Sep 17 00:00:00 2001 From: rjoshi2 Date: Sun, 21 Jan 2018 18:53:06 -0800 Subject: [PATCH] [SPARK-20129][Core] JavaSparkContext should use SparkContext.getOrCreate --- .../scala/org/apache/spark/api/java/JavaSparkContext.scala | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala b/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala index f1936bf58728..bc48a8cbea6b 100644 --- a/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala +++ b/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala @@ -50,12 +50,12 @@ class JavaSparkContext(val sc: SparkContext) * Create a JavaSparkContext that loads settings from system properties (for instance, when * launching with ./bin/spark-submit). */ - def this() = this(new SparkContext()) + def this() = this(SparkContext.getOrCreate()) /** * @param conf a [[org.apache.spark.SparkConf]] object specifying Spark parameters */ - def this(conf: SparkConf) = this(new SparkContext(conf)) + def this(conf: SparkConf) = this(SparkContext.getOrCreate(conf)) /** * @param master Cluster URL to connect to (e.g. mesos://host:port, spark://host:port, local[4]).