-
Notifications
You must be signed in to change notification settings - Fork 2.8k
ZEPPELIN-1267. PySparkInterpreter doesn't work in spark 2.0 #1260
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
\cc @jongyoul for review |
|
Looks good to me. |
|
I've tested it. LGTM. |
|
Let's merge if there is no further discussion! |
|
@zjffdu I tried current master branch with master set to local[*], Spark standalone and yarn-client both with/without |
|
I also use the latest master and build zeppelin with this command: |
|
@zjffdu thank you for quick response, I was able to reproduce it and tested this patch fixes the issue! |
### What is this PR for? PySparkInterpreter doesn't work in spark 2.0 because pyspark and py4j is not distributed to executors. This PR extract the setup staff for pyspark interpreter into method setupConfForPySpark and use it for both spark1 and spark2. But this is just a short term solution, as I think this should be handled by spark rather than zeppelin, here zeppelin duplicate part of spark's work. In the long term, I'd like to resolve it in `ZEPPELIN-1263`. ### What type of PR is it? [Bug Fix] ### Todos * https://issues.apache.org/jira/browse/ZEPPELIN-1263 ### What is the Jira issue? * https://issues.apache.org/jira/browse/ZEPPELIN-1267 ### How should this be tested? Verify it manually. ### Screenshots (if appropriate)  ### Questions: * Does the licenses files need update? No * Is there breaking changes for older versions? No * Does this needs documentation? No Author: Jeff Zhang <zjffdu@apache.org> Closes #1260 from zjffdu/ZEPPELIN-1267 and squashes the following commits: 81d1d56 [Jeff Zhang] ZEPPELIN-1267. PySparkInterpreter doesn't work in spark 2.0 (cherry picked from commit 161dd0e) Signed-off-by: Mina Lee <minalee@apache.org>
### What is this PR for? PySparkInterpreter doesn't work in spark 2.0 because pyspark and py4j is not distributed to executors. This PR extract the setup staff for pyspark interpreter into method setupConfForPySpark and use it for both spark1 and spark2. But this is just a short term solution, as I think this should be handled by spark rather than zeppelin, here zeppelin duplicate part of spark's work. In the long term, I'd like to resolve it in `ZEPPELIN-1263`. ### What type of PR is it? [Bug Fix] ### Todos * https://issues.apache.org/jira/browse/ZEPPELIN-1263 ### What is the Jira issue? * https://issues.apache.org/jira/browse/ZEPPELIN-1267 ### How should this be tested? Verify it manually. ### Screenshots (if appropriate)  ### Questions: * Does the licenses files need update? No * Is there breaking changes for older versions? No * Does this needs documentation? No Author: Jeff Zhang <zjffdu@apache.org> Closes apache#1260 from zjffdu/ZEPPELIN-1267 and squashes the following commits: 81d1d56 [Jeff Zhang] ZEPPELIN-1267. PySparkInterpreter doesn't work in spark 2.0
What is this PR for?
PySparkInterpreter doesn't work in spark 2.0 because pyspark and py4j is not distributed to executors. This PR extract the setup staff for pyspark interpreter into method setupConfForPySpark and use it for both spark1 and spark2. But this is just a short term solution, as I think this should be handled by spark rather than zeppelin, here zeppelin duplicate part of spark's work. In the long term, I'd like to resolve it in
ZEPPELIN-1263.What type of PR is it?
[Bug Fix]
Todos
What is the Jira issue?
How should this be tested?
Verify it manually.
Screenshots (if appropriate)
Questions: