-
Notifications
You must be signed in to change notification settings - Fork 26
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error: 'JavaPackage' object is not callable #242
Comments
The python pip package does only contain the stubs for code completion. Spark requires the java package to be installed (the python package is not necessary on Databricks). Add a Maven Library and pass |
@liteart How do I achieve this on Databricks? Do I need to add the package at cluster level then? |
In your setup (Scala 2.12, Spark 3.4.1), this should be |
Provides a meaningful error message when user accesses a spark extension function in Python that requires the Java / Scala package: RuntimeError: Java / Scala package not found! You need to add the Maven spark-extension package to your PySpark environment: https://github.com/G-Research/spark-extension#python Before, the error was: TypeError: 'JavaPackage' object is not callable Improves #242 Supersedes #244
Description
I have two PySpark dataframes, source_df and target_df. I ran
pip install pyspark-extension
to install diff.Spark Version - 3.4.1
Scala Version - 2.12
When I run
source_df.diff(target_df)
, I get the below error -Any help would be appreciated.
The text was updated successfully, but these errors were encountered: