You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Using this extension in a notebook that also uses pySpark leads to continuous Spark queries. By default, Spark is a lazy-evaluation system, and only runs queries when there is an output operation on a dataframe. With the extension loaded, however, there are Spark queries running continuously.
My guess is the extension appears to try to continuously convert/show any variables that are Spark dataframes. This is a special problem with larger dataframes, as it keeps the Spark instance continuously busy.
Using Jupyter Lab version 3 and lckr-jupyterlab-variableinspector 3.0.9
One mitigation may be to have a setting/option to skip variables that point to Spark DFs.
The text was updated successfully, but these errors were encountered:
Using this extension in a notebook that also uses pySpark leads to continuous Spark queries. By default, Spark is a lazy-evaluation system, and only runs queries when there is an output operation on a dataframe. With the extension loaded, however, there are Spark queries running continuously.
My guess is the extension appears to try to continuously convert/show any variables that are Spark dataframes. This is a special problem with larger dataframes, as it keeps the Spark instance continuously busy.
Using Jupyter Lab version 3 and lckr-jupyterlab-variableinspector 3.0.9
One mitigation may be to have a setting/option to skip variables that point to Spark DFs.
The text was updated successfully, but these errors were encountered: