From a967d4f8fd0854e9bc367e037863af232ad22440 Mon Sep 17 00:00:00 2001 From: Dongjoon Hyun Date: Fri, 15 Nov 2019 15:09:43 -0800 Subject: [PATCH 1/6] Set io.netty.tryReflectionSetAccessible for Arrow --- pom.xml | 1 + 1 file changed, 1 insertion(+) diff --git a/pom.xml b/pom.xml index 5110285547ab..21a4709aa578 100644 --- a/pom.xml +++ b/pom.xml @@ -2343,6 +2343,7 @@ test true ${project.build.directory}/tmp + true ${spark.test.home} 1 false From 80641b91b745c138affb5f58f9306e614f0c6a1c Mon Sep 17 00:00:00 2001 From: Dongjoon Hyun Date: Fri, 15 Nov 2019 17:29:59 -0800 Subject: [PATCH 2/6] Move to argLine --- pom.xml | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/pom.xml b/pom.xml index 21a4709aa578..2d14f7eec517 100644 --- a/pom.xml +++ b/pom.xml @@ -2326,7 +2326,7 @@ **/*Suite.java ${project.build.directory}/surefire-reports - -ea -Xmx4g -Xss4m -XX:ReservedCodeCacheSize=${CodeCacheSize} + -ea -Xmx4g -Xss4m -XX:ReservedCodeCacheSize=${CodeCacheSize} -Dio.netty.tryReflectionSetAccessible=true - -da -Xmx4g -XX:ReservedCodeCacheSize=${CodeCacheSize} + -da -Xmx4g -XX:ReservedCodeCacheSize=${CodeCacheSize} -Dio.netty.tryReflectionSetAccessible=true From e5079d3318580fd195e8af85229632e0b90a127f Mon Sep 17 00:00:00 2001 From: Dongjoon Hyun Date: Fri, 15 Nov 2019 21:30:26 -0800 Subject: [PATCH 5/6] fix pylint --- python/run-tests.py | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/python/run-tests.py b/python/run-tests.py index b1119b044d71..282fe5afa196 100755 --- a/python/run-tests.py +++ b/python/run-tests.py @@ -87,9 +87,10 @@ def run_individual_python_test(target_dir, test_name, pyspark_python): env["TMPDIR"] = tmp_dir # Also override the JVM's temp directory by setting driver and executor options. + java_options = "-Djava.io.tmpdir={0} -Dio.netty.tryReflectionSetAccessible=true".format(tmp_dir) spark_args = [ - "--conf", "spark.driver.extraJavaOptions=-Djava.io.tmpdir={0}".format(tmp_dir), - "--conf", "spark.executor.extraJavaOptions=-Djava.io.tmpdir={0}".format(tmp_dir), + "--conf", "spark.driver.extraJavaOptions='{0}'".format(java_options), + "--conf", "spark.executor.extraJavaOptions='{0}'".format(java_options), "pyspark-shell" ] env["PYSPARK_SUBMIT_ARGS"] = " ".join(spark_args) From 2c5112e061e1d39d1cc2b35e6f3428d7089f0d96 Mon Sep 17 00:00:00 2001 From: Dongjoon Hyun Date: Fri, 15 Nov 2019 21:51:21 -0800 Subject: [PATCH 6/6] Add R fix --- R/run-tests.sh | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/R/run-tests.sh b/R/run-tests.sh index 86bd8aad5f11..51ca7d600caf 100755 --- a/R/run-tests.sh +++ b/R/run-tests.sh @@ -23,7 +23,7 @@ FAILED=0 LOGFILE=$FWDIR/unit-tests.out rm -f $LOGFILE -SPARK_TESTING=1 NOT_CRAN=true $FWDIR/../bin/spark-submit --driver-java-options "-Dlog4j.configuration=file:$FWDIR/log4j.properties" --conf spark.hadoop.fs.defaultFS="file:///" $FWDIR/pkg/tests/run-all.R 2>&1 | tee -a $LOGFILE +SPARK_TESTING=1 NOT_CRAN=true $FWDIR/../bin/spark-submit --driver-java-options "-Dlog4j.configuration=file:$FWDIR/log4j.properties" --conf spark.hadoop.fs.defaultFS="file:///" --conf spark.driver.extraJavaOptions="-Dio.netty.tryReflectionSetAccessible=true" --conf spark.executor.extraJavaOptions="-Dio.netty.tryReflectionSetAccessible=true" $FWDIR/pkg/tests/run-all.R 2>&1 | tee -a $LOGFILE FAILED=$((PIPESTATUS[0]||$FAILED)) NUM_TEST_WARNING="$(grep -c -e 'Warnings ----------------' $LOGFILE)"