diff --git a/docs/howto/export.md b/docs/howto/export.md index afad4d6afd8..c0b9fcf1e08 100644 --- a/docs/howto/export.md +++ b/docs/howto/export.md @@ -55,7 +55,7 @@ The complete `spark-submit` command would look as follows: spark-submit --conf spark.hadoop.lakefs.api.url=https:///api/v1 \ --conf spark.hadoop.lakefs.api.access_key= \ --conf spark.hadoop.lakefs.api.secret_key= \ - --packages io.lakefs:lakefs-spark-client_2.12:0.10.0 \ + --packages io.lakefs:lakefs-spark-client_2.12:0.11.0 \ --class io.treeverse.clients.Main export-app example-repo s3://example-bucket/prefix \ --branch=example-branch ``` diff --git a/docs/howto/garbage-collection/gc.md b/docs/howto/garbage-collection/gc.md index 8cbd8e0a424..11d3a7e4a2c 100644 --- a/docs/howto/garbage-collection/gc.md +++ b/docs/howto/garbage-collection/gc.md @@ -121,7 +121,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \ -c spark.hadoop.lakefs.api.secret_key= \ -c spark.hadoop.fs.s3a.access.key= \ -c spark.hadoop.fs.s3a.secret.key= \ - http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.10.0/lakefs-spark-client-assembly-0.10.0.jar \ + http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.11.0/lakefs-spark-client-assembly-0.11.0.jar \ example-repo us-east-1 ``` @@ -136,7 +136,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \ -c spark.hadoop.lakefs.api.access_key= \ -c spark.hadoop.lakefs.api.secret_key= \ -c spark.hadoop.fs.azure.account.key..dfs.core.windows.net= \ - http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.10.0/lakefs-spark-client-assembly-0.10.0.jar \ + http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.11.0/lakefs-spark-client-assembly-0.11.0.jar \ example-repo ``` @@ -153,7 +153,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \ -c spark.hadoop.fs.azure.account.oauth2.client.id..dfs.core.windows.net= \ -c spark.hadoop.fs.azure.account.oauth2.client.secret..dfs.core.windows.net= \ -c spark.hadoop.fs.azure.account.oauth2.client.endpoint..dfs.core.windows.net=https://login.microsoftonline.com//oauth2/token \ - http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.10.0/lakefs-spark-client-assembly-0.10.0.jar \ + http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.11.0/lakefs-spark-client-assembly-0.11.0.jar \ example-repo ``` @@ -181,7 +181,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \ -c spark.hadoop.fs.gs.impl=com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystem \ -c spark.hadoop.fs.AbstractFileSystem.gs.impl=com.google.cloud.hadoop.fs.gcs.GoogleHadoopFS \ -c spark.hadoop.lakefs.gc.do_sweep=false \ - http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.10.0/lakefs-spark-client-assembly-0.10.0.jar \ + http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.11.0/lakefs-spark-client-assembly-0.11.0.jar \ example-repo ``` diff --git a/docs/reference/spark-client.md b/docs/reference/spark-client.md index 29333affd45..418260d1ca0 100644 --- a/docs/reference/spark-client.md +++ b/docs/reference/spark-client.md @@ -23,13 +23,13 @@ Start Spark Shell / PySpark with the `--packages` flag: This client is compiled for Spark 3.1.2 with Hadoop 3.2.1, but can work for other Spark versions and higher Hadoop versions. - + ```bash -spark-shell --packages io.lakefs:lakefs-spark-client_2.12:0.10.0 +spark-shell --packages io.lakefs:lakefs-spark-client_2.12:0.11.0 ``` Alternatively an assembled jar is available on S3, at -`s3://treeverse-clients-us-east/lakefs-spark-client/0.10.0/lakefs-spark-client-assembly-0.10.0.jar` +`s3://treeverse-clients-us-east/lakefs-spark-client/0.11.0/lakefs-spark-client-assembly-0.11.0.jar` ## Configuration