We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Sometimes it is needed to use 'magic' committer for some tables in the metastore, but to use default Spark configuration to write to other tables.
It could be helpful if special Spark configuration can be specified for individual tables in the metastore.
The configuration should be restored after the write.
Allow to configure Spark configuration for metatables.
pramen.metastore { tables = [ { name = "my_table1" format = "parquet" path = "s3://bucket1/path1" }, { name = "my_table2" format = "parquet" path = "s3a://bucket2/path2" spark.conf { spark.sql.sources.commitProtocolClass = "org.apache.spark.internal.io.cloud.PathOutputCommitProtocol" spark.sql.parquet.output.committer.class = "org.apache.spark.internal.io.cloud.BindingParquetOutputCommitter" } } ] }
The text was updated successfully, but these errors were encountered:
#447 Add Spark configuration for metastore table definitions.
0e17269
1740b66
spark.conf
a9558d9
Successfully merging a pull request may close this issue.
Background
Sometimes it is needed to use 'magic' committer for some tables in the metastore, but to use default Spark configuration to write to other tables.
It could be helpful if special Spark configuration can be specified for individual tables in the metastore.
The configuration should be restored after the write.
Feature
Allow to configure Spark configuration for metatables.
Example
Proposed Solution
The text was updated successfully, but these errors were encountered: