Compressors are not getting recycled while writing parquet files. This is causing native/physical memory leak in my spark app which is parquet write intensive since its creating new compressors everytime i write parquet files.
The actual code issue is that we are creating 'codecFactory' in 'getRecordWriter' method of ParquetOutputFormat.java but not calling codecFactory.release() which is responsible for recycling compressors.
Reporter: Nitin Goyal / @nitin2goyal
Assignee: Nitin Goyal / @nitin2goyal
Related issues:
PRs and other links:
Note: This issue was originally created as PARQUET-353. Please see the migration documentation for further details.