Skip to content

Compressors not getting recycled while writing parquet files, causing memory leak #1869

@asfimport

Description

@asfimport

Compressors are not getting recycled while writing parquet files. This is causing native/physical memory leak in my spark app which is parquet write intensive since its creating new compressors everytime i write parquet files.

The actual code issue is that we are creating 'codecFactory' in 'getRecordWriter' method of ParquetOutputFormat.java but not calling codecFactory.release() which is responsible for recycling compressors.

Reporter: Nitin Goyal / @nitin2goyal
Assignee: Nitin Goyal / @nitin2goyal

Related issues:

PRs and other links:

Note: This issue was originally created as PARQUET-353. Please see the migration documentation for further details.

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions