-
Notifications
You must be signed in to change notification settings - Fork 267
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Garbled data when reading large cdf-5 files #453
Comments
This looks vaguely familiar; thanks for the bug report, we definitely want to correct this. Does this issue work if the file is generate serially? Or is that impractical? Perhaps it is impractical and it shouldn't matter, but I need to start narrowing things down. The fact that ncmpidump works is reassuring; the issue is in ncdump reading the data, and not how netcdf writes the data (a worst case scenario). Is it possible to put this large file somewhere where I can download it? I will start trying to diagnose this. Oh yes, I said this looked vaguely familiar. I just released the 4.5.0-rc2 yesterday, and it contains a lot of fixes which have not made its way back into Thanks! |
4.5.0-rc2 has the same problem. I'll upload the file and will post a link to it here later today. |
I fixed it myself, was an easy one. A check for NC_64BIT_DATA was missing in v1hpg.c. Just issued pull request #457 that should fix this. |
Thank you, that is a fantastic help, I will get it worked into the 4.5.0 release branch. If possible, can we still get the file? Or perhaps it would be easy to capture the output form 'ncdump -h [filename]' and attach it here? I'd like to characterize the file so that I can add a test to ensure this doesn't regress in the future. |
Let me try to generate a file with a minimal size that shows this error, will get that one to you. |
Thank you @pastewka ; I appreciate that, and certainly take your time. |
You can download the example file here. File size is 20GB. |
Given the fix, it should be possible to reengineer a |
Could write a cdf-5 file of appropriate size and then read. That should break. Is cdf-5 included in large file tests and if so, what is largest record size that is written?
…________________________________
Von: DennisHeimbigner <notifications@github.com>
Gesendet: 12.08.2017 12:22 vorm.
An: Unidata/netcdf-c
Cc: Lars Pastewka; Mention
Betreff: Re: [Unidata/netcdf-c] Garbled data when reading large cdf-5 files (#453)
|
The fix has been merged into master, thank you! |
Perfect, thank you! |
Environment Information
configure
)CFLAGS: -D_LARGEFILE64_SOURCE -D_FILE_OFFSET_BITS=64 -g -Wall -Wconversion
C
code to recreate the issue?Summary of Issue
I am having a problem reading cdf-5 files that were written with parallel NetCDF 1.8.1 on an IBM BlueGene/Q. All unlimited frames except the first one contain garbled data. I believe this problem only appears when a single frame of an unlimited dimension becomes large. (Probably >2GB but have not systematically tested this.)
The example file under consideration has a size of 82G and 4 unlimited frames. Output of
ncdump -h
follows:Inspection for example the
cell_angles
variable withncdump -v cell_angles
yields:The dump command from parallel NetCDF is able to read this file.
ncmpidump -v cell_angles
yields:which is the correct information.
The error occurs in NetCDF 4.4.1.1 and latest github master.
The text was updated successfully, but these errors were encountered: