You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
There is a memory leak when writing SOMA sparse arrays (maybe other types). Perhaps the Arrow table being written is leaked. Test code below.
Note:
Issue only occurs in 1.13 or later
Workaround is to pin to <1.13
The leak is large enough that it prevents copying any large-ish array. For example, trying to read and rewrite a Census X layer (with a new schema), and that OOMs quickly on a 128GiB host.
I had similar issues in both tiledb-r and tiledbsoma-r which were identifiable and addressed by measuring with valgrind and making (much) more extended use of nanoarrow. In principle it is very easy to ensure each arrow allocated object has a finalizer, in practice I found not all ended up being freed. Of course, 'code being code', this leak may also be caused by something completely different but valgrind is still gold.
(I had noticed the nighlies for tiledb-r changed from '216 bytes definitely lost' to '224 bytes' but as I just verfied both are due to the same initialization within the AWS SDK we can do nothing about, and for which CRAN does not point fingers at us.)
Problem
There is a memory leak when writing SOMA sparse arrays (maybe other types). Perhaps the Arrow table being written is leaked. Test code below.
Note:
The leak is large enough that it prevents copying any large-ish array. For example, trying to read and rewrite a Census X layer (with a new schema), and that OOMs quickly on a 128GiB host.
Reported by @bkmartinjr.
[sc-53264]
Repro
Test code (will create a
test_array
in the current working directory):The text was updated successfully, but these errors were encountered: