You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Aug 13, 2019. It is now read-only.
is going to extract every single record from the Kinto database into a monster list of dicts.
I measured it, with memory_profiler in my docker with a postgres database of about 500,000 records. Generating that new_records dict eats up about 1700MB.
See https://irccloud.mozilla.com/pastebin/ZhAoBDia/
The solution has to be to consume that paginated result as a stream and for each patch of 10,000 records make the hashes and then reuse that allocated memory for the next 10,000 records.
@leplatrem I don't see a whole lot in the documentation about to do the pagination i any other way. Do you have some ideas?
The text was updated successfully, but these errors were encountered:
Suppose you delete the
.records-hashes-$SERVER_NAME.json
file. That means this code right here:buildhub/jobs/buildhub/to_kinto.py
Lines 158 to 161 in ee484fa
I measured it, with
memory_profiler
in my docker with a postgres database of about 500,000 records. Generating thatnew_records
dict eats up about 1700MB.See https://irccloud.mozilla.com/pastebin/ZhAoBDia/
The solution has to be to consume that paginated result as a stream and for each patch of 10,000 records make the hashes and then reuse that allocated memory for the next 10,000 records.
@leplatrem I don't see a whole lot in the documentation about to do the pagination i any other way. Do you have some ideas?
The text was updated successfully, but these errors were encountered: