You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Apr 26, 2024. It is now read-only.
When running multiple media workers it is possible to end up with both fetching and persisting the same media. #8682 fixed it such that doing so a) didn't throw an error to the client and b) deleted the duplicated media from disk, however by this point the media is already queued up to be uploaded to s3.
We could change it to only upload to storage providers after persisting to the DB, however that runs the risk of having media in the DB that hasn't been uploaded to a storage provider if e.g. the worker dies half way through upload.
I think the "fix" here is to add a delete_file function to the storage provider interface that can be called to delete a file (or cancel its upload if it hasn't finished uploading).
The text was updated successfully, but these errors were encountered:
When running multiple media workers it is possible to end up with both fetching and persisting the same media. #8682 fixed it such that doing so a) didn't throw an error to the client and b) deleted the duplicated media from disk, however by this point the media is already queued up to be uploaded to s3.
We could change it to only upload to storage providers after persisting to the DB, however that runs the risk of having media in the DB that hasn't been uploaded to a storage provider if e.g. the worker dies half way through upload.
I think the "fix" here is to add a
delete_file
function to the storage provider interface that can be called to delete a file (or cancel its upload if it hasn't finished uploading).The text was updated successfully, but these errors were encountered: