-
Notifications
You must be signed in to change notification settings - Fork 280
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
download_checkmd5 ContentTooShortError #559
Comments
Some auto retry logic could be added. Wget defaults to 20 retries. http://www.gnu.org/software/wget/manual/html_node/Download-Options.html |
In order to work reliably the code would need logic to resume a previously aborted download. It could e.g. use PyCurl for this (see http://stackoverflow.com/questions/4533304/python-urlretrieve-limit-rate-and-resume-partial-download). A pull request for this would be welcome 😄 |
I believe this stackoverflow is talking about this issue and how you might implement this behavior with Python's urllib: |
@dirk-thomas OK I'm starting on it |
One comment regarding the code @wjwwood referenced: it does not check if the server actually support the range header. It should do that in order not to stitch together wrong parts of the downloaded file. |
update download script to resume downloads when server supports it (fix #559)
Hi
I'm using download_checkmd5.py to download a 200MB file from my server, and it always seems to fail with a ContentTooShortError. My server is quite reliable though: using wget I have no issues at all downloading the file.
I tried using urllib2 but it's the same.
I am now working on a solution using wget...
any idea?
Brice
The text was updated successfully, but these errors were encountered: