-
-
Notifications
You must be signed in to change notification settings - Fork 4.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Chunked upload performance improvements #47682
Comments
Good stuff. A couple thoughts:
Footnotes
|
@joshtrichards thanks for your input! Highly appreciated having pointers to related topics :) |
Hi, @nextcloud/desktop would it be possible to sit down with some of you at the Conference in two weeks and discuss this? I think we can figure out something together, solve the problems and unify the behavior across clients and the web. |
Also @joshtrichards (and anyone else) feel free to join in case this takes place. You seem to know quite a lot of paint points of this problem which would be really helpful to consider. |
@provokateurin Matthieu and me will be at the conference, could you put something on the calendar for this? |
I'll try to find a time slot (the calendar is really full huh) and get it added to the official agenda so everyone interested can join. To avoid collisions with any other event it probably has to be late afternoon. |
Ok, the meeting will take place Wednesday 18.9. 16:00 at Motion Lab. The room is still to be determined, but will be available in the public agenda and calendar on time. |
I wish I could join, but I'll not in be town for the conference. |
I just realized I might have made a mistake. Instead of the 10 workers I think only 4 were available. Given that the chunked upload tries to upload 5 concurrent chunks at least one of the might have been stalled the whole time. |
During my analysis of #47856, I have found some interesting points regarding file upload optimization in Nextcloud: One problem in case of local storage is that Another problem is in |
@cfiehe thanks for your insights! We will consider them in the discussion, although this issue itself is more about the chunked upload performance itself. I only noticed similar issues and noted them down here so they can be further investigated in the future. |
Hi everyone,
Items 1, 2, 3 and 4 can all be implemented indepently of another, while 1 and 3 and very easy, 2 needs quite some work on the backend and client sides and 4 needs more research/talking to experts to figure out if option 4.ii is possible. Thank you again to all the people who attended, the outcome is way better than what I had hoped for. |
I implemented this, but it actually makes no difference, because the chunk is read into memory anyway. It even seemed to slow down the process a few ms, maybe because it no longer used streams but the entire data directly. |
One additional limitation there is on the PHP side (ref #43636 (comment)). I could imagine some performance gain if we'd use a POST for upload of chunks where PHP exposes the temporary file directly, but doesn't seem to be possible with the current PUT approach for chunks. In the current state there is always a file written by PHP itself which we then read into memory through |
@juliushaertl thanks for the insight. I will keep it in mind for point 2, then it can be even faster 👍 |
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
For point 2: Thanks @Leptopoda for pointing me to this RFC :) |
Actually now that I think about it, we can probably get rid of the assembly step completely, which makes point 4 irrelevant (except for older chunked upload implementations). With a single upload connection we can just append the already uploaded data. Having the chunks separate and then merging them is only necessary due to parallel connections. I need to think a bit more about it, but I hope it is possible to save a lot of work by just taking the simple route which is also way more performant and less resource intensive. |
Implementation of the draft RFC is here: #49753 |
Hi I would like to know if this would possibly also address the following issue I've been having: https://help.nextcloud.com/t/long-running-request-to-move-file-after-large-file-upload-times-out-causing-del-to-be-called/212394 To summarize the post from the community forum: A long-running request to Would the changes proposed in the PR have a positive impact in any way in this situation? Thanks. |
Yes the PR above will fix that issue. The problem you are describing is unfortunately not fixable with the current architecture of chunked upload. The only way to work around it is to increase timeouts if possible (but that can have other downsides as well). |
Thanks for your reply. I am glad to hear that would solve the issue. That's been a blocker for me to move myself and me entire family to NC, but knowing there's a solution to be released eventually is nice. In order to speed up things, I can offer my help. I'm not experienced with PHP at all so I couldn't help there but I'm very experienced with a range of frontend tech and some open source contributions as all, if needed at all I could try to provide some contribution. |
Thanks for the offer to help, but currently the problem is that the RFC is still in draft and we want to wait for it to be finalized. Hopefully that is going to happen within the next few months and the feature will land in 31 or 32 later next year. |
This comment has been minimized.
This comment has been minimized.
@brunofin is experiencing a different problem which can not be fixed by changing chunk sizes. In this case the assembly of the chunks can never finish because the timeouts always end the request early. Let's stay on topic please and use other issues or the forum to discuss existing problems that are not directly related to this issue. |
Motivation
Chunked upload is useful in many cases, but also slower than uploading the entire file directly.
Overall it is slower because multiple network requests have to be made.
The speed of requests varies during their lifetime, as it is slower at the beginning and then flattens out at the maximum over time.
The smaller the requests are and the more we make, the worse performance penalty.
Thus to increase chunked upload speed the size of chunks should be increased.
A single upload using cURL is the upper limit of the possible upload speed for any configuration and upload method.
A chunked upload with the chunk size equal to or greater than the file size represents the upper limit for chunked uploads as it only uploads a single chunk.
While reaching the former would be nice, only the latter is achievable (without general performance improvements in WebDAV regardless of the maximum chunk size) and thus represents the theoretical goal.
Testing methodology
Input
Scenarios
All tests are running on a local instance using the PHP standalone web server with 10 workers and no extra apps enabled.
The machine has a Ryzen 5 5800X (8 threads, 16 cores), 48GB RAM and a NVMe M.2 Samsung 980 SSD 1TB.
Hardware should not be a bottleneck on this setup and external networking can not have an effect either.
1. cURL single upload
Take the
Real
timing value.Runs:
5.412s
5.223s
5.100s
Average: 5.245s
Note: I once saw an outlier that only took about 4.7s, but this never happened again.
Chunked upload via browser
Open Firefox Devtools and filter network requests by
dav/uploads
.Upload
1G.bin
via web interface.Take
Started
of first (MKCOL
) and theDownloaded
of the last (MOVE
) request and subtract them (See theTimings
tab of each request).This includes some constant overhead for the
MKCOL
andMOVE
requests which is not relevant for comparing chunked upload timing results as they all have the same overhead, but when comparing to the cURL scenario it accurately measures the overall time for the upload process.According to https://firefox-source-docs.mozilla.org/devtools-user/network_monitor/throttling/index.html "Wi-Fi" throttling means a maximum speed of 15 Mbps.
Sadly this is the "fastest" speed one can select for throttling and there is no way to set a custom speed.
It should represent a worst-case, while most uploads are probably done with 2-3x that speed in the real world if the Nextcloud instance is not on the same network.
Adjusting the default maximum chunk size can be done in
server/apps/files/lib/App.php
Line 45 in 7964058
2. Chunk size 10MiB (current default), unlimited bandwidth
Chunks: 103
Runs:
47.16s
47.65s
47.33s
Average: 47.38s
3. Chunk size 100MiB, unlimited bandwidth
Chunks: 11
Runs:
8.53s
8.64s
8.63s
Average: 8.6s
4. Chunk size 1024MiB, unlimited bandwidth
Chunks: 1
Runs:
6.37s
6.34s
6.34s
Average: 6.35s
5. Chunk size 10MiB (current default), throttled "Wi-Fi"
Chunks: 103
Runs:
551.40s
551.40s
551.40s
Average: 551.40s
6. Chunk size 100MiB, throttled "Wi-Fi"
Chunks: 11
Runs:
552.60s
549.60s
551.40s
Average: 551.2s
7. Chunk size 1024MiB, throttled "Wi-Fi"
Chunks: 1
Runs:
568.20s
555.60s
553.11s
Average: 558.97s
Conclusions
Upload speed in Nextcloud is very consistent regardless of upload method. Great!
Chunked upload in general takes about 21% longer in scenarios with unlimited bandwidth (scenario 1 and 4). Whether this overhead can be eliminated easily is not clear, but at least there is no hard limitation since both uploads are done through WebDAV and thus use the same underlying stack (also see other interesting findings section below).
In the current default configuration with unlimited bandwidth chunked upload is takes 646% longer than the maximum speed (scenario 2 and 4). By increasing the chunk size by 10x keeping the bandwidth ulimited it only takes 35% longer than the maximum speed (scenario 3 and 4). This is a 5.5x increase in total throughput (scenario 2 and 3).
In bandwidth limited scenarios increasing the chunk size has almost no positive (and no negative effect; scenario 5 and 6). This is expected as the slow speed at the beginning of each chunk is a lot smaller on relation to the overall speed or even exactly the same.
Increasing the chunk size helps uploads on fast connections while it has no downsides on slow connections speed wise. Slow networks can be correlated with unstable networks, so having fewer and larger chunks could result in a higher rate of aborted chunk uploads. This downside should be taken into consideration when choosing a new maximum chunk size.
A new maximum chunk size still needs to be figured out by collecting more data for different chunk sizes. It needs to hit a sweet spot of maximum speed with minimum size to account for the before mentioned drawback on unstable networks (basically the point of diminishing returns). This investigation was only to prove that we can increase the chunked upload speed.
Other interesting findings
While uploading with a single chunk and unlimited bandwidth Firefox displayed that the request needed 637ms to send but had to wait 2.10s after that (reproducible). This might show that we have a bottleneck in processing the uploads on the backend side. Maybe it would be possible to stream the request data directly into the file while should cut down the waiting time a lot. It should be possible to profile these requests and figure out where the time is spent.
For single chunks the
MOVE
request still takes quite some time. I assume this happens because it concatenates the chunks while there is only one (which is slow because it has to read and write all the data). This case could be detected and the file moved without reading and writing it (which is not possible for all storages AFAICT, i.e. it needs to be on the same physical disk to take advantage of it). This only affects uploads where the file size is less than the maximum chunk size. Due to the current low limit it is not really noticable, but with a higher maximum chunk size this would affect a lot more and bigger uploads and could lead to quite a performance improvement for those.Upload ETA calculations are all over the place due to the varying speed of uploading chunks over their lifetime. It could be improved by taking the overall time a chunk needs to upload and multiplying it by number of remaining chunks. This should be a lot more reliable as every chunk should have a similar upload characteristics. To smooth out the the value it could take into account the last 3-5 chunks.
The text was updated successfully, but these errors were encountered: