smaller segments would facilitate better client progress indicators and better alacrity #252
Labels
No Label
0.2.0
0.3.0
0.4.0
0.5.0
0.5.1
0.6.0
0.6.1
0.7.0
0.8.0
0.9.0
1.0.0
1.1.0
1.10.0
1.10.1
1.10.2
1.10a2
1.11.0
1.12.0
1.12.1
1.13.0
1.14.0
1.15.0
1.15.1
1.2.0
1.3.0
1.4.1
1.5.0
1.6.0
1.6.1
1.7.0
1.7.1
1.7β
1.8.0
1.8.1
1.8.2
1.8.3
1.8β
1.9.0
1.9.0-s3branch
1.9.0a1
1.9.0a2
1.9.0b1
1.9.1
1.9.2
1.9.2a1
LeastAuthority.com automation
blocker
cannot reproduce
cloud-branch
code
code-dirnodes
code-encoding
code-frontend
code-frontend-cli
code-frontend-ftp-sftp
code-frontend-magic-folder
code-frontend-web
code-mutable
code-network
code-nodeadmin
code-peerselection
code-storage
contrib
critical
defect
dev-infrastructure
documentation
duplicate
enhancement
fixed
invalid
major
minor
n/a
normal
operational
packaging
somebody else's problem
supercritical
task
trivial
unknown
was already fixed
website
wontfix
worksforme
No Milestone
No Assignees
2 Participants
Notifications
Due Date
No due date set.
Reference: tahoe-lafs/trac-2024-07-25#252
Loading…
Reference in New Issue
Block a user
No description provided.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
The FireFox download progress meter, and also the upload progress meter is either absent entirely or confused, according to a report from user "cryptomail".
I suspect this is because the 1 MiB segment size is too large for FireFox's progress estimator on typical links.
Reducing it to 256 KiB or 128 KiB would increase overhead a bit but not, if I recall correctly a lot.
Someone should try that (i.e., change this constant) and see if FireFox (or other HTTP client) progress meters work better.
What was the environment like? Specifically, were they using a local tahoe node, or a remote one? If the bottleneck is between the node and the storage servers, then yeah, the browser would see nothing for a while, then suddenly a whole segment, then nothing, then a segment. But if it's using a remote node that is fairly close to the storage servers, I'd expect to see smoother downloading.
We should also verify that we're providing a Content-Length header properly. Without that, the download progress meter doesn't have a chance (and would probably revert to the "spinner" behavior, where it indicates that progress is being made but doesn't try to imply percentage completion at all).
I believe cryptomail was using http://nooxie.zooko.com:8123, which is pretty close to the storage servers.
I just now tested download of this text file:
There is a Content-Length header.
The constant in question has moved to here: source:src/allmydata/client.py@2169#L48
http://allmydata.org/trac/tahoe/browser/src/allmydata/client.py#L48
smaller segments would enable HTTP client progress indicatorsto smaller segments would facilitate better client progress indicators and better alacrityUsing 128KiB makes firefox's progress meter much smoother. Our work DSL line gives us about 1Mbps downstream, so that makes for one segment per second, which is probably a good goal.
According to '
python misc/sizes.py -m gamma
' (patched to use 3-of-10,set MAX_SEGSIZE to either 1MiB or 128KiB, account for the fact that we keep
three sets of hashes, and fix some basic confusion):
I uploaded a 2MB file (2035089 bytes, to be exact) at a 128KiB segsize, and
each of the 10 resulting shares were 682633 bytes each, so 6.83MB total.
Discounting the 3.3x expansion factor, this is a net overhead of .63%
So I'm ok with changing the default segsize for immutable files to be 128KiB.
Can we do this safely right now? I think so.. CHK keys will change (so
already-uploaded files will be uploaded again), but we hash the encoding
parameters into the CHK key, so these should not cause multiple encodings for
the same SI or anything too weird.
Mutable files need to remain at 1MB, since the current design only allows one
segment, and 128KiB would impose too small of a size limit on our directory
nodes.
I'll do this change today.
Uh oh, it appears that lowering the mag-seg-size to 128KiB also dropped our
in-colo upload and download speeds drastically. DSL speeds were unaffected,
which suggests to me that the extra round-trip times are a problem (we're not
taking full advantage of the link capacity, which is a bigger deal in-colo
because the links are so much faster).
At 1MiB MSS, speeds were mostly constant across file sizes (1MB, 10MB,
100MB). Upload varied by about 5% (e.g. 1MB at 2.01MBps, and 100MB at
1.86MBps), and download by about 4% (1MB at 4.80MBps, 10MB at 4.50MBps, and
100MB at 4.61MBps).
At 128 KiB MSS, upload speeds were mostly constant, but download speeds were
significantly faster for smaller files: 1MB ran at 3.74MBps, 10MB ran at
3.44MBps, while 100MB ran at 2.07MBps.
Zooko pointed out that pipelining our block sends could probably help this.. we'd just have to limit the maximum number of outstanding segments to keep our overall memory consumption reasonable (with 128KiB segments, allowing two to be outstanding at once has a good chance of keeping the pipeline full, while not allowing more than 128KiBexpansion2 to be in the transport buffers at any one time, which remains much smaller than the 1MiB segsize case).
The code needs to be a bit more complex, however.
I'm not sure how to prioritize this. For the benefit of end-users, I'd like the smaller segsize. For the benefit of our own servers, I'd like faster in-colo upload/download speeds. It might be possible to accomplish both, later, by adding pipelining code. But if that doesn't work, we might be stuck with the slower approach.
I guess I'll leave it at 128KiB for now and hope to implement pipelining within the next two weeks.
closing this ticket, since we're now using 128KiB segments. Created ticket #392 to track pipelining multiple segments to regain the speed lost by using 128KiB segments.
Milestone 1.0.1 deleted