repair/rebalancing service #543
Labels
No Label
0.2.0
0.3.0
0.4.0
0.5.0
0.5.1
0.6.0
0.6.1
0.7.0
0.8.0
0.9.0
1.0.0
1.1.0
1.10.0
1.10.1
1.10.2
1.10a2
1.11.0
1.12.0
1.12.1
1.13.0
1.14.0
1.15.0
1.15.1
1.2.0
1.3.0
1.4.1
1.5.0
1.6.0
1.6.1
1.7.0
1.7.1
1.7β
1.8.0
1.8.1
1.8.2
1.8.3
1.8β
1.9.0
1.9.0-s3branch
1.9.0a1
1.9.0a2
1.9.0b1
1.9.1
1.9.2
1.9.2a1
LeastAuthority.com automation
blocker
cannot reproduce
cloud-branch
code
code-dirnodes
code-encoding
code-frontend
code-frontend-cli
code-frontend-ftp-sftp
code-frontend-magic-folder
code-frontend-web
code-mutable
code-network
code-nodeadmin
code-peerselection
code-storage
contrib
critical
defect
dev-infrastructure
documentation
duplicate
enhancement
fixed
invalid
major
minor
n/a
normal
operational
packaging
somebody else's problem
supercritical
task
trivial
unknown
was already fixed
website
wontfix
worksforme
No Milestone
No Assignees
3 Participants
Notifications
Due Date
No due date set.
Reference: tahoe-lafs/trac-2024-07-25#543
Loading…
Reference in New Issue
Block a user
No description provided.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
So, in doing a bunch of manual GC work over the last week, I'm starting to
think about what a "rebalancing manager" service would look like.
The basic idea is that storage servers would give a central service access to
some special facet, through which the manager could enumerate the shares
present on each one. The manager would slowly cycle through the entire
storage-index space (over the course of a month, I imagine), probably one
prefixdir at a time.
It would ask all the storage servers about which shares they hold, and figure
out which other servers also hold those shares (this query is an online
version of the 'tahoe debug catalog-shares' CLI tool). Then it would make
decisions about which shares ought to go where. There are two goals
(sometimes competing). The first is to move shares closer to the start of the
permuted peer-selection order, so that clients don't have to search as far to
find them. The second is to smooth out disk usage among all servers (more by
percentage than by absolute usage).
Once the manager works out the minimum-effort rearrangement, it will inform
the two servers that they should move a share between them. The servers can
then use a direct connection to copy the share to its new home and then
delete the original. In grids without full bidirectional connectivity, the
manager could conceivably act as a relay.
When a new (empty) disk is added to the grid, it will accumulate shares very
slowly, and only get shares for new files (those which are created after the
new node comes online). A rebalancing manager would make better use of the
new disk: filling it with old shares too, thus freeing space on old servers
so they can continue to participate in the grid (instead of being read-only).
There may be ways to perform this task without a central manager. For
example, we could treat balancing as an aspect of repair, such that the
repair process ought to include moving shares around to better places. In
this approach, the client that performs a repair would also do rebalancing.
It is not clear if the clients ought to have the same level of authority as a
trusted repair-manager: for example, should clients have the ability to
delete shares of immutable files? Making the clients drive the rebalancing
process would insure that no effort is expended on unwanted files. On the
other hand, 1) clients must then take an active interest in rebalancing, and
2) the load generated by rebalancing would be pretty choppy (a central
manager could do it smoothly, over time, whereas a client would want to
finish their repair/rebalancing pass as quickly as possible).
This will also interact with accounting. A privileged rebalancing manager
could be given the authority to clone a share (account labels and all) to a
new server, whereas a client performing rebalancing themselves would
naturally be restricted to whatever storage that client was normally allowed
to consume. I'm not sure whether this issue is significant or not.
On the implementation side, I'd expect the rebalancing-manager to be a Tahoe
node (made with 'tahoe create rebalancer', or the like), which advertises
itself via the introducer. Storage Servers would have a configuration setting
that says "give rebalancing-authority to any rebalancer that is advertised
with a signature from blesser key X". This would require each storage server
to be configured with pubkey X, but would not require any changes on the
balancer node when new storage servers are added.
It might also be a good idea to tell the rebalancer how many storage servers
it should expect to see, so it can refrain from doing anything unless it's
fully connected.
I'm also thinking that the enumerate-your-shares interface could be used to
generate estimates of how many files are in the grid. The rebalancer (or some
other node with similar enumeration authority, perhaps a stats-gatherer or
disk-watcher) could query for all shares in the aa-ab prefix range, merge the
responses from all servers, then multiply by the number of prefixes. If the
servers could efficiently distinguish mutable shares from immutable shares,
we could get estimates of both filetypes.
See also #699 (optionally rebalance during repair or upload).
The following clump of tickets might be of interest to people who are interested in this ticket: #711 (repair to different levels of M), #699 (optionally rebalance during repair or upload), #543 ('rebalancing manager'), #232 (Peer selection doesn't rebalance shares on overwrite of mutable file.), #678 (converge same file, same K, different M), #610 (upload should take better advantage of existing shares), #573 (Allow client to control which storage servers receive shares).
Also related: #778 ("shares of happiness" is the wrong measure; "servers of happiness" is better).
#778 ("shares of happiness" is the wrong measure; "servers of happiness" is better) is fixed!
If a repair operation were to also rebalance shares (as of v1.8.1 it does not), then #483 (repairer service) would be a duplicate of this ticket. So we should close #483 as a duplicate iff we decide that repair should rebalance.
(These are not to be confused with #643, which is a less ambitious ticket about scheduling the existing deep-checker/repairer using a cron job or the Windows scheduler.)
Shu Lin suggested on tahoe-dev that one criterion for rebalancing should be that a file has been recently accessed.
This could have some disadvantages for debugging, since reads would have a side effect, so I think it would have to be optional. It is probably better to add files to a rebalancing queue (and keep track of how recently they were last rebalanced) rather than repairing/rebalancing them immediately after the read, because the latter would do unnecessary work if a file is read several times.
#483 (repairer service) was closed as a duplicate.
'rebalancing manager'to repair/rebalancing service#661 was a duplicate.