client should handle migrated shares by updating the write-enabler #269
Labels
No Label
0.2.0
0.3.0
0.4.0
0.5.0
0.5.1
0.6.0
0.6.1
0.7.0
0.8.0
0.9.0
1.0.0
1.1.0
1.10.0
1.10.1
1.10.2
1.10a2
1.11.0
1.12.0
1.12.1
1.13.0
1.14.0
1.15.0
1.15.1
1.2.0
1.3.0
1.4.1
1.5.0
1.6.0
1.6.1
1.7.0
1.7.1
1.7β
1.8.0
1.8.1
1.8.2
1.8.3
1.8β
1.9.0
1.9.0-s3branch
1.9.0a1
1.9.0a2
1.9.0b1
1.9.1
1.9.2
1.9.2a1
LeastAuthority.com automation
blocker
cannot reproduce
cloud-branch
code
code-dirnodes
code-encoding
code-frontend
code-frontend-cli
code-frontend-ftp-sftp
code-frontend-magic-folder
code-frontend-web
code-mutable
code-network
code-nodeadmin
code-peerselection
code-storage
contrib
critical
defect
dev-infrastructure
documentation
duplicate
enhancement
fixed
invalid
major
minor
n/a
normal
operational
packaging
somebody else's problem
supercritical
task
trivial
unknown
was already fixed
website
wontfix
worksforme
No Milestone
No Assignees
2 Participants
Notifications
Due Date
No due date set.
Reference: tahoe-lafs/trac-2024-07-25#269
Loading…
Reference in New Issue
Block a user
No description provided.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
On 12/17/07, in zooko's patch named "put all private state in
$BASEDIR/private", the location of the Tub's certificate moved from node.pem
to private/node.pem . Production nodes that were created before this change
had their cert in the old location. When I upgraded most of these nodes to
the newer code (some time after 12/17), I forgot about the change, and as a
result these nodes generated new certificate files (and thus changed their
tubid).
However, these nodes had already been running for a while, and had accepted
shares with the old tubid. As a result, the leases recorded in these shares
were made for a different nodeid, so the cancel/renew secrets are different,
and the write_enabler used for mutable slots is different.
The consequence is that certain files created before this change can no
longer be modified. They also have leases which can no longer be renewed or
canceled, but that's not as big a deal because (I think) the client nodes can
just establish new leases.
Effectively, we've accidentally migrated these shares from one storage server
to another. We made provisions for this, by recording the nodeid of the
storage server which created the slot inside the share (next to the write
enabler) and also in all the leases. The exception raised by a bad write
enabler includes this recorded nodeid (of the old server), so the client can
compute the same write enabler. Eventually, we'll write migration tools which
allow the client to demonstrate knowledge of the old write_enabler and then
ask the server to update the share to the new (correct) write-enabler.
But we haven't written these tools yet, and we're hoping to avoid it until we
actually need to start migrating shares.
I'm adding this ticket to remind me what the problem is. We can close it once
we've written and deployed the migration tools.
Note that CHK/immutable slots do not record the nodeid information, because
concern, and 2) we didn't think about it at the time. We should probably add
these identifiers to the CHK leases, and include CHK lease migration in the
tools we build. The basic idea is that if you can demonstrate knowledge of
the old lease secret, then you are allowed to change it to be whatever you
want.
it might also be a good idea to delete the current testnet shares suffering from this problem, since they're causing our lack of mutable-file recovery code to become a bigger issue than it otherwise needs to be: see #272 for details.
accidental share migrationto share migrationshare migrationto share migration (update write-enabler)The clients could respond to the wrong-write-enabler error by computing the new one and using it immediately: work around the problem rather than fixing it. This would not require any new storage-server API, and clients would continue to be able modify all their shares (it would just take an extra roundtrip).
share migration (update write-enabler)to client should handle migrated shares by updating the write-enablerThe new plan is to define a new mutable-file format (wiki/NewCapDesign) which doesn't have write-enablers (instead it has a digital signature checkable by the server), so I think we should close this as wontfix.
This means that people using the old cap format (which is everyone, currently) can't expect that a future version of Tahoe-LAFS will have the ability to fix the writer-enablers on migrated mutable shares.