cloud backend: crawlers are running too often #1886
Labels
No Label
0.2.0
0.3.0
0.4.0
0.5.0
0.5.1
0.6.0
0.6.1
0.7.0
0.8.0
0.9.0
1.0.0
1.1.0
1.10.0
1.10.1
1.10.2
1.10a2
1.11.0
1.12.0
1.12.1
1.13.0
1.14.0
1.15.0
1.15.1
1.2.0
1.3.0
1.4.1
1.5.0
1.6.0
1.6.1
1.7.0
1.7.1
1.7β
1.8.0
1.8.1
1.8.2
1.8.3
1.8β
1.9.0
1.9.0-s3branch
1.9.0a1
1.9.0a2
1.9.0b1
1.9.1
1.9.2
1.9.2a1
LeastAuthority.com automation
blocker
cannot reproduce
cloud-branch
code
code-dirnodes
code-encoding
code-frontend
code-frontend-cli
code-frontend-ftp-sftp
code-frontend-magic-folder
code-frontend-web
code-mutable
code-network
code-nodeadmin
code-peerselection
code-storage
contrib
critical
defect
dev-infrastructure
documentation
duplicate
enhancement
fixed
invalid
major
minor
n/a
normal
operational
packaging
somebody else's problem
supercritical
task
trivial
unknown
was already fixed
website
wontfix
worksforme
No Milestone
No Assignees
4 Participants
Notifications
Due Date
No due date set.
Reference: tahoe-lafs/trac-2024-07-25#1886
Loading…
Reference in New Issue
No description provided.
Delete Branch "%!s(<nil>)"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
I
flogtool tail
ed the merged cloud backend and leasedb (as of 1819-cloud-merge/022796fb), and observed that the bucket crawler and accounting crawler were running much too often. This was confirmed by looking at the storage status page at/storage
, which said that the bucket crawler and accounting crawler would run in 2 minutes and 7 minutes respectively, after they had run a few minutes ago.The bucket crawler is supposed to have a minimum cycle time of one hour, and the accounting crawler 12 hours.
This would be fixed by #1834.
Replying to zooko:
Well, not unless we removed all use of crawlers, which I don't think #1834 proposed to do.
Replying to [daira]comment:4:
IIRC, it proposes to remover all crawlers except for this one: #1835, and that one would be executed only in specific circumstances, not executing automatically over-and-over, so this ticket would almost certainly be obsolete.
Replying to [zooko]comment:5:
Oh, good point. Nevertheless I want to fix this bug before then, because I think it's shallow and could easily be fixed in an earlier release.
Hm, I wonder if this could have to do with the out-of-file-descriptors issue that my client ("WAG") is experiencing — #2342.
Hello, first time tahoe user here,
I've testing cloud storage for 3 or 4 days (branch 2237-cloud-backend-s4) into my raspberry-pi, running raspbian.
Despite it having very long startup times, I noticed .tahoe/logs/twistd.log filled up of lines like this,
that seems indicating the shares crawler is being run to often. This has the nasty side effecting that on amazon billing page
I am already with 80000 requests. Rough estimate this could mean 3 to 5 USD per month on an idle storage node alone.
This seems to be related to #1835 and #1886 - sorry for the crossposting.
2014-12-22 15:48:37+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fod%2F>
2014-12-22 15:48:37+0000 HTTPPageGetter,client Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fod%2F>
2014-12-22 15:48:37+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Foe%2F>
2014-12-22 15:48:38+0000 HTTPPageGetter,client Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Foe%2F>
2014-12-22 15:48:38+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fof%2F>
2014-12-22 15:48:38+0000 HTTPPageGetter,client Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fof%2F>
2014-12-22 15:48:38+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fog%2F>
2014-12-22 15:48:38+0000 HTTPPageGetter,client Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fog%2F>
...
Milestone renamed
renaming milestone
Moving open issues out of closed milestones.
The established line of development on the "cloud backend" branch has been abandoned. This ticket is being closed as part of a batch-ticket cleanup for "cloud backend"-related tickets.
If this is a bug, it is probably genuinely no longer relevant. The "cloud backend" branch is too large and unwieldy to ever be merged into the main line of development (particularly now that the Python 3 porting effort is significantly underway).
If this is a feature, it may be relevant to some future efforts - if they are sufficiently similar to the "cloud backend" effort - but I am still closing it because there are no immediate plans for a new development effort in such a direction.
Tickets related to the "leasedb" are included in this set because the "leasedb" code is in the "cloud backend" branch and fairly well intertwined with the "cloud backend". If there is interest in lease implementation change at some future time then that effort will essentially have to be restarted as well.