Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Shorter availability data retention period for testnets #3270

Closed
sandreim opened this issue Feb 8, 2024 · 5 comments · Fixed by #7353
Closed

Shorter availability data retention period for testnets #3270

sandreim opened this issue Feb 8, 2024 · 5 comments · Fixed by #7353
Assignees
Labels
T0-node This PR/Issue is related to the topic “node”. T8-polkadot This PR/Issue is related to/affects the Polkadot network.

Comments

@sandreim
Copy link
Contributor

sandreim commented Feb 8, 2024

Currently we keep the finalized data for 25h. This raises testnet disk space requirements needlessly, especially if huge PoVs are being used.

The way this parameter is passed should have some checks to prevent misuse, especially in non-test chains where we actually do need 25h retention.

@sandreim sandreim added T0-node This PR/Issue is related to the topic “node”. T8-polkadot This PR/Issue is related to/affects the Polkadot network. labels Feb 8, 2024
@s0me0ne-unkn0wn s0me0ne-unkn0wn self-assigned this Feb 8, 2024
@eskimor
Copy link
Member

eskimor commented Mar 18, 2024

Also for production we should consider only preserving the last 10 blocks or something for each parachain.

@pepyakin
Copy link
Contributor

For production maybe it's worth to err on the safer side though. A parachain could be frozen in case a collator manages to withold the data from other parachain nodes. Therefore 60s sounds a bit too cowboy.

@burdges
Copy link

burdges commented Mar 18, 2024

Also for production we should consider only preserving the last 10 blocks or something for each parachain.

If we tune down availability, then some validators will tune down their storage. This matters little for operator's physical nodes, since drives are cheap, but matters lots for cloud validators, since cloud providers charge lots for the very reliable storage they provide. I'm almost always pro-efficency, but in this case maybe better to artificially waste cloud valdiators money since we're not big fans of them. Also, cores are underutilized, so we can tune down this paramater quickly later if we've a burst of usage.

Also, we'll hopefully soon have other availability data which last longer than a few min, like off-chain messages, so we either block that future work on adding multiple availability segments with different retentions, or else we keep some longer reteention period now, and then do multiple availability segments as a later storage optimization.

@eskimor
Copy link
Member

eskimor commented Mar 18, 2024

For production maybe it's worth to err on the safer side though. A parachain could be frozen in case a collator manages to withold the data from other parachain nodes. Therefore 60s sounds a bit too cowboy.

yeah. I said "or something" does not need to be 10, even 100 would be much less data than what we have now. Either way 10 blocks, should mean 10 different collators, which all would need to collude. I would certainly recommend some proper reasoning before going for a particular number ;-)

@burdges
Copy link

burdges commented Mar 19, 2024

We should definitely shortten testnet availability duration anyways. We reduced the validator count & soundness for the community testnet, so they'll be hammered pretty seriously if a bunch of parachains show up. 10 min is definitely reasonable there. I guess 1 min works too, but maybe it'll need bumping for XCMP.

We could discuss the number of sessions for production, but two full sessions would still be less than 24h. We could also reserve a fixed amount of space and recycle old stuff when the space is needed, but this could create strange bugs when people expect it to be longer.

As for duration arguments, we should price storage using commercial drive prices, but not anything fancier like RAID arrays or cloud storage.

github-merge-queue bot pushed a commit that referenced this issue Feb 13, 2025
Closes #3270

---------

Co-authored-by: command-bot <>
@github-project-automation github-project-automation bot moved this from Backlog to Completed in parachains team board Feb 13, 2025
clangenb pushed a commit to clangenb/polkadot-sdk that referenced this issue Feb 19, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
T0-node This PR/Issue is related to the topic “node”. T8-polkadot This PR/Issue is related to/affects the Polkadot network.
Projects
Status: Completed
Development

Successfully merging a pull request may close this issue.

5 participants