Skip to content

Commit

Permalink
core/filtermaps: use DeleteRange, checkpoint init of log value pointer
Browse files Browse the repository at this point in the history
  • Loading branch information
zsfelfoldi committed Oct 25, 2024
1 parent f6c08b2 commit 34a2d4d
Show file tree
Hide file tree
Showing 2 changed files with 78 additions and 34 deletions.
83 changes: 54 additions & 29 deletions core/filtermaps/filtermaps.go
Original file line number Diff line number Diff line change
Expand Up @@ -17,6 +17,7 @@
package filtermaps

import (
"bytes"
"errors"
"sync"
"time"
Expand All @@ -27,10 +28,37 @@ import (
"github.com/ethereum/go-ethereum/core/rawdb"
"github.com/ethereum/go-ethereum/core/types"
"github.com/ethereum/go-ethereum/ethdb"
"github.com/ethereum/go-ethereum/ethdb/leveldb"
"github.com/ethereum/go-ethereum/event"
"github.com/ethereum/go-ethereum/log"
)

// checkpoint allows the log indexer to start indexing from the given block
// instead of genesis at the correct absolute log value index.
type checkpoint struct {
blockNumber uint64
blockHash common.Hash
nextLvIndex uint64 // next log value index after the given block
}

var checkpoints = []checkpoint{
{ // Mainnet
blockNumber: 21019982,
blockHash: common.HexToHash("0xc684e4db692fe347e740082665acf91e27c0d9ad2a118822abdd7bb06c2a9250"),
nextLvIndex: 15878969230,
},
{ // Sepolia
blockNumber: 6939193,
blockHash: common.HexToHash("0x659b6e8a711efe8184368ac286f1f4aee74be50d38bb7fe4b24f53e73dfa58b8"),
nextLvIndex: 3392298216,
},
{ // Holesky
blockNumber: 2607449,
blockHash: common.HexToHash("0xa48c4e1ff3857ba44346bc25346d9947cd12c08f5ce8c10e8acaf40e2d6c7dc4"),
nextLvIndex: 966700355,
},
}

const headCacheSize = 8 // maximum number of recent filter maps cached in memory

// blockchain defines functions required by the FilterMaps log indexer.
Expand Down Expand Up @@ -175,6 +203,9 @@ func NewFilterMaps(db ethdb.KeyValueStore, chain blockchain, params Params, hist
log.Error("Error fetching tail block pointer, resetting log index", "error", err)
fm.filterMapsRange = filterMapsRange{} // updateLoop resets the database
}
headBlockPtr, _ := fm.getBlockLvPointer(fm.headBlockNumber)

Check failure on line 206 in core/filtermaps/filtermaps.go

View workflow job for this annotation

GitHub Actions / build

declared and not used: headBlockPtr
log.Trace("Log index head", "number", fm.headBlockNumber, "hash", fm.headBlockHash.String(), "log value pointer", fm.headLvPointer)
log.Trace("Log index tail", "number", fm.tailBlockNumber, "parentHash", fm.tailParentHash.String(), "log value pointer", fm.tailBlockLvPointer)
}
return fm
}
Expand Down Expand Up @@ -218,43 +249,37 @@ func (f *FilterMaps) removeBloomBits() {
// removeDbWithPrefix removes data with the given prefix from the database and
// returns true if everything was successfully removed.
func (f *FilterMaps) removeDbWithPrefix(prefix []byte, action string) bool {
var (
logged bool
lastLogged time.Time
removed uint64
)
it := f.db.NewIterator(prefix, nil)
hasData := it.Next()
it.Release()
if !hasData {
return true
}

end := bytes.Clone(prefix)
end[len(end)-1]++
start := time.Now()
var retry bool
for {
err := f.db.DeleteRange(prefix, end)
if err == nil {
log.Info(action+" finished", "elapsed", time.Since(start))
return true
}
if err != leveldb.ErrTooManyKeys {
log.Error(action+" failed", "error", err)
return false
}
select {
case <-f.closeCh:
return false
default:
}
it := f.db.NewIterator(prefix, nil)
batch := f.db.NewBatch()
var count int
for ; count < 250000 && it.Next(); count++ {
batch.Delete(it.Key())
removed++
}
it.Release()
if count == 0 {
break
if !retry {
log.Info(action + " in progress...")
retry = true
}
if !logged {
log.Info(action + "...")
logged = true
lastLogged = time.Now()
}
if time.Since(lastLogged) >= time.Second*10 {
log.Info(action+" in progress", "removed keys", removed)
lastLogged = time.Now()
}
batch.Write()
}
if logged {
log.Info(action + " finished")
}
return true
}

// setRange updates the covered range and also adds the changes to the given batch.
Expand Down
29 changes: 24 additions & 5 deletions core/filtermaps/indexer.go
Original file line number Diff line number Diff line change
Expand Up @@ -203,18 +203,37 @@ func (f *FilterMaps) tryInit(head *types.Header) bool {
if !f.reset() {
return false
}
head = f.chain.GetHeader(f.chain.GetCanonicalHash(0), 0)
receipts := f.chain.GetReceiptsByHash(head.Hash())

hc := newHeaderChain(f.chain, head.Number.Uint64(), head.Hash())
var (
initHeader *types.Header
nextLvPtr uint64
)
for _, cp := range checkpoints {
if initHeader == nil || cp.blockNumber >= initHeader.Number.Uint64() {
if h := f.chain.GetHeader(hc.getBlockHash(cp.blockNumber+1), cp.blockNumber+1); h != nil && h.ParentHash == cp.blockHash {
initHeader, nextLvPtr = h, cp.nextLvIndex
}
}
}
if initHeader == nil {
initHeader = f.chain.GetHeader(f.chain.GetCanonicalHash(0), 0)
}
if initHeader == nil {
log.Error("Could not retrieve init header")
return true
}
receipts := f.chain.GetReceiptsByHash(initHeader.Hash())
if receipts == nil {
log.Error("Could not retrieve block receipts for init block", "number", head.Number, "hash", head.Hash())
log.Error("Could not retrieve block receipts for init block", "number", initHeader.Number, "hash", initHeader.Hash())
return true
}
update := f.newUpdateBatch()
if err := update.initWithBlock(head, receipts, 0); err != nil {
if err := update.initWithBlock(initHeader, receipts, nextLvPtr); err != nil {
log.Error("Could not initialize log index", "error", err)
}
f.applyUpdateBatch(update)
log.Info("Initialized log index", "head", head.Number.Uint64())
log.Info("Initialized log index", "head", initHeader.Number.Uint64(), "log value pointer", nextLvPtr)
return true
}

Expand Down

0 comments on commit 34a2d4d

Please sign in to comment.