mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2024-12-23 11:57:18 +00:00
0625a6906c
* add better pruning * add test * gaz * victor's review Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> Co-authored-by: prylabs-bulldozer[bot] <58059840+prylabs-bulldozer[bot]@users.noreply.github.com>
155 lines
4.5 KiB
Go
155 lines
4.5 KiB
Go
package cache
|
|
|
|
import (
|
|
"sync"
|
|
"time"
|
|
|
|
lru "github.com/hashicorp/golang-lru"
|
|
"github.com/patrickmn/go-cache"
|
|
types "github.com/prysmaticlabs/eth2-types"
|
|
"github.com/prysmaticlabs/prysm/shared/params"
|
|
"github.com/prysmaticlabs/prysm/shared/sliceutil"
|
|
)
|
|
|
|
type subnetIDs struct {
|
|
attester *lru.Cache
|
|
attesterLock sync.RWMutex
|
|
aggregator *lru.Cache
|
|
aggregatorLock sync.RWMutex
|
|
persistentSubnets *cache.Cache
|
|
subnetsLock sync.RWMutex
|
|
}
|
|
|
|
// SubnetIDs for attester and aggregator.
|
|
var SubnetIDs = newSubnetIDs()
|
|
|
|
func newSubnetIDs() *subnetIDs {
|
|
// Given a node can calculate committee assignments of current epoch and next epoch.
|
|
// Max size is set to 2 epoch length.
|
|
cacheSize := int(params.BeaconConfig().SlotsPerEpoch.Mul(params.BeaconConfig().MaxCommitteesPerSlot * 2))
|
|
attesterCache, err := lru.New(cacheSize)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
aggregatorCache, err := lru.New(cacheSize)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
epochDuration := time.Duration(params.BeaconConfig().SlotsPerEpoch.Mul(params.BeaconConfig().SecondsPerSlot))
|
|
subLength := epochDuration * time.Duration(params.BeaconConfig().EpochsPerRandomSubnetSubscription)
|
|
persistentCache := cache.New(subLength*time.Second, epochDuration*time.Second)
|
|
return &subnetIDs{attester: attesterCache, aggregator: aggregatorCache, persistentSubnets: persistentCache}
|
|
}
|
|
|
|
// AddAttesterSubnetID adds the subnet index for subscribing subnet for the attester of a given slot.
|
|
func (s *subnetIDs) AddAttesterSubnetID(slot types.Slot, subnetID uint64) {
|
|
s.attesterLock.Lock()
|
|
defer s.attesterLock.Unlock()
|
|
|
|
ids := []uint64{subnetID}
|
|
val, exists := s.attester.Get(slot)
|
|
if exists {
|
|
ids = sliceutil.UnionUint64(append(val.([]uint64), ids...))
|
|
}
|
|
s.attester.Add(slot, ids)
|
|
}
|
|
|
|
// GetAttesterSubnetIDs gets the subnet IDs for subscribed subnets for attesters of the slot.
|
|
func (s *subnetIDs) GetAttesterSubnetIDs(slot types.Slot) []uint64 {
|
|
s.attesterLock.RLock()
|
|
defer s.attesterLock.RUnlock()
|
|
|
|
val, exists := s.attester.Get(slot)
|
|
if !exists {
|
|
return nil
|
|
}
|
|
if v, ok := val.([]uint64); ok {
|
|
return v
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// AddAggregatorSubnetID adds the subnet ID for subscribing subnet for the aggregator of a given slot.
|
|
func (s *subnetIDs) AddAggregatorSubnetID(slot types.Slot, subnetID uint64) {
|
|
s.aggregatorLock.Lock()
|
|
defer s.aggregatorLock.Unlock()
|
|
|
|
ids := []uint64{subnetID}
|
|
val, exists := s.aggregator.Get(slot)
|
|
if exists {
|
|
ids = sliceutil.UnionUint64(append(val.([]uint64), ids...))
|
|
}
|
|
s.aggregator.Add(slot, ids)
|
|
}
|
|
|
|
// GetAggregatorSubnetIDs gets the subnet IDs for subscribing subnet for aggregator of the slot.
|
|
func (s *subnetIDs) GetAggregatorSubnetIDs(slot types.Slot) []uint64 {
|
|
s.aggregatorLock.RLock()
|
|
defer s.aggregatorLock.RUnlock()
|
|
|
|
val, exists := s.aggregator.Get(slot)
|
|
if !exists {
|
|
return []uint64{}
|
|
}
|
|
return val.([]uint64)
|
|
}
|
|
|
|
// GetPersistentSubnets retrieves the persistent subnet and expiration time of that validator's
|
|
// subscription.
|
|
func (s *subnetIDs) GetPersistentSubnets(pubkey []byte) ([]uint64, bool, time.Time) {
|
|
s.subnetsLock.RLock()
|
|
defer s.subnetsLock.RUnlock()
|
|
|
|
id, duration, ok := s.persistentSubnets.GetWithExpiration(string(pubkey))
|
|
if !ok {
|
|
return []uint64{}, ok, time.Time{}
|
|
}
|
|
return id.([]uint64), ok, duration
|
|
}
|
|
|
|
// GetAllSubnets retrieves all the non-expired subscribed subnets of all the validators
|
|
// in the cache.
|
|
func (s *subnetIDs) GetAllSubnets() []uint64 {
|
|
s.subnetsLock.RLock()
|
|
defer s.subnetsLock.RUnlock()
|
|
|
|
itemsMap := s.persistentSubnets.Items()
|
|
var committees []uint64
|
|
|
|
for _, v := range itemsMap {
|
|
if v.Expired() {
|
|
continue
|
|
}
|
|
committees = append(committees, v.Object.([]uint64)...)
|
|
}
|
|
return sliceutil.SetUint64(committees)
|
|
}
|
|
|
|
// AddPersistentCommittee adds the relevant committee for that particular validator along with its
|
|
// expiration period.
|
|
func (s *subnetIDs) AddPersistentCommittee(pubkey []byte, comIndex []uint64, duration time.Duration) {
|
|
s.subnetsLock.Lock()
|
|
defer s.subnetsLock.Unlock()
|
|
|
|
s.persistentSubnets.Set(string(pubkey), comIndex, duration)
|
|
}
|
|
|
|
// EmptyAllCaches empties out all the related caches and flushes any stored
|
|
// entries on them. This should only ever be used for testing, in normal
|
|
// production, handling of the relevant subnets for each role is done
|
|
// separately.
|
|
func (s *subnetIDs) EmptyAllCaches() {
|
|
// Clear the caches.
|
|
s.attesterLock.Lock()
|
|
s.attester.Purge()
|
|
s.attesterLock.Unlock()
|
|
|
|
s.aggregatorLock.Lock()
|
|
s.aggregator.Purge()
|
|
s.aggregatorLock.Unlock()
|
|
|
|
s.subnetsLock.Lock()
|
|
s.persistentSubnets.Flush()
|
|
s.subnetsLock.Unlock()
|
|
}
|