mirror of
https://gitlab.com/pulsechaincom/go-pulse.git
synced 2024-12-24 12:27:17 +00:00
c2003ed63b
This change - implements concurrent LES request serving even for a single peer. - replaces the request cost estimation method with a cost table based on benchmarks which gives much more consistent results. Until now the allowed number of light peers was just a guess which probably contributed a lot to the fluctuating quality of available service. Everything related to request cost is implemented in a single object, the 'cost tracker'. It uses a fixed cost table with a global 'correction factor'. Benchmark code is included and can be run at any time to adapt costs to low-level implementation changes. - reimplements flowcontrol.ClientManager in a cleaner and more efficient way, with added capabilities: There is now control over bandwidth, which allows using the flow control parameters for client prioritization. Target utilization over 100 percent is now supported to model concurrent request processing. Total serving bandwidth is reduced during block processing to prevent database contention. - implements an RPC API for the LES servers allowing server operators to assign priority bandwidth to certain clients and change prioritized status even while the client is connected. The new API is meant for cases where server operators charge for LES using an off-protocol mechanism. - adds a unit test for the new client manager. - adds an end-to-end test using the network simulator that tests bandwidth control functions through the new API.
526 lines
15 KiB
Go
526 lines
15 KiB
Go
// Copyright 2016 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package les
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"flag"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"math/rand"
|
|
"os"
|
|
"sync"
|
|
"sync/atomic"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
|
"github.com/ethereum/go-ethereum/common/hexutil"
|
|
"github.com/ethereum/go-ethereum/consensus/ethash"
|
|
"github.com/ethereum/go-ethereum/eth"
|
|
"github.com/ethereum/go-ethereum/eth/downloader"
|
|
"github.com/ethereum/go-ethereum/les/flowcontrol"
|
|
"github.com/ethereum/go-ethereum/log"
|
|
"github.com/ethereum/go-ethereum/node"
|
|
"github.com/ethereum/go-ethereum/p2p/enode"
|
|
"github.com/ethereum/go-ethereum/p2p/simulations"
|
|
"github.com/ethereum/go-ethereum/p2p/simulations/adapters"
|
|
"github.com/ethereum/go-ethereum/rpc"
|
|
colorable "github.com/mattn/go-colorable"
|
|
)
|
|
|
|
/*
|
|
This test is not meant to be a part of the automatic testing process because it
|
|
runs for a long time and also requires a large database in order to do a meaningful
|
|
request performance test. When testServerDataDir is empty, the test is skipped.
|
|
*/
|
|
|
|
const (
|
|
testServerDataDir = "" // should always be empty on the master branch
|
|
testServerCapacity = 200
|
|
testMaxClients = 10
|
|
testTolerance = 0.1
|
|
minRelCap = 0.2
|
|
)
|
|
|
|
func TestCapacityAPI3(t *testing.T) {
|
|
testCapacityAPI(t, 3)
|
|
}
|
|
|
|
func TestCapacityAPI6(t *testing.T) {
|
|
testCapacityAPI(t, 6)
|
|
}
|
|
|
|
func TestCapacityAPI10(t *testing.T) {
|
|
testCapacityAPI(t, 10)
|
|
}
|
|
|
|
// testCapacityAPI runs an end-to-end simulation test connecting one server with
|
|
// a given number of clients. It sets different priority capacities to all clients
|
|
// except a randomly selected one which runs in free client mode. All clients send
|
|
// similar requests at the maximum allowed rate and the test verifies whether the
|
|
// ratio of processed requests is close enough to the ratio of assigned capacities.
|
|
// Running multiple rounds with different settings ensures that changing capacity
|
|
// while connected and going back and forth between free and priority mode with
|
|
// the supplied API calls is also thoroughly tested.
|
|
func testCapacityAPI(t *testing.T, clientCount int) {
|
|
if testServerDataDir == "" {
|
|
// Skip test if no data dir specified
|
|
return
|
|
}
|
|
|
|
for !testSim(t, 1, clientCount, []string{testServerDataDir}, nil, func(ctx context.Context, net *simulations.Network, servers []*simulations.Node, clients []*simulations.Node) bool {
|
|
if len(servers) != 1 {
|
|
t.Fatalf("Invalid number of servers: %d", len(servers))
|
|
}
|
|
server := servers[0]
|
|
|
|
clientRpcClients := make([]*rpc.Client, len(clients))
|
|
|
|
serverRpcClient, err := server.Client()
|
|
if err != nil {
|
|
t.Fatalf("Failed to obtain rpc client: %v", err)
|
|
}
|
|
headNum, headHash := getHead(ctx, t, serverRpcClient)
|
|
totalCap := getTotalCap(ctx, t, serverRpcClient)
|
|
minCap := getMinCap(ctx, t, serverRpcClient)
|
|
testCap := totalCap * 3 / 4
|
|
fmt.Printf("Server testCap: %d minCap: %d head number: %d head hash: %064x\n", testCap, minCap, headNum, headHash)
|
|
reqMinCap := uint64(float64(testCap) * minRelCap / (minRelCap + float64(len(clients)-1)))
|
|
if minCap > reqMinCap {
|
|
t.Fatalf("Minimum client capacity (%d) bigger than required minimum for this test (%d)", minCap, reqMinCap)
|
|
}
|
|
|
|
freeIdx := rand.Intn(len(clients))
|
|
freeCap := getFreeCap(ctx, t, serverRpcClient)
|
|
|
|
for i, client := range clients {
|
|
var err error
|
|
clientRpcClients[i], err = client.Client()
|
|
if err != nil {
|
|
t.Fatalf("Failed to obtain rpc client: %v", err)
|
|
}
|
|
|
|
fmt.Println("connecting client", i)
|
|
if i != freeIdx {
|
|
setCapacity(ctx, t, serverRpcClient, client.ID(), testCap/uint64(len(clients)))
|
|
}
|
|
net.Connect(client.ID(), server.ID())
|
|
|
|
for {
|
|
select {
|
|
case <-ctx.Done():
|
|
t.Fatalf("Timeout")
|
|
default:
|
|
}
|
|
num, hash := getHead(ctx, t, clientRpcClients[i])
|
|
if num == headNum && hash == headHash {
|
|
fmt.Println("client", i, "synced")
|
|
break
|
|
}
|
|
time.Sleep(time.Millisecond * 200)
|
|
}
|
|
}
|
|
|
|
var wg sync.WaitGroup
|
|
stop := make(chan struct{})
|
|
|
|
reqCount := make([]uint64, len(clientRpcClients))
|
|
|
|
for i, c := range clientRpcClients {
|
|
wg.Add(1)
|
|
i, c := i, c
|
|
go func() {
|
|
queue := make(chan struct{}, 100)
|
|
var count uint64
|
|
for {
|
|
select {
|
|
case queue <- struct{}{}:
|
|
select {
|
|
case <-stop:
|
|
wg.Done()
|
|
return
|
|
case <-ctx.Done():
|
|
wg.Done()
|
|
return
|
|
default:
|
|
wg.Add(1)
|
|
go func() {
|
|
ok := testRequest(ctx, t, c)
|
|
wg.Done()
|
|
<-queue
|
|
if ok {
|
|
count++
|
|
atomic.StoreUint64(&reqCount[i], count)
|
|
}
|
|
}()
|
|
}
|
|
case <-stop:
|
|
wg.Done()
|
|
return
|
|
case <-ctx.Done():
|
|
wg.Done()
|
|
return
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
|
|
processedSince := func(start []uint64) []uint64 {
|
|
res := make([]uint64, len(reqCount))
|
|
for i := range reqCount {
|
|
res[i] = atomic.LoadUint64(&reqCount[i])
|
|
if start != nil {
|
|
res[i] -= start[i]
|
|
}
|
|
}
|
|
return res
|
|
}
|
|
|
|
weights := make([]float64, len(clients))
|
|
for c := 0; c < 5; c++ {
|
|
setCapacity(ctx, t, serverRpcClient, clients[freeIdx].ID(), freeCap)
|
|
freeIdx = rand.Intn(len(clients))
|
|
var sum float64
|
|
for i := range clients {
|
|
if i == freeIdx {
|
|
weights[i] = 0
|
|
} else {
|
|
weights[i] = rand.Float64()*(1-minRelCap) + minRelCap
|
|
}
|
|
sum += weights[i]
|
|
}
|
|
for i, client := range clients {
|
|
weights[i] *= float64(testCap-freeCap-100) / sum
|
|
capacity := uint64(weights[i])
|
|
if i != freeIdx && capacity < getCapacity(ctx, t, serverRpcClient, client.ID()) {
|
|
setCapacity(ctx, t, serverRpcClient, client.ID(), capacity)
|
|
}
|
|
}
|
|
setCapacity(ctx, t, serverRpcClient, clients[freeIdx].ID(), 0)
|
|
for i, client := range clients {
|
|
capacity := uint64(weights[i])
|
|
if i != freeIdx && capacity > getCapacity(ctx, t, serverRpcClient, client.ID()) {
|
|
setCapacity(ctx, t, serverRpcClient, client.ID(), capacity)
|
|
}
|
|
}
|
|
weights[freeIdx] = float64(freeCap)
|
|
for i := range clients {
|
|
weights[i] /= float64(testCap)
|
|
}
|
|
|
|
time.Sleep(flowcontrol.DecParamDelay)
|
|
fmt.Println("Starting measurement")
|
|
fmt.Printf("Relative weights:")
|
|
for i := range clients {
|
|
fmt.Printf(" %f", weights[i])
|
|
}
|
|
fmt.Println()
|
|
start := processedSince(nil)
|
|
for {
|
|
select {
|
|
case <-ctx.Done():
|
|
t.Fatalf("Timeout")
|
|
default:
|
|
}
|
|
|
|
totalCap = getTotalCap(ctx, t, serverRpcClient)
|
|
if totalCap < testCap {
|
|
fmt.Println("Total capacity underrun")
|
|
close(stop)
|
|
wg.Wait()
|
|
return false
|
|
}
|
|
|
|
processed := processedSince(start)
|
|
var avg uint64
|
|
fmt.Printf("Processed")
|
|
for i, p := range processed {
|
|
fmt.Printf(" %d", p)
|
|
processed[i] = uint64(float64(p) / weights[i])
|
|
avg += processed[i]
|
|
}
|
|
avg /= uint64(len(processed))
|
|
|
|
if avg >= 10000 {
|
|
var maxDev float64
|
|
for _, p := range processed {
|
|
dev := float64(int64(p-avg)) / float64(avg)
|
|
fmt.Printf(" %7.4f", dev)
|
|
if dev < 0 {
|
|
dev = -dev
|
|
}
|
|
if dev > maxDev {
|
|
maxDev = dev
|
|
}
|
|
}
|
|
fmt.Printf(" max deviation: %f totalCap: %d\n", maxDev, totalCap)
|
|
if maxDev <= testTolerance {
|
|
fmt.Println("success")
|
|
break
|
|
}
|
|
} else {
|
|
fmt.Println()
|
|
}
|
|
time.Sleep(time.Millisecond * 200)
|
|
}
|
|
}
|
|
|
|
close(stop)
|
|
wg.Wait()
|
|
|
|
for i, count := range reqCount {
|
|
fmt.Println("client", i, "processed", count)
|
|
}
|
|
return true
|
|
}) {
|
|
fmt.Println("restarting test")
|
|
}
|
|
}
|
|
|
|
func getHead(ctx context.Context, t *testing.T, client *rpc.Client) (uint64, common.Hash) {
|
|
res := make(map[string]interface{})
|
|
if err := client.CallContext(ctx, &res, "eth_getBlockByNumber", "latest", false); err != nil {
|
|
t.Fatalf("Failed to obtain head block: %v", err)
|
|
}
|
|
numStr, ok := res["number"].(string)
|
|
if !ok {
|
|
t.Fatalf("RPC block number field invalid")
|
|
}
|
|
num, err := hexutil.DecodeUint64(numStr)
|
|
if err != nil {
|
|
t.Fatalf("Failed to decode RPC block number: %v", err)
|
|
}
|
|
hashStr, ok := res["hash"].(string)
|
|
if !ok {
|
|
t.Fatalf("RPC block number field invalid")
|
|
}
|
|
hash := common.HexToHash(hashStr)
|
|
return num, hash
|
|
}
|
|
|
|
func testRequest(ctx context.Context, t *testing.T, client *rpc.Client) bool {
|
|
//res := make(map[string]interface{})
|
|
var res string
|
|
var addr common.Address
|
|
rand.Read(addr[:])
|
|
c, _ := context.WithTimeout(ctx, time.Second*12)
|
|
// if err := client.CallContext(ctx, &res, "eth_getProof", addr, nil, "latest"); err != nil {
|
|
err := client.CallContext(c, &res, "eth_getBalance", addr, "latest")
|
|
if err != nil {
|
|
fmt.Println("request error:", err)
|
|
}
|
|
return err == nil
|
|
}
|
|
|
|
func setCapacity(ctx context.Context, t *testing.T, server *rpc.Client, clientID enode.ID, cap uint64) {
|
|
if err := server.CallContext(ctx, nil, "les_setClientCapacity", clientID, cap); err != nil {
|
|
t.Fatalf("Failed to set client capacity: %v", err)
|
|
}
|
|
}
|
|
|
|
func getCapacity(ctx context.Context, t *testing.T, server *rpc.Client, clientID enode.ID) uint64 {
|
|
var s string
|
|
if err := server.CallContext(ctx, &s, "les_getClientCapacity", clientID); err != nil {
|
|
t.Fatalf("Failed to get client capacity: %v", err)
|
|
}
|
|
cap, err := hexutil.DecodeUint64(s)
|
|
if err != nil {
|
|
t.Fatalf("Failed to decode client capacity: %v", err)
|
|
}
|
|
return cap
|
|
}
|
|
|
|
func getTotalCap(ctx context.Context, t *testing.T, server *rpc.Client) uint64 {
|
|
var s string
|
|
if err := server.CallContext(ctx, &s, "les_totalCapacity"); err != nil {
|
|
t.Fatalf("Failed to query total capacity: %v", err)
|
|
}
|
|
total, err := hexutil.DecodeUint64(s)
|
|
if err != nil {
|
|
t.Fatalf("Failed to decode total capacity: %v", err)
|
|
}
|
|
return total
|
|
}
|
|
|
|
func getMinCap(ctx context.Context, t *testing.T, server *rpc.Client) uint64 {
|
|
var s string
|
|
if err := server.CallContext(ctx, &s, "les_minimumCapacity"); err != nil {
|
|
t.Fatalf("Failed to query minimum capacity: %v", err)
|
|
}
|
|
min, err := hexutil.DecodeUint64(s)
|
|
if err != nil {
|
|
t.Fatalf("Failed to decode minimum capacity: %v", err)
|
|
}
|
|
return min
|
|
}
|
|
|
|
func getFreeCap(ctx context.Context, t *testing.T, server *rpc.Client) uint64 {
|
|
var s string
|
|
if err := server.CallContext(ctx, &s, "les_freeClientCapacity"); err != nil {
|
|
t.Fatalf("Failed to query free client capacity: %v", err)
|
|
}
|
|
free, err := hexutil.DecodeUint64(s)
|
|
if err != nil {
|
|
t.Fatalf("Failed to decode free client capacity: %v", err)
|
|
}
|
|
return free
|
|
}
|
|
|
|
func init() {
|
|
flag.Parse()
|
|
// register the Delivery service which will run as a devp2p
|
|
// protocol when using the exec adapter
|
|
adapters.RegisterServices(services)
|
|
|
|
log.PrintOrigins(true)
|
|
log.Root().SetHandler(log.LvlFilterHandler(log.Lvl(*loglevel), log.StreamHandler(colorable.NewColorableStderr(), log.TerminalFormat(true))))
|
|
}
|
|
|
|
var (
|
|
adapter = flag.String("adapter", "exec", "type of simulation: sim|socket|exec|docker")
|
|
loglevel = flag.Int("loglevel", 0, "verbosity of logs")
|
|
nodes = flag.Int("nodes", 0, "number of nodes")
|
|
)
|
|
|
|
var services = adapters.Services{
|
|
"lesclient": newLesClientService,
|
|
"lesserver": newLesServerService,
|
|
}
|
|
|
|
func NewNetwork() (*simulations.Network, func(), error) {
|
|
adapter, adapterTeardown, err := NewAdapter(*adapter, services)
|
|
if err != nil {
|
|
return nil, adapterTeardown, err
|
|
}
|
|
defaultService := "streamer"
|
|
net := simulations.NewNetwork(adapter, &simulations.NetworkConfig{
|
|
ID: "0",
|
|
DefaultService: defaultService,
|
|
})
|
|
teardown := func() {
|
|
adapterTeardown()
|
|
net.Shutdown()
|
|
}
|
|
|
|
return net, teardown, nil
|
|
}
|
|
|
|
func NewAdapter(adapterType string, services adapters.Services) (adapter adapters.NodeAdapter, teardown func(), err error) {
|
|
teardown = func() {}
|
|
switch adapterType {
|
|
case "sim":
|
|
adapter = adapters.NewSimAdapter(services)
|
|
// case "socket":
|
|
// adapter = adapters.NewSocketAdapter(services)
|
|
case "exec":
|
|
baseDir, err0 := ioutil.TempDir("", "les-test")
|
|
if err0 != nil {
|
|
return nil, teardown, err0
|
|
}
|
|
teardown = func() { os.RemoveAll(baseDir) }
|
|
adapter = adapters.NewExecAdapter(baseDir)
|
|
/*case "docker":
|
|
adapter, err = adapters.NewDockerAdapter()
|
|
if err != nil {
|
|
return nil, teardown, err
|
|
}*/
|
|
default:
|
|
return nil, teardown, errors.New("adapter needs to be one of sim, socket, exec, docker")
|
|
}
|
|
return adapter, teardown, nil
|
|
}
|
|
|
|
func testSim(t *testing.T, serverCount, clientCount int, serverDir, clientDir []string, test func(ctx context.Context, net *simulations.Network, servers []*simulations.Node, clients []*simulations.Node) bool) bool {
|
|
net, teardown, err := NewNetwork()
|
|
defer teardown()
|
|
if err != nil {
|
|
t.Fatalf("Failed to create network: %v", err)
|
|
}
|
|
timeout := 1800 * time.Second
|
|
ctx, cancel := context.WithTimeout(context.Background(), timeout)
|
|
defer cancel()
|
|
|
|
servers := make([]*simulations.Node, serverCount)
|
|
clients := make([]*simulations.Node, clientCount)
|
|
|
|
for i := range clients {
|
|
clientconf := adapters.RandomNodeConfig()
|
|
clientconf.Services = []string{"lesclient"}
|
|
if len(clientDir) == clientCount {
|
|
clientconf.DataDir = clientDir[i]
|
|
}
|
|
client, err := net.NewNodeWithConfig(clientconf)
|
|
if err != nil {
|
|
t.Fatalf("Failed to create client: %v", err)
|
|
}
|
|
clients[i] = client
|
|
}
|
|
|
|
for i := range servers {
|
|
serverconf := adapters.RandomNodeConfig()
|
|
serverconf.Services = []string{"lesserver"}
|
|
if len(serverDir) == serverCount {
|
|
serverconf.DataDir = serverDir[i]
|
|
}
|
|
server, err := net.NewNodeWithConfig(serverconf)
|
|
if err != nil {
|
|
t.Fatalf("Failed to create server: %v", err)
|
|
}
|
|
servers[i] = server
|
|
}
|
|
|
|
for _, client := range clients {
|
|
if err := net.Start(client.ID()); err != nil {
|
|
t.Fatalf("Failed to start client node: %v", err)
|
|
}
|
|
}
|
|
for _, server := range servers {
|
|
if err := net.Start(server.ID()); err != nil {
|
|
t.Fatalf("Failed to start server node: %v", err)
|
|
}
|
|
}
|
|
|
|
return test(ctx, net, servers, clients)
|
|
}
|
|
|
|
func newLesClientService(ctx *adapters.ServiceContext) (node.Service, error) {
|
|
config := eth.DefaultConfig
|
|
config.SyncMode = downloader.LightSync
|
|
config.Ethash.PowMode = ethash.ModeFake
|
|
return New(ctx.NodeContext, &config)
|
|
}
|
|
|
|
func newLesServerService(ctx *adapters.ServiceContext) (node.Service, error) {
|
|
config := eth.DefaultConfig
|
|
config.SyncMode = downloader.FullSync
|
|
config.LightServ = testServerCapacity
|
|
config.LightPeers = testMaxClients
|
|
ethereum, err := eth.New(ctx.NodeContext, &config)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
server, err := NewLesServer(ethereum, &config)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
ethereum.AddLesServer(server)
|
|
return ethereum, nil
|
|
}
|