Compare commits
10 Commits
1daf2a4470
...
eb87329ae3
Author | SHA1 | Date |
---|---|---|
lightclient | eb87329ae3 | |
lightclient | 4abf2e5afd | |
Arran Schlosberg | 23800122b3 | |
Jordan Krage | 3c754e2a09 | |
Hyunsoo Shin (Lake) | 19fa71b917 | |
Martin HS | 02159d553f | |
Martin HS | ab4a1cc01f | |
lightclient | 1635da8021 | |
Martin Holst Swende | f6971a729c | |
lightclient | 83b532c4da |
|
@ -42,7 +42,7 @@ func MakeTopics(query ...[]interface{}) ([][]common.Hash, error) {
|
||||||
case common.Address:
|
case common.Address:
|
||||||
copy(topic[common.HashLength-common.AddressLength:], rule[:])
|
copy(topic[common.HashLength-common.AddressLength:], rule[:])
|
||||||
case *big.Int:
|
case *big.Int:
|
||||||
copy(topic[:], math.U256Bytes(rule))
|
copy(topic[:], math.U256Bytes(new(big.Int).Set(rule)))
|
||||||
case bool:
|
case bool:
|
||||||
if rule {
|
if rule {
|
||||||
topic[common.HashLength-1] = 1
|
topic[common.HashLength-1] = 1
|
||||||
|
|
|
@ -149,6 +149,23 @@ func TestMakeTopics(t *testing.T) {
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
t.Run("does not mutate big.Int", func(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
want := [][]common.Hash{{common.HexToHash("ffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffff")}}
|
||||||
|
|
||||||
|
in := big.NewInt(-1)
|
||||||
|
got, err := MakeTopics([]interface{}{in})
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("makeTopics() error = %v", err)
|
||||||
|
}
|
||||||
|
if !reflect.DeepEqual(got, want) {
|
||||||
|
t.Fatalf("makeTopics() = %v, want %v", got, want)
|
||||||
|
}
|
||||||
|
if orig := big.NewInt(-1); in.Cmp(orig) != 0 {
|
||||||
|
t.Fatalf("makeTopics() mutated an input parameter from %v to %v", orig, in)
|
||||||
|
}
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
type args struct {
|
type args struct {
|
||||||
|
|
|
@ -282,14 +282,13 @@ func importChain(ctx *cli.Context) error {
|
||||||
if ctx.Args().Len() < 1 {
|
if ctx.Args().Len() < 1 {
|
||||||
utils.Fatalf("This command requires an argument.")
|
utils.Fatalf("This command requires an argument.")
|
||||||
}
|
}
|
||||||
// Start metrics export if enabled
|
stack, cfg := makeConfigNode(ctx)
|
||||||
utils.SetupMetrics(ctx)
|
|
||||||
// Start system runtime metrics collection
|
|
||||||
go metrics.CollectProcessMetrics(3 * time.Second)
|
|
||||||
|
|
||||||
stack, _ := makeConfigNode(ctx)
|
|
||||||
defer stack.Close()
|
defer stack.Close()
|
||||||
|
|
||||||
|
// Start metrics export if enabled
|
||||||
|
utils.SetupMetrics(&cfg.Metrics)
|
||||||
|
go metrics.CollectProcessMetrics(3 * time.Second)
|
||||||
|
|
||||||
chain, db := utils.MakeChain(ctx, stack, false)
|
chain, db := utils.MakeChain(ctx, stack, false)
|
||||||
defer db.Close()
|
defer db.Close()
|
||||||
|
|
||||||
|
|
|
@ -25,6 +25,7 @@ import (
|
||||||
"runtime"
|
"runtime"
|
||||||
"slices"
|
"slices"
|
||||||
"strings"
|
"strings"
|
||||||
|
"time"
|
||||||
"unicode"
|
"unicode"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/accounts"
|
"github.com/ethereum/go-ethereum/accounts"
|
||||||
|
@ -150,7 +151,7 @@ func loadBaseConfig(ctx *cli.Context) gethConfig {
|
||||||
// Load config file.
|
// Load config file.
|
||||||
if file := ctx.String(configFileFlag.Name); file != "" {
|
if file := ctx.String(configFileFlag.Name); file != "" {
|
||||||
if err := loadConfig(file, &cfg); err != nil {
|
if err := loadConfig(file, &cfg); err != nil {
|
||||||
utils.Fatalf("%v", err)
|
utils.Fatalf("failed to load config: %v", err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -192,6 +193,10 @@ func makeFullNode(ctx *cli.Context) *node.Node {
|
||||||
cfg.Eth.OverrideVerkle = &v
|
cfg.Eth.OverrideVerkle = &v
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Start metrics export if enabled
|
||||||
|
utils.SetupMetrics(&cfg.Metrics)
|
||||||
|
go metrics.CollectProcessMetrics(3 * time.Second)
|
||||||
|
|
||||||
backend, eth := utils.RegisterEthService(stack, &cfg.Eth)
|
backend, eth := utils.RegisterEthService(stack, &cfg.Eth)
|
||||||
|
|
||||||
// Create gauge with geth system and build information
|
// Create gauge with geth system and build information
|
||||||
|
|
|
@ -34,7 +34,6 @@ import (
|
||||||
"github.com/ethereum/go-ethereum/internal/debug"
|
"github.com/ethereum/go-ethereum/internal/debug"
|
||||||
"github.com/ethereum/go-ethereum/internal/flags"
|
"github.com/ethereum/go-ethereum/internal/flags"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"github.com/ethereum/go-ethereum/metrics"
|
|
||||||
"github.com/ethereum/go-ethereum/node"
|
"github.com/ethereum/go-ethereum/node"
|
||||||
"go.uber.org/automaxprocs/maxprocs"
|
"go.uber.org/automaxprocs/maxprocs"
|
||||||
|
|
||||||
|
@ -325,12 +324,6 @@ func prepare(ctx *cli.Context) {
|
||||||
ctx.Set(utils.CacheFlag.Name, strconv.Itoa(4096))
|
ctx.Set(utils.CacheFlag.Name, strconv.Itoa(4096))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Start metrics export if enabled
|
|
||||||
utils.SetupMetrics(ctx)
|
|
||||||
|
|
||||||
// Start system runtime metrics collection
|
|
||||||
go metrics.CollectProcessMetrics(3 * time.Second)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// geth is the main entry point into the system if no special subcommand is run.
|
// geth is the main entry point into the system if no special subcommand is run.
|
||||||
|
|
|
@ -1969,64 +1969,60 @@ func RegisterFullSyncTester(stack *node.Node, eth *eth.Ethereum, target common.H
|
||||||
log.Info("Registered full-sync tester", "hash", target)
|
log.Info("Registered full-sync tester", "hash", target)
|
||||||
}
|
}
|
||||||
|
|
||||||
func SetupMetrics(ctx *cli.Context) {
|
func SetupMetrics(cfg *metrics.Config) {
|
||||||
if metrics.Enabled {
|
if !cfg.Enabled {
|
||||||
log.Info("Enabling metrics collection")
|
return
|
||||||
|
}
|
||||||
|
log.Info("Enabling metrics collection")
|
||||||
|
var (
|
||||||
|
enableExport = cfg.EnableInfluxDB
|
||||||
|
enableExportV2 = cfg.EnableInfluxDBV2
|
||||||
|
)
|
||||||
|
if enableExport && enableExportV2 {
|
||||||
|
Fatalf("Flags %v can't be used at the same time", strings.Join([]string{MetricsEnableInfluxDBFlag.Name, MetricsEnableInfluxDBV2Flag.Name}, ", "))
|
||||||
|
}
|
||||||
|
if enableExport || enableExportV2 {
|
||||||
|
v1FlagIsSet := cfg.InfluxDBUsername != "" || cfg.InfluxDBPassword != ""
|
||||||
|
v2FlagIsSet := cfg.InfluxDBToken != "" || cfg.InfluxDBOrganization != "" || cfg.InfluxDBBucket != ""
|
||||||
|
|
||||||
var (
|
if enableExport && v2FlagIsSet {
|
||||||
enableExport = ctx.Bool(MetricsEnableInfluxDBFlag.Name)
|
Fatalf("Flags --influxdb.metrics.organization, --influxdb.metrics.token, --influxdb.metrics.bucket are only available for influxdb-v2")
|
||||||
enableExportV2 = ctx.Bool(MetricsEnableInfluxDBV2Flag.Name)
|
} else if enableExportV2 && v1FlagIsSet {
|
||||||
)
|
Fatalf("Flags --influxdb.metrics.username, --influxdb.metrics.password are only available for influxdb-v1")
|
||||||
|
|
||||||
if enableExport || enableExportV2 {
|
|
||||||
CheckExclusive(ctx, MetricsEnableInfluxDBFlag, MetricsEnableInfluxDBV2Flag)
|
|
||||||
|
|
||||||
v1FlagIsSet := ctx.IsSet(MetricsInfluxDBUsernameFlag.Name) ||
|
|
||||||
ctx.IsSet(MetricsInfluxDBPasswordFlag.Name)
|
|
||||||
|
|
||||||
v2FlagIsSet := ctx.IsSet(MetricsInfluxDBTokenFlag.Name) ||
|
|
||||||
ctx.IsSet(MetricsInfluxDBOrganizationFlag.Name) ||
|
|
||||||
ctx.IsSet(MetricsInfluxDBBucketFlag.Name)
|
|
||||||
|
|
||||||
if enableExport && v2FlagIsSet {
|
|
||||||
Fatalf("Flags --influxdb.metrics.organization, --influxdb.metrics.token, --influxdb.metrics.bucket are only available for influxdb-v2")
|
|
||||||
} else if enableExportV2 && v1FlagIsSet {
|
|
||||||
Fatalf("Flags --influxdb.metrics.username, --influxdb.metrics.password are only available for influxdb-v1")
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
var (
|
var (
|
||||||
endpoint = ctx.String(MetricsInfluxDBEndpointFlag.Name)
|
endpoint = cfg.InfluxDBEndpoint
|
||||||
database = ctx.String(MetricsInfluxDBDatabaseFlag.Name)
|
database = cfg.InfluxDBDatabase
|
||||||
username = ctx.String(MetricsInfluxDBUsernameFlag.Name)
|
username = cfg.InfluxDBUsername
|
||||||
password = ctx.String(MetricsInfluxDBPasswordFlag.Name)
|
password = cfg.InfluxDBPassword
|
||||||
|
|
||||||
token = ctx.String(MetricsInfluxDBTokenFlag.Name)
|
token = cfg.InfluxDBToken
|
||||||
bucket = ctx.String(MetricsInfluxDBBucketFlag.Name)
|
bucket = cfg.InfluxDBBucket
|
||||||
organization = ctx.String(MetricsInfluxDBOrganizationFlag.Name)
|
organization = cfg.InfluxDBOrganization
|
||||||
)
|
)
|
||||||
|
|
||||||
if enableExport {
|
if enableExport {
|
||||||
tagsMap := SplitTagsFlag(ctx.String(MetricsInfluxDBTagsFlag.Name))
|
tagsMap := SplitTagsFlag(cfg.InfluxDBTags)
|
||||||
|
|
||||||
log.Info("Enabling metrics export to InfluxDB")
|
log.Info("Enabling metrics export to InfluxDB")
|
||||||
|
|
||||||
go influxdb.InfluxDBWithTags(metrics.DefaultRegistry, 10*time.Second, endpoint, database, username, password, "geth.", tagsMap)
|
go influxdb.InfluxDBWithTags(metrics.DefaultRegistry, 10*time.Second, endpoint, database, username, password, "geth.", tagsMap)
|
||||||
} else if enableExportV2 {
|
} else if enableExportV2 {
|
||||||
tagsMap := SplitTagsFlag(ctx.String(MetricsInfluxDBTagsFlag.Name))
|
tagsMap := SplitTagsFlag(cfg.InfluxDBTags)
|
||||||
|
|
||||||
log.Info("Enabling metrics export to InfluxDB (v2)")
|
log.Info("Enabling metrics export to InfluxDB (v2)")
|
||||||
|
|
||||||
go influxdb.InfluxDBV2WithTags(metrics.DefaultRegistry, 10*time.Second, endpoint, token, bucket, organization, "geth.", tagsMap)
|
go influxdb.InfluxDBV2WithTags(metrics.DefaultRegistry, 10*time.Second, endpoint, token, bucket, organization, "geth.", tagsMap)
|
||||||
}
|
}
|
||||||
|
|
||||||
if ctx.IsSet(MetricsHTTPFlag.Name) {
|
if cfg.HTTP != "" {
|
||||||
address := net.JoinHostPort(ctx.String(MetricsHTTPFlag.Name), fmt.Sprintf("%d", ctx.Int(MetricsPortFlag.Name)))
|
address := net.JoinHostPort(cfg.HTTP, fmt.Sprintf("%d", cfg.Port))
|
||||||
log.Info("Enabling stand-alone metrics HTTP endpoint", "address", address)
|
log.Info("Enabling stand-alone metrics HTTP endpoint", "address", address)
|
||||||
exp.Setup(address)
|
exp.Setup(address)
|
||||||
} else if ctx.IsSet(MetricsPortFlag.Name) {
|
} else if cfg.HTTP == "" && cfg.Port != 0 {
|
||||||
log.Warn(fmt.Sprintf("--%s specified without --%s, metrics server will not start.", MetricsPortFlag.Name, MetricsHTTPFlag.Name))
|
log.Warn(fmt.Sprintf("--%s specified without --%s, metrics server will not start.", MetricsPortFlag.Name, MetricsHTTPFlag.Name))
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -206,47 +206,24 @@ func New(config Config, diskdb ethdb.KeyValueStore, triedb *triedb.Database, roo
|
||||||
log.Warn("Snapshot maintenance disabled (syncing)")
|
log.Warn("Snapshot maintenance disabled (syncing)")
|
||||||
return snap, nil
|
return snap, nil
|
||||||
}
|
}
|
||||||
// Create the building waiter iff the background generation is allowed
|
|
||||||
if !config.NoBuild && !config.AsyncBuild {
|
|
||||||
defer snap.waitBuild()
|
|
||||||
}
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Warn("Failed to load snapshot", "err", err)
|
log.Warn("Failed to load snapshot", "err", err)
|
||||||
if !config.NoBuild {
|
if config.NoBuild {
|
||||||
snap.Rebuild(root)
|
return nil, err
|
||||||
return snap, nil
|
|
||||||
}
|
}
|
||||||
return nil, err // Bail out the error, don't rebuild automatically.
|
wait := snap.Rebuild(root)
|
||||||
|
if !config.AsyncBuild {
|
||||||
|
wait()
|
||||||
|
}
|
||||||
|
return snap, nil
|
||||||
}
|
}
|
||||||
// Existing snapshot loaded, seed all the layers
|
// Existing snapshot loaded, seed all the layers
|
||||||
for head != nil {
|
for ; head != nil; head = head.Parent() {
|
||||||
snap.layers[head.Root()] = head
|
snap.layers[head.Root()] = head
|
||||||
head = head.Parent()
|
|
||||||
}
|
}
|
||||||
return snap, nil
|
return snap, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// waitBuild blocks until the snapshot finishes rebuilding. This method is meant
|
|
||||||
// to be used by tests to ensure we're testing what we believe we are.
|
|
||||||
func (t *Tree) waitBuild() {
|
|
||||||
// Find the rebuild termination channel
|
|
||||||
var done chan struct{}
|
|
||||||
|
|
||||||
t.lock.RLock()
|
|
||||||
for _, layer := range t.layers {
|
|
||||||
if layer, ok := layer.(*diskLayer); ok {
|
|
||||||
done = layer.genPending
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
t.lock.RUnlock()
|
|
||||||
|
|
||||||
// Wait until the snapshot is generated
|
|
||||||
if done != nil {
|
|
||||||
<-done
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Disable interrupts any pending snapshot generator, deletes all the snapshot
|
// Disable interrupts any pending snapshot generator, deletes all the snapshot
|
||||||
// layers in memory and marks snapshots disabled globally. In order to resume
|
// layers in memory and marks snapshots disabled globally. In order to resume
|
||||||
// the snapshot functionality, the caller must invoke Rebuild.
|
// the snapshot functionality, the caller must invoke Rebuild.
|
||||||
|
@ -688,8 +665,9 @@ func (t *Tree) Journal(root common.Hash) (common.Hash, error) {
|
||||||
|
|
||||||
// Rebuild wipes all available snapshot data from the persistent database and
|
// Rebuild wipes all available snapshot data from the persistent database and
|
||||||
// discard all caches and diff layers. Afterwards, it starts a new snapshot
|
// discard all caches and diff layers. Afterwards, it starts a new snapshot
|
||||||
// generator with the given root hash.
|
// generator with the given root hash. The returned function blocks until
|
||||||
func (t *Tree) Rebuild(root common.Hash) {
|
// regeneration is complete.
|
||||||
|
func (t *Tree) Rebuild(root common.Hash) (wait func()) {
|
||||||
t.lock.Lock()
|
t.lock.Lock()
|
||||||
defer t.lock.Unlock()
|
defer t.lock.Unlock()
|
||||||
|
|
||||||
|
@ -721,9 +699,11 @@ func (t *Tree) Rebuild(root common.Hash) {
|
||||||
// Start generating a new snapshot from scratch on a background thread. The
|
// Start generating a new snapshot from scratch on a background thread. The
|
||||||
// generator will run a wiper first if there's not one running right now.
|
// generator will run a wiper first if there's not one running right now.
|
||||||
log.Info("Rebuilding state snapshot")
|
log.Info("Rebuilding state snapshot")
|
||||||
|
disk := generateSnapshot(t.diskdb, t.triedb, t.config.CacheSize, root)
|
||||||
t.layers = map[common.Hash]snapshot{
|
t.layers = map[common.Hash]snapshot{
|
||||||
root: generateSnapshot(t.diskdb, t.triedb, t.config.CacheSize, root),
|
root: disk,
|
||||||
}
|
}
|
||||||
|
return func() { <-disk.genPending }
|
||||||
}
|
}
|
||||||
|
|
||||||
// AccountIterator creates a new account iterator for the specified root hash and
|
// AccountIterator creates a new account iterator for the specified root hash and
|
||||||
|
|
|
@ -363,26 +363,35 @@ func (t *mdLogger) OnEnter(depth int, typ byte, from common.Address, to common.A
|
||||||
if depth != 0 {
|
if depth != 0 {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
create := vm.OpCode(typ) == vm.CREATE
|
if create := vm.OpCode(typ) == vm.CREATE; !create {
|
||||||
if !create {
|
fmt.Fprintf(t.out, "Pre-execution info:\n"+
|
||||||
fmt.Fprintf(t.out, "From: `%v`\nTo: `%v`\nData: `%#x`\nGas: `%d`\nValue `%v` wei\n",
|
" - from: `%v`\n"+
|
||||||
from.String(), to.String(),
|
" - to: `%v`\n"+
|
||||||
input, gas, value)
|
" - data: `%#x`\n"+
|
||||||
|
" - gas: `%d`\n"+
|
||||||
|
" - value: `%v` wei\n",
|
||||||
|
from.String(), to.String(), input, gas, value)
|
||||||
} else {
|
} else {
|
||||||
fmt.Fprintf(t.out, "From: `%v`\nCreate at: `%v`\nData: `%#x`\nGas: `%d`\nValue `%v` wei\n",
|
fmt.Fprintf(t.out, "Pre-execution info:\n"+
|
||||||
from.String(), to.String(),
|
" - from: `%v`\n"+
|
||||||
input, gas, value)
|
" - create: `%v`\n"+
|
||||||
|
" - data: `%#x`\n"+
|
||||||
|
" - gas: `%d`\n"+
|
||||||
|
" - value: `%v` wei\n",
|
||||||
|
from.String(), to.String(), input, gas, value)
|
||||||
}
|
}
|
||||||
|
|
||||||
fmt.Fprintf(t.out, `
|
fmt.Fprintf(t.out, `
|
||||||
| Pc | Op | Cost | Stack | RStack | Refund |
|
| Pc | Op | Cost | Refund | Stack |
|
||||||
|-------|-------------|------|-----------|-----------|---------|
|
|-------|-------------|------|-----------|-----------|
|
||||||
`)
|
`)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t *mdLogger) OnExit(depth int, output []byte, gasUsed uint64, err error, reverted bool) {
|
func (t *mdLogger) OnExit(depth int, output []byte, gasUsed uint64, err error, reverted bool) {
|
||||||
if depth == 0 {
|
if depth == 0 {
|
||||||
fmt.Fprintf(t.out, "\nOutput: `%#x`\nConsumed gas: `%d`\nError: `%v`\n",
|
fmt.Fprintf(t.out, "\nPost-execution info:\n"+
|
||||||
|
" - output: `%#x`\n"+
|
||||||
|
" - consumed gas: `%d`\n"+
|
||||||
|
" - error: `%v`\n",
|
||||||
output, gasUsed, err)
|
output, gasUsed, err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -390,7 +399,8 @@ func (t *mdLogger) OnExit(depth int, output []byte, gasUsed uint64, err error, r
|
||||||
// OnOpcode also tracks SLOAD/SSTORE ops to track storage change.
|
// OnOpcode also tracks SLOAD/SSTORE ops to track storage change.
|
||||||
func (t *mdLogger) OnOpcode(pc uint64, op byte, gas, cost uint64, scope tracing.OpContext, rData []byte, depth int, err error) {
|
func (t *mdLogger) OnOpcode(pc uint64, op byte, gas, cost uint64, scope tracing.OpContext, rData []byte, depth int, err error) {
|
||||||
stack := scope.StackData()
|
stack := scope.StackData()
|
||||||
fmt.Fprintf(t.out, "| %4d | %10v | %3d |", pc, vm.OpCode(op).String(), cost)
|
fmt.Fprintf(t.out, "| %4d | %10v | %3d |%10v |", pc, vm.OpCode(op).String(),
|
||||||
|
cost, t.env.StateDB.GetRefund())
|
||||||
|
|
||||||
if !t.cfg.DisableStack {
|
if !t.cfg.DisableStack {
|
||||||
// format stack
|
// format stack
|
||||||
|
@ -401,7 +411,6 @@ func (t *mdLogger) OnOpcode(pc uint64, op byte, gas, cost uint64, scope tracing.
|
||||||
b := fmt.Sprintf("[%v]", strings.Join(a, ","))
|
b := fmt.Sprintf("[%v]", strings.Join(a, ","))
|
||||||
fmt.Fprintf(t.out, "%10v |", b)
|
fmt.Fprintf(t.out, "%10v |", b)
|
||||||
}
|
}
|
||||||
fmt.Fprintf(t.out, "%10v |", t.env.StateDB.GetRefund())
|
|
||||||
fmt.Fprintln(t.out, "")
|
fmt.Fprintln(t.out, "")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
fmt.Fprintf(t.out, "Error: %v\n", err)
|
fmt.Fprintf(t.out, "Error: %v\n", err)
|
||||||
|
|
|
@ -71,7 +71,7 @@ func NewJSONLogger(cfg *Config, writer io.Writer) *tracing.Hooks {
|
||||||
l.hooks = &tracing.Hooks{
|
l.hooks = &tracing.Hooks{
|
||||||
OnTxStart: l.OnTxStart,
|
OnTxStart: l.OnTxStart,
|
||||||
OnSystemCallStart: l.onSystemCallStart,
|
OnSystemCallStart: l.onSystemCallStart,
|
||||||
OnExit: l.OnEnd,
|
OnExit: l.OnExit,
|
||||||
OnOpcode: l.OnOpcode,
|
OnOpcode: l.OnOpcode,
|
||||||
OnFault: l.OnFault,
|
OnFault: l.OnFault,
|
||||||
}
|
}
|
||||||
|
@ -152,13 +152,6 @@ func (l *jsonLogger) OnEnter(depth int, typ byte, from common.Address, to common
|
||||||
l.encoder.Encode(frame)
|
l.encoder.Encode(frame)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (l *jsonLogger) OnEnd(depth int, output []byte, gasUsed uint64, err error, reverted bool) {
|
|
||||||
if depth > 0 {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
l.OnExit(depth, output, gasUsed, err, false)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (l *jsonLogger) OnExit(depth int, output []byte, gasUsed uint64, err error, reverted bool) {
|
func (l *jsonLogger) OnExit(depth int, output []byte, gasUsed uint64, err error, reverted bool) {
|
||||||
type endLog struct {
|
type endLog struct {
|
||||||
Output string `json:"output"`
|
Output string `json:"output"`
|
||||||
|
|
|
@ -21,7 +21,6 @@ import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"maps"
|
|
||||||
"math/big"
|
"math/big"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
@ -186,7 +185,7 @@ func (sim *simulator) processBlock(ctx context.Context, block *simBlock, header,
|
||||||
Tracer: tracer.Hooks(),
|
Tracer: tracer.Hooks(),
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
var tracingStateDB = vm.StateDB(sim.state)
|
tracingStateDB := vm.StateDB(sim.state)
|
||||||
if hooks := tracer.Hooks(); hooks != nil {
|
if hooks := tracer.Hooks(); hooks != nil {
|
||||||
tracingStateDB = state.NewHookedState(sim.state, hooks)
|
tracingStateDB = state.NewHookedState(sim.state, hooks)
|
||||||
}
|
}
|
||||||
|
@ -289,7 +288,7 @@ func (sim *simulator) activePrecompiles(base *types.Header) vm.PrecompiledContra
|
||||||
isMerge = (base.Difficulty.Sign() == 0)
|
isMerge = (base.Difficulty.Sign() == 0)
|
||||||
rules = sim.chainConfig.Rules(base.Number, isMerge, base.Time)
|
rules = sim.chainConfig.Rules(base.Number, isMerge, base.Time)
|
||||||
)
|
)
|
||||||
return maps.Clone(vm.ActivePrecompiledContracts(rules))
|
return vm.ActivePrecompiledContracts(rules)
|
||||||
}
|
}
|
||||||
|
|
||||||
// sanitizeChain checks the chain integrity. Specifically it checks that
|
// sanitizeChain checks the chain integrity. Specifically it checks that
|
||||||
|
|
|
@ -6,15 +6,16 @@
|
||||||
package metrics
|
package metrics
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"flag"
|
||||||
|
"io"
|
||||||
"os"
|
"os"
|
||||||
"runtime/metrics"
|
"runtime/metrics"
|
||||||
"runtime/pprof"
|
"runtime/pprof"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
|
||||||
"syscall"
|
"syscall"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/naoina/toml"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Enabled is checked by the constructor functions for all of the
|
// Enabled is checked by the constructor functions for all of the
|
||||||
|
@ -24,34 +25,63 @@ import (
|
||||||
// for less cluttered pprof profiles.
|
// for less cluttered pprof profiles.
|
||||||
var Enabled = false
|
var Enabled = false
|
||||||
|
|
||||||
// enablerFlags is the CLI flag names to use to enable metrics collections.
|
|
||||||
var enablerFlags = []string{"metrics"}
|
|
||||||
|
|
||||||
// enablerEnvVars is the env var names to use to enable metrics collections.
|
|
||||||
var enablerEnvVars = []string{"GETH_METRICS"}
|
|
||||||
|
|
||||||
// init enables or disables the metrics system. Since we need this to run before
|
// init enables or disables the metrics system. Since we need this to run before
|
||||||
// any other code gets to create meters and timers, we'll actually do an ugly hack
|
// any other code gets to create meters and timers, we'll actually do an ugly hack
|
||||||
// and peek into the command line args for the metrics flag.
|
// and peek into the command line args for the metrics flag.
|
||||||
func init() {
|
func init() {
|
||||||
for _, enabler := range enablerEnvVars {
|
if val, found := syscall.Getenv("GETH_METRICS"); found && !Enabled {
|
||||||
if val, found := syscall.Getenv(enabler); found && !Enabled {
|
if enable, _ := strconv.ParseBool(val); enable { // ignore error, flag parser will choke on it later
|
||||||
if enable, _ := strconv.ParseBool(val); enable { // ignore error, flag parser will choke on it later
|
Enabled = true
|
||||||
log.Info("Enabling metrics collection")
|
|
||||||
Enabled = true
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for _, arg := range os.Args {
|
|
||||||
flag := strings.TrimLeft(arg, "-")
|
|
||||||
|
|
||||||
for _, enabler := range enablerFlags {
|
fs := flag.NewFlagSet("", flag.ContinueOnError)
|
||||||
if !Enabled && flag == enabler {
|
fs.SetOutput(io.Discard)
|
||||||
log.Info("Enabling metrics collection")
|
fs.Bool("metrics", false, "")
|
||||||
Enabled = true
|
fs.String("config", "", "")
|
||||||
}
|
|
||||||
|
// The flag package will quit parsing immediately if it encounters a flag that
|
||||||
|
// was not declared to it ahead of time. We could be fancy and try to look
|
||||||
|
// through the args to find the ones we care about, but then we'd need to
|
||||||
|
// handle the various ways that flags can be defined -- which was the point of
|
||||||
|
// using the flags package in the first place! So instead, let's chop off the
|
||||||
|
// first element in the args list each time a parse fails. This way we will
|
||||||
|
// eventually parse every arg and get the ones we care about.
|
||||||
|
for i := range os.Args[1:] {
|
||||||
|
if err := fs.Parse(os.Args[i+1:]); err == nil {
|
||||||
|
break
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Now visit the flags we defined which are present in the args and see if we
|
||||||
|
// should enable metrics.
|
||||||
|
fs.Visit(func(f *flag.Flag) {
|
||||||
|
g, ok := f.Value.(flag.Getter)
|
||||||
|
if !ok {
|
||||||
|
// Don't expect this to fail, but skip over just in case it does.
|
||||||
|
return
|
||||||
|
}
|
||||||
|
switch f.Name {
|
||||||
|
case "metrics":
|
||||||
|
Enabled = g.Get().(bool)
|
||||||
|
case "config":
|
||||||
|
data, err := os.ReadFile(g.Get().(string))
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
var cfg map[string]map[string]any
|
||||||
|
if err := toml.Unmarshal(data, &cfg); err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
// A config file is definitely being used and was parsed correct. Let's
|
||||||
|
// try to peek inside and see if metrics are listed as enabled.
|
||||||
|
if m, ok := cfg["Metrics"]; ok {
|
||||||
|
if v, ok := m["Enabled"].(bool); ok && v {
|
||||||
|
Enabled = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
var threadCreateProfile = pprof.Lookup("threadcreate")
|
var threadCreateProfile = pprof.Lookup("threadcreate")
|
||||||
|
|
|
@ -676,7 +676,7 @@ func (typedData *TypedData) EncodePrimitiveValue(encType string, encValue interf
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
return math.U256Bytes(b), nil
|
return math.U256Bytes(new(big.Int).Set(b)), nil
|
||||||
}
|
}
|
||||||
return nil, fmt.Errorf("unrecognized type '%s'", encType)
|
return nil, fmt.Errorf("unrecognized type '%s'", encType)
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue