Code Monkey home page Code Monkey logo

erigon's Introduction

Erigon

Documentation: erigon.gitbook.io

Erigon is an implementation of Ethereum (execution layer with embeddable consensus layer), on the efficiency frontier. Archive Node by default.


Build status Coverage

Disclaimer: this software is currently a tech preview. We will do our best to keep it stable and make no breaking changes but we don't guarantee anything. Things can and will break.

Important defaults: Erigon is an Archive Node by default (to remove history see: --prune flags in erigon --help). We don't allow change this flag after first start.

In-depth links are marked by the microscope sign (🔬)

System Requirements

  • For an Archive node of Ethereum Mainnet we recommend >=3.5TB storage space: 2.3TiB state (as of March 2024), 643GiB snapshots (can symlink or mount folder <datadir>/snapshots to another disk), 200GB temp files (can symlink or mount folder <datadir>/temp to another disk). Ethereum Mainnet Full node (see Pruned Node): 1.1TiB not including temp files (June 2024).

  • Gnosis Chain Archive: 1.7TiB (March 2024). Gnosis Chain Full node (see Pruned Node): 300GiB (June 2024).

  • Polygon Mainnet Archive: 8.5TiB (December 2023). Polygon Mainnet Full node (see Pruned Node) with --prune.*.older 15768000: 5.1Tb (September 2023).

SSD or NVMe. Do not recommend HDD - on HDD Erigon will always stay N blocks behind chain tip, but not fall behind. Bear in mind that SSD performance deteriorates when close to capacity.

RAM: >=16GB, 64-bit architecture.

Golang version >= 1.21; GCC 10+ or Clang; On Linux: kernel > v4

🔬 more details on disk storage here and here.

Usage

Getting Started

For building the latest release (this will be suitable for most users just wanting to run a node):

git clone --branch release/<x.xx> --single-branch https://github.com/erigontech/erigon.git
cd erigon
make erigon
./build/bin/erigon

You can check the list of releases for release notes.

For building the bleeding edge development branch:

git clone --recurse-submodules https://github.com/erigontech/erigon.git
cd erigon
git checkout main
make erigon
./build/bin/erigon

Default --snapshots for mainnet, gnosis, chiado. Other networks now have default --snapshots=false. Increase download speed by flag --torrent.download.rate=20mb. 🔬 See Downloader docs

Use --datadir to choose where to store data.

Use --chain=gnosis for Gnosis Chain, --chain=bor-mainnet for Polygon Mainnet, and --chain=amoy for Polygon Amoy. For Gnosis Chain you need a Consensus Layer client alongside Erigon (https://docs.gnosischain.com/node/manual/beacon).

Running make help will list and describe the convenience commands available in the Makefile.

Datadir structure

  • chaindata: recent blocks, state, recent state history. low-latency disk recommended.
  • snapshots: old blocks, old state history. can symlink/mount it to cheaper disk. mostly immutable. must have ~100gb free space (for merge recent files to bigger one).
  • temp: can grow to ~100gb, but usually empty. can symlink/mount it to cheaper disk.
  • txpool: pending transactions. safe to remove.
  • nodes: p2p peers. safe to remove.

Logging

Flags:

  • verbosity
  • log.console.verbosity (overriding alias for verbosity)
  • log.json
  • log.console.json (alias for log.json)
  • log.dir.path
  • log.dir.prefix
  • log.dir.verbosity
  • log.dir.json

In order to log only to the stdout/stderr the --verbosity (or log.console.verbosity) flag can be used to supply an int value specifying the highest output log level:

  LvlCrit = 0
  LvlError = 1
  LvlWarn = 2
  LvlInfo = 3
  LvlDebug = 4
  LvlTrace = 5

To set an output dir for logs to be collected on disk, please set --log.dir.path If you want to change the filename produced from erigon you should also set the --log.dir.prefix flag to an alternate name. The flag --log.dir.verbosity is also available to control the verbosity of this logging, with the same int value as above, or the string value e.g. ' debug' or 'info'. Default verbosity is 'debug' (4), for disk logging.

Log format can be set to json by the use of the boolean flags log.json or log.console.json, or for the disk output --log.dir.json.

Modularity

Erigon by default is "all in one binary" solution, but it's possible start TxPool as separated processes. Same true about: JSON RPC layer (RPCDaemon), p2p layer (Sentry), history download layer (Downloader), consensus. Don't start services as separated processes unless you have clear reason for it: resource limiting, scale, replace by your own implementation, security. How to start Erigon's services as separated processes, see in docker-compose.yml.

Embedded Consensus Layer

On Ethereum Mainnet and Sepolia, the Engine API can be disabled in favour of the Erigon native Embedded Consensus Layer. If you want to use the internal Consensus Layer, run Erigon with flag --internalcl. Warning: Staking (block production) is not possible with the embedded CL.

Testnets

If you would like to give Erigon a try, but do not have spare 2TB on your drive, a good option is to start syncing one of the public testnets, Sepolia. It syncs much quicker, and does not take so much disk space:

git clone --recurse-submodules -j8 https://github.com/erigontech/erigon.git
cd erigon
make erigon
./build/bin/erigon --datadir=<your_datadir> --chain=sepolia

Please note the --datadir option that allows you to store Erigon files in a non-default location, in this example, in sepolia subdirectory of the current directory. Name of the directory --datadir does not have to match the name of the chain in --chain.

Block Production (PoW Miner or PoS Validator)

Disclaimer: Not supported/tested for Gnosis Chain and Polygon Network (In Progress)

Support only remote-miners.

  • To enable, add --mine --miner.etherbase=... or --mine --miner.miner.sigkey=... flags.
  • Other supported options: --miner.extradata, --miner.notify, --miner.gaslimit, --miner.gasprice , --miner.gastarget
  • JSON-RPC supports methods: eth_coinbase , eth_hashrate, eth_mining, eth_getWork, eth_submitWork, eth_submitHashrate
  • JSON-RPC supports websocket methods: newPendingTransaction

🔬 Detailed explanation is here.

Windows

Windows users may run erigon in 3 possible ways:

  • Build executable binaries natively for Windows using provided wmake.ps1 PowerShell script. Usage syntax is the same as make command so you have to run .\wmake.ps1 [-target] <targetname>. Example: .\wmake.ps1 erigon builds erigon executable. All binaries are placed in .\build\bin\ subfolder. There are some requirements for a successful native build on windows :

    • Git for Windows must be installed. If you're cloning this repository is very likely you already have it
    • GO Programming Language must be installed. Minimum required version is 1.21
    • GNU CC Compiler at least version 13 (is highly suggested that you install chocolatey package manager - see following point)
    • If you need to build MDBX tools (i.e. .\wmake.ps1 db-tools) then Chocolatey package manager for Windows must be installed. By Chocolatey you need to install the following components : cmake, make, mingw by choco install cmake make mingw. Make sure Windows System "Path" variable has: C:\ProgramData\chocolatey\lib\mingw\tools\install\mingw64\bin

    Important note about Anti-Viruses During MinGW's compiler detection phase some temporary executables are generated to test compiler capabilities. It's been reported some anti-virus programs detect those files as possibly infected by Win64/Kryptic.CIS trojan horse (or a variant of it). Although those are false positives we have no control over 100+ vendors of security products for Windows and their respective detection algorithms and we understand this might make your experience with Windows builds uncomfortable. To workaround the issue you might either set exclusions for your antivirus specifically for build\bin\mdbx\CMakeFiles sub-folder of the cloned repo or you can run erigon using the following other two options

  • Use Docker : see docker-compose.yml

  • Use WSL (Windows Subsystem for Linux) strictly on version 2. Under this option you can build Erigon just as you would on a regular Linux distribution. You can point your data also to any of the mounted Windows partitions ( eg. /mnt/c/[...], /mnt/d/[...] etc) but in such case be advised performance is impacted: this is due to the fact those mount points use DrvFS which is a network file system and, additionally, MDBX locks the db for exclusive access which implies only one process at a time can access data. This has consequences on the running of rpcdaemon which has to be configured as Remote DB even if it is executed on the very same computer. If instead your data is hosted on the native Linux filesystem non limitations apply. Please also note the default WSL2 environment has its own IP address which does not match the one of the network interface of Windows host: take this into account when configuring NAT for port 30303 on your router.

Using TOML or YAML Config Files

You can set Erigon flags through a YAML or TOML configuration file with the flag --config. The flags set in the configuration file can be overwritten by writing the flags directly on Erigon command line

Example

./build/bin/erigon --config ./config.yaml --chain=sepolia

Assuming we have chain : "mainnet" in our configuration file, by adding --chain=sepolia allows the overwrite of the flag inside of the yaml configuration file and sets the chain to sepolia

TOML

Example of setting up TOML config file

datadir = 'your datadir'
port = 1111
chain = "mainnet"
http = true
"private.api.addr"="localhost:9090"

"http.api" = ["eth","debug","net"]

YAML

Example of setting up a YAML config file

datadir : 'your datadir'
port : 1111
chain : "mainnet"
http : true
private.api.addr : "localhost:9090"

http.api : ["eth","debug","net"]

Beacon Chain (Consensus Layer)

Erigon can be used as an Execution Layer (EL) for Consensus Layer clients (CL). Default configuration is OK.

If your CL client is on a different device, add --authrpc.addr 0.0.0.0 (Engine API listens on localhost by default) as well as --authrpc.vhosts <CL host> where <CL host> is your source host or any.

In order to establish a secure connection between the Consensus Layer and the Execution Layer, a JWT secret key is automatically generated.

The JWT secret key will be present in the datadir by default under the name of jwt.hex and its path can be specified with the flag --authrpc.jwtsecret.

This piece of info needs to be specified in the Consensus Layer as well in order to establish connection successfully. More information can be found here.

Once Erigon is running, you need to point your CL client to <erigon address>:8551, where <erigon address> is either localhost or the IP address of the device running Erigon, and also point to the JWT secret path created by Erigon.

Caplin

Caplin is a full-fledged validating Consensus Client like Prysm, Lighthouse, Teku, Nimbus and Lodestar. Its goal is:

  • provide better stability
  • Validation of the chain
  • Stay in sync
  • keep the execution of blocks on chain tip
  • serve the Beacon API using a fast and compact data model alongside low CPU and memory usage.

The main reason why developed a new Consensus Layer is to experiment with the possible benefits that could come with it. For example, The Engine API does not work well with Erigon. The Engine API sends data one block at a time, which does not suit how Erigon works. Erigon is designed to handle many blocks simultaneously and needs to sort and process data efficiently. Therefore, it would be better for Erigon to handle the blocks independently instead of relying on the Engine API.

Caplin's Usage.

Caplin is be enabled by default. to disable it and enable the Engine API, use the --externalcl flag. from that point on, an external Consensus Layer will not be need anymore.

Caplin also has an archivial mode for historical states and blocks. it can be enabled through the --caplin.archive flag. In order to enable the caplin's Beacon API, the flag --beacon.api=<namespaces> must be added. e.g: --beacon.api=beacon,builder,config,debug,node,validator,lighthouse will enable all endpoints. **NOTE: Caplin is not staking-ready so aggregation endpoints are still to be implemented. Additionally enabling the Beacon API will lead to a 6 GB higher RAM usage.

Multiple Instances / One Machine

Define 6 flags to avoid conflicts: --datadir --port --http.port --authrpc.port --torrent.port --private.api.addr. Example of multiple chains on the same machine:

# mainnet
./build/bin/erigon --datadir="<your_mainnet_data_path>" --chain=mainnet --port=30303 --http.port=8545 --authrpc.port=8551 --torrent.port=42069 --private.api.addr=127.0.0.1:9090 --http --ws --http.api=eth,debug,net,trace,web3,erigon


# sepolia
./build/bin/erigon --datadir="<your_sepolia_data_path>" --chain=sepolia --port=30304 --http.port=8546 --authrpc.port=8552 --torrent.port=42068 --private.api.addr=127.0.0.1:9091 --http --ws --http.api=eth,debug,net,trace,web3,erigon

Quote your path if it has spaces.

Dev Chain

🔬 Detailed explanation is DEV_CHAIN.

Key features

🔬 See more detailed overview of functionality and current limitations. It is being updated on recurring basis.

More Efficient State Storage

Flat KV storage. Erigon uses a key-value database and storing accounts and storage in a simple way.

🔬 See our detailed DB walkthrough here.

Preprocessing. For some operations, Erigon uses temporary files to preprocess data before inserting it into the main DB. That reduces write amplification and DB inserts are orders of magnitude quicker.

🔬 See our detailed ETL explanation here.

Plain state.

Single accounts/state trie. Erigon uses a single Merkle trie for both accounts and the storage.

Faster Initial Sync

Erigon uses a rearchitected full sync algorithm from Go-Ethereum that is split into "stages".

🔬 See more detailed explanation in the Staged Sync Readme

It uses the same network primitives and is compatible with regular go-ethereum nodes that are using full sync, you do not need any special sync capabilities for Erigon to sync.

When reimagining the full sync, with focus on batching data together and minimize DB overwrites. That makes it possible to sync Ethereum mainnet in under 2 days if you have a fast enough network connection and an SSD drive.

Examples of stages are:

  • Downloading headers;

  • Downloading block bodies;

  • Recovering senders' addresses;

  • Executing blocks;

  • Validating root hashes and building intermediate hashes for the state Merkle trie;

  • [...]

JSON-RPC daemon

Most of Erigon's components (txpool, rpcdaemon, snapshots downloader, sentry, ...) can work inside Erigon and as independent process.

To enable built-in RPC server: --http and --ws (sharing same port with http)

Run RPCDaemon as separated process: this daemon can use local DB (with running Erigon or on snapshot of a database) or remote DB (run on another server). 🔬 See RPC-Daemon docs

For remote DB

This works regardless of whether RPC daemon is on the same computer with Erigon, or on a different one. They use TPC socket connection to pass data between them. To use this mode, run Erigon in one terminal window

make erigon
./build/bin/erigon --private.api.addr=localhost:9090 --http=false
make rpcdaemon
./build/bin/rpcdaemon --private.api.addr=localhost:9090 --http.api=eth,erigon,web3,net,debug,trace,txpool

gRPC ports

9090 erigon, 9091 sentry, 9092 consensus engine, 9093 torrent downloader, 9094 transactions pool

Supported JSON-RPC calls (eth, debug , net, web3):

For a details on the implementation status of each command, see this table.

Run all components by docker-compose

Docker allows for building and running Erigon via containers. This alleviates the need for installing build dependencies onto the host OS.

Optional: Setup dedicated user

User UID/GID need to be synchronized between the host OS and container so files are written with correct permission.

You may wish to setup a dedicated user/group on the host OS, in which case the following make targets are available.

# create "erigon" user
make user_linux
# or
make user_macos

Environment Variables

There is a .env.example file in the root of the repo.

  • DOCKER_UID - The UID of the docker user
  • DOCKER_GID - The GID of the docker user
  • XDG_DATA_HOME - The data directory which will be mounted to the docker containers

If not specified, the UID/GID will use the current user.

A good choice for XDG_DATA_HOME is to use the ~erigon/.ethereum directory created by helper targets make user_linux or make user_macos.

Check: Permissions

In all cases, XDG_DATA_HOME (specified or default) must be writeable by the user UID/GID in docker, which will be determined by the DOCKER_UID and DOCKER_GID at build time.

If a build or service startup is failing due to permissions, check that all the directories, UID, and GID controlled by these environment variables are correct.

Run

Next command starts: Erigon on port 30303, rpcdaemon on port 8545, prometheus on port 9090, and grafana on port 3000.

#
# Will mount ~/.local/share/erigon to /home/erigon/.local/share/erigon inside container
#
make docker-compose

#
# or
#
# if you want to use a custom data directory
# or, if you want to use different uid/gid for a dedicated user
#
# To solve this, pass in the uid/gid parameters into the container.
#
# DOCKER_UID: the user id
# DOCKER_GID: the group id
# XDG_DATA_HOME: the data directory (default: ~/.local/share)
#
# Note: /preferred/data/folder must be read/writeable on host OS by user with UID/GID given
#       if you followed above instructions
#
# Note: uid/gid syntax below will automatically use uid/gid of running user so this syntax
#       is intended to be run via the dedicated user setup earlier
#
DOCKER_UID=$(id -u) DOCKER_GID=$(id -g) XDG_DATA_HOME=/preferred/data/folder DOCKER_BUILDKIT=1 COMPOSE_DOCKER_CLI_BUILD=1 make docker-compose

#
# if you want to run the docker, but you are not logged in as the $ERIGON_USER
# then you'll need to adjust the syntax above to grab the correct uid/gid
#
# To run the command via another user, use
#
ERIGON_USER=erigon
sudo -u ${ERIGON_USER} DOCKER_UID=$(id -u ${ERIGON_USER}) DOCKER_GID=$(id -g ${ERIGON_USER}) XDG_DATA_HOME=~${ERIGON_USER}/.ethereum DOCKER_BUILDKIT=1 COMPOSE_DOCKER_CLI_BUILD=1 make docker-compose

Makefile creates the initial directories for erigon, prometheus and grafana. The PID namespace is shared between erigon and rpcdaemon which is required to open Erigon's DB from another process (RPCDaemon local-mode). See: https://github.com/erigontech/erigon/pull/2392/files

If your docker installation requires the docker daemon to run as root (which is by default), you will need to prefix the command above with sudo. However, it is sometimes recommended running docker (and therefore its containers) as a non-root user for security reasons. For more information about how to do this, refer to this article.

Windows support for docker-compose is not ready yet. Please help us with .ps1 port.

Grafana dashboard

docker compose up prometheus grafana, detailed docs.

old data

Disabled by default. To enable see ./build/bin/erigon --help for flags --prune

Documentation

The ./docs directory includes a lot of useful but outdated documentation. For code located in the ./cmd directory, their respective documentation can be found in ./cmd/*/README.md. A more recent collation of developments and happenings in Erigon can be found in the Erigon Blog.

FAQ

How much RAM do I need

  • Baseline (ext4 SSD): 16Gb RAM sync takes 6 days, 32Gb - 5 days, 64Gb - 4 days
  • +1 day on "zfs compression=off". +2 days on "zfs compression=on" (2x compression ratio). +3 days on btrfs.
  • -1 day on NVMe

Detailed explanation: ./docs/programmers_guide/db_faq.md

Default Ports and Firewalls

erigon ports

Component Port Protocol Purpose Should Expose
engine 9090 TCP gRPC Server Private
engine 42069 TCP & UDP Snap sync (Bittorrent) Public
engine 8551 TCP Engine API (JWT auth) Private
sentry 30303 TCP & UDP eth/68 peering Public
sentry 30304 TCP & UDP eth/67 peering Public
sentry 9091 TCP incoming gRPC Connections Private
rpcdaemon 8545 TCP HTTP & WebSockets & GraphQL Private

Typically, 30303 and 30304 are exposed to the internet to allow incoming peering connections. 9090 is exposed only internally for rpcdaemon or other connections, (e.g. rpcdaemon -> erigon). Port 8551 (JWT authenticated) is exposed only internally for Engine API JSON-RPC queries from the Consensus Layer node.

caplin ports

Component Port Protocol Purpose Should Expose
sentinel 4000 UDP Peering Public
sentinel 4001 TCP Peering Public

In order to configure the ports, use:

   --caplin.discovery.addr value                                                    Address for Caplin DISCV5 protocol (default: "127.0.0.1")
   --caplin.discovery.port value                                                    Port for Caplin DISCV5 protocol (default: 4000)
   --caplin.discovery.tcpport value                                                 TCP Port for Caplin DISCV5 protocol (default: 4001)

beaconAPI ports

Component Port Protocol Purpose Should Expose
REST 5555 TCP REST Public

shared ports

Component Port Protocol Purpose Should Expose
all 6060 TCP pprof Private
all 6060 TCP metrics Private

Optional flags can be enabled that enable pprof or metrics (or both) - however, they both run on 6060 by default, so

you'll have to change one if you want to run both at the same time. use --help with the binary for more info.

other ports

Reserved for future use: gRPC ports: 9092 consensus engine, 9093 snapshot downloader, 9094 TxPool

Hetzner expecting strict firewall rules

0.0.0.0/8             "This" Network             RFC 1122, Section 3.2.1.3
10.0.0.0/8            Private-Use Networks       RFC 1918
100.64.0.0/10         Carrier-Grade NAT (CGN)    RFC 6598, Section 7
127.16.0.0/12         Private-Use Networks       RFC 1918
169.254.0.0/16        Link Local                 RFC 3927
172.16.0.0/12         Private-Use Networks       RFC 1918
192.0.0.0/24          IETF Protocol Assignments  RFC 5736
192.0.2.0/24          TEST-NET-1                 RFC 5737
192.88.99.0/24        6to4 Relay Anycast         RFC 3068
192.168.0.0/16        Private-Use Networks       RFC 1918
198.18.0.0/15         Network Interconnect
Device Benchmark Testing   RFC 2544
198.51.100.0/24       TEST-NET-2                 RFC 5737
203.0.113.0/24        TEST-NET-3                 RFC 5737
224.0.0.0/4           Multicast                  RFC 3171
240.0.0.0/4           Reserved for Future Use    RFC 1112, Section 4
255.255.255.255/32    Limited Broadcast          RFC 919, Section 7
RFC 922, Section 7

Same in IpTables syntax

How to run erigon as a separate user? (e.g. as a systemd daemon)

Running erigon from build/bin as a separate user might produce an error:

error while loading shared libraries: libsilkworm_capi.so: cannot open shared object file: No such file or directory

The library needs to be installed for another user using make DIST=<path> install. You could use $HOME/erigon or /opt/erigon as the installation path, for example:

make DIST=/opt/erigon install

and then run /opt/erigon/erigon.

How to get diagnostic for bug report?

  • Get stack trace: kill -SIGUSR1 <pid>, get trace and stop: kill -6 <pid>
  • Get CPU profiling: add --pprof flag run go tool pprof -png http://127.0.0.1:6060/debug/pprof/profile\?seconds\=20 > cpu.png
  • Get RAM profiling: add --pprof flag run go tool pprof -inuse_space -png http://127.0.0.1:6060/debug/pprof/heap > mem.png

How to run local devnet?

🔬 Detailed explanation is here.

Docker permissions error

Docker uses user erigon with UID/GID 1000 (for security reasons). You can see this user being created in the Dockerfile. Can fix by giving a host's user ownership of the folder, where the host's user UID/GID is the same as the docker's user UID/GID (1000). More details in post

How to run public RPC api

  • --txpool.nolocals=true
  • don't add admin in --http.api list
  • to increase throughput may need increase/decrease: --db.read.concurrency, --rpc.batch.concurrency, --rpc.batch.limit

Run RaspberyPI

https://github.com/mathMakesArt/Erigon-on-RPi-4

How to change db pagesize

post

Getting in touch

Erigon Discord Server

The main discussions are happening on our Discord server. To get an invite, send an email to bloxster [at] proton.me with your name, occupation, a brief explanation of why you want to join the Discord, and how you heard about Erigon.

Reporting security issues/concerns

Send an email to security [at] torquem.ch.

Known issues

htop shows incorrect memory usage

Erigon's internal DB (MDBX) using MemoryMap - when OS does manage all read, write, cache operations instead of Application (linux , windows)

htop on column res shows memory of "App + OS used to hold page cache for given App", but it's not informative, because if htop says that app using 90% of memory you still can run 3 more instances of app on the same machine - because most of that 90% is "OS pages cache". OS automatically frees this cache any time it needs memory. Smaller "page cache size" may not impact performance of Erigon at all.

Next tools show correct memory usage of Erigon:

  • vmmap -summary PID | grep -i "Physical footprint". Without grep you can see details

    • section MALLOC ZONE column Resident Size shows App memory usage, section REGION TYPE column Resident Size shows OS pages cache size.
  • Prometheus dashboard shows memory of Go app without OS pages cache (make prometheus, open in browser localhost:3000, credentials admin/admin)

  • cat /proc/<PID>/smaps

    Erigon uses ~4Gb of RAM during genesis sync and ~1Gb during normal work. OS pages cache can utilize unlimited amount of memory.

    Warning: Multiple instances of Erigon on same machine will touch Disk concurrently, it impacts performance - one of main Erigon optimisations: "reduce Disk random access". "Blocks Execution stage" still does many random reads - this is reason why it's slowest stage. We do not recommend running multiple genesis syncs on same Disk. If genesis sync passed, then it's fine to run multiple Erigon instances on same Disk.

Blocks Execution is slow on cloud-network-drives

Please read #1516 (comment) In short: network-disks are bad for blocks execution - because blocks execution reading data from db non-parallel non-batched way.

Filesystem's background features are expensive

For example: btrfs's autodefrag option - may increase write IO 100x times

Gnome Tracker can kill Erigon

Gnome Tracker - detecting miners and kill them.

the --mount option requires BuildKit error

For anyone else that was getting the BuildKit error when trying to start Erigon the old way you can use the below...

XDG_DATA_HOME=/preferred/data/folder DOCKER_BUILDKIT=1 COMPOSE_DOCKER_CLI_BUILD=1 make docker-compose

Erigon3 user's guide

Git branch main. Just start erigon as you usually do.

RAM requirement is higher: 32gb and better 64gb. We will work on this topic a bit later.

Golang 1.21

Almost all RPC methods are implemented - if something doesn't work - just drop it on our head.

Supported networks: all.

E3 changes from E2:

  • Sync from scratch doesn't require re-exec all history. Latest state and it's history are in snapshots - can download.
  • ExecutionStage - now including many E2 stages: stage_hash_state, stage_trie, stage_log_index, stage_history_index, stage_trace_index
  • E3 can execute 1 historical transaction - without executing it's block - because history/indices have transaction-granularity, instead of block-granularity.
  • E3 doesn't store Logs (aka Receipts) - it always re-executing historical txn (but it's cheaper then in E2 - see point above). Known perf issues: #10747
  • --sync.loop.block.limit is enabled by default. (Default: 5_000. Set --sync.loop.block.limit=10_000 --batchSize=2g to increase sync speed on good hardware).
  • datadir/chaindata is small now - to prevent it's grow: we recommend set --batchSize <= 2G. And it's fine to rm -rf chaindata
  • can symlink/mount latest state to fast drive and history to cheap drive
  • Archive Node is default. Full Node: --prune.mode=full, Minimal Node (EIP-4444): --prune.mode=minimal

Known Problems of E3:

  • don't rm -rf downloader - it will cause re-downloading of files: #10976

E3 datadir structure

datadir        
    chaindata   # "Recently-updated Latest State" and "Recent History"
    snapshots   
        domain    # Latest State: link to fast disk
        history   # Historical values 
        idx       # InvertedIndices: can search/filtering/union/intersect them - to find historical data. like eth_getLogs or trace_transaction
        accessors # Additional (generated) indices of history - have "random-touch" read-pattern. They can serve only `Get` requests (no search/filters).
    temp # buffers to sort data >> RAM. sequential-buffered IO - is slow-disk-friendly
   
# There is 4 domains: account, storage, code, commitment 

E3 can store state on fast disk and history on cheap disk

If you can afford store datadir on 1 nvme-raid - great. If can't - it's possible to store history on cheap drive.

# place (or ln -s) `datadir` on slow disk. link some sub-folders to fast disk.
# Example: what need link to fast disk to speedup execution
datadir        
    chaindata   # link to fast disk
    snapshots   
        domain    # link to fast disk
        history   
        idx       
        accessors 
    temp   

# Example: how to speedup history access: 
#   - go step-by-step - first try store `accessors` on fast disk
#   - if speed is not good enough: `idx`
#   - if still not enough: `history` 

E3 datadir size

# eth-mainnet - archive - April 2024

du -hsc /erigon/* 
6G  	/erigon/caplin
50G 	/erigon/chaindata
1.8T	/erigon/snapshots
1.9T	total

du -hsc /erigon/snapshots/* 
100G 	/erigon/snapshots/accessor
240G	/erigon/snapshots/domain
260G	/erigon/snapshots/history
410G	/erigon/snapshots/idx
1.7T	/erigon/snapshots
# bor-mainnet - archive - Jun 2024

du -hsc /erigon/* 

160M	/erigon/bor
50G 	/erigon/chaindata
3.7T	/erigon/snapshots
3.8T	total

du -hsc /erigon/snapshots/* 
260G	/erigon-data/snapshots/accessor
850G	/erigon-data/snapshots/domain
650G	/erigon-data/snapshots/history
1.4T	/erigon-data/snapshots/idx
4.1T	/erigon/snapshots

E3 other perf trics

  • --sync.loop.block.limit=10_000 --batchSize=2g - likely will help for sync speed.
  • on cloud-drives (good throughput, bad latency) - can enable OS's brain to pre-fetch: SNAPSHOT_MADV_RND=false
  • can lock latest state in RAM - to prevent from eviction (node may face high historical RPC traffic without impacting Chain-Tip perf):
vmtouch -vdlw /mnt/erigon/snapshots/domain/*bt
ls /mnt/erigon/snapshots/domain/*.kv | parallel vmtouch -vdlw

# if it failing with "can't allocate memory", try: 
sync && sudo sysctl vm.drop_caches=3
echo 1 > /proc/sys/vm/compact_memory

erigon's People

Contributors

alexeyakhunov avatar askalexsharov avatar awskii avatar b00ris avatar bas-vk avatar battlmonstr avatar cjentzsch avatar debris avatar dvovk avatar enriavil1 avatar fjl avatar gballet avatar giulio2002 avatar holiman avatar janos avatar jekamas avatar karalabe avatar mandrigin avatar mariusvanderwijden avatar mh0lt avatar nonsense avatar obscuren avatar rjl493456442 avatar somnathb1 avatar taratorio avatar tgerring avatar vorot93 avatar yperbasis avatar zelig avatar zsfelfoldi avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

erigon's Issues

Not to mutate state trie when verifying the block

When we import a block, we execute txs in it and mutate the trie. However, if the state root hash does not match, we need to undo the changes. This does not happen now. Alternatively, we need to find a way to compute the state root without mutating the trie. This would also be important for mining.

Add badgerDb as an alternative to BoltDb

Badger DB satisfies the main requirements to a database for Turbo-Geth, namely:

  1. ACID transactions. This is important for preventing database corruption due to panic, and abrupt shutdowns
  2. MVCC (Multi-Version Concurrency Control). This is important for our project of creating a read-only node (with remote over-the-network access to the database of the write-node).

The main difference between BoltDB and BadgerDB is that BoltDB uses B+trees as the data structure, whereas BadgerDB - LSM (Log Structured Merge) trees. We have reasons to believe that LSM trees can perform better with the typical workload of syncing of Turbo-Geth node.

Pruning proceeds ahead of last sync blocks

When syncing to Ropsten, where it takes a long time for sync to start, I observed that pruning has started ahead of sync:

INFO [10-16|10:46:00.988] Initialising Ethereum protocol           versions="[63 62]" network=3
INFO [10-16|10:46:01.010] Loaded most recent local header          number=0 hash=419410…ca4a2d td=1048576 age=50y6mo5d
INFO [10-16|10:46:01.011] Loaded most recent local full block      number=0 hash=419410…ca4a2d td=1048576 age=50y6mo5d
INFO [10-16|10:46:01.011] Loaded most recent local fast block      number=0 hash=419410…ca4a2d td=1048576 age=50y6mo5d
INFO [10-16|10:46:01.011] Pruner started 
INFO [10-16|10:46:01.011] Loaded local transaction journal         transactions=0 dropped=0
INFO [10-16|10:46:01.012] Regenerated local transaction journal    transactions=0 accounts=0
INFO [10-16|10:46:01.012] Initialising Firehose protocol           versions=[1]
INFO [10-16|10:46:01.088] New local node record                    seq=23 id=a3c8dc9a49dedf3b ip=127.0.0.1 udp=30303 tcp=30303
INFO [10-16|10:46:01.088] Started P2P networking                   self=enode://d3228a32673a4c6ec8556a6796bf3a4ea94d288aed9db43a98d7ba742abd296624b2c7f02bd3ad88250ce4b74a777dfdb30fef0a4cad7f484b8c748221f6c626@127.0.0.1:30303
INFO [10-16|10:46:01.091] IPC endpoint opened                      url=/Users/alexeyakhunov/Library/Ethereum/testnet/geth.ipc
INFO [10-16|10:51:01.036] Save last pruned block num               num=3000
INFO [10-16|10:56:01.028] Save last pruned block num               num=6000
INFO [10-16|11:01:01.016] Save last pruned block num               num=9000
INFO [10-16|11:06:01.007] Save last pruned block num               num=12000

The reason is underflow when computing diff in this code:

func calculateNumOfPrunedBlocks(curentBlock, lastPrunedBlock uint64, blocksBeforePruning uint64, blocksBatch uint64) (uint64, uint64, bool) {
	diff := curentBlock - lastPrunedBlock - blocksBeforePruning
	switch {
	case diff >= blocksBatch:
		return lastPrunedBlock, lastPrunedBlock + blocksBatch, true
	case diff > 0 && diff < blocksBatch:
		return lastPrunedBlock, lastPrunedBlock + diff, true
	default:
		return lastPrunedBlock, lastPrunedBlock, false
	}
}

Also there is a typo in the word curentBlock :)

Make encoding of values in the SUFFIX (changeset) bucket ordered by keys and binary-searchable

We are currently in the process of extending the values in the SUFFUX bucket to contain not just the list of keys that have changed in given block, but also previous values (before the change). This allows decoupling of rewinding from the history of state buckets hAT and hST, making history of state optional, and also simplifying the rewinding.

We should go further and make another two improvements in the way we encode the list of key-value pairs:

  1. Make the list sorted by keys. This will allow more efficient merging of subsequent change-sets without deserialising every change-set first (in rewindData).
  2. Prepend (or append) the list by the positional index, so that the binary search on the keys is possible. We expect the values become quite large (100ks - megabytes potentially). Binary search will be useful to implement historical resolver `trie.Resolver' in the way which is not coupled to the history of state buckets 'hAT' and 'hST'. The idea is to perform rewind of a sub-trie, and for that, we would need to be able to efficiently derive change-sets only for certain range of keys. Start of the range can be found by binary search and then keys can be iterated until we reach the end of the desired range.

Investigate a broken snapshot after archive sync

After having performed archive sync, I transmitted the database (around 500Gb) to another computer and tried to resume sync from it.
However, the sync very soon failed because the resolution of one of the contracts returned a wrong storageRoot.
The success of transmission has been checked by performing SHA256 on both files, and the results matched.
The plan is to compare storageRoot for all contracts with the computed values, then choose the smallest contract where the mismatch happens, and attempt to find the cause for the mismatch.

Fix TestCopy datarace

==================
WARNING: DATA RACE
Write at 0x00c0001f3558 by goroutine 7:
  github.com/ledgerwatch/turbo-geth/trie.(*Trie).insert()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/trie/trie.go:319 +0x1e08
  github.com/ledgerwatch/turbo-geth/trie.(*Trie).insert()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/trie/trie.go:380 +0x929
  github.com/ledgerwatch/turbo-geth/trie.(*Trie).insert()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/trie/trie.go:420 +0x1038
  github.com/ledgerwatch/turbo-geth/trie.(*Trie).insert()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/trie/trie.go:420 +0x1038
  github.com/ledgerwatch/turbo-geth/trie.(*Trie).Update()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/trie/trie.go:131 +0x19b
  github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).computeTrieRoots()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:551 +0xf3b
  github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).ComputeTrieRoots()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:296 +0x6f
  github.com/ledgerwatch/turbo-geth/core/state.TestCopy()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/core/state/statedb_test.go:611 +0x181e
  testing.tRunner()
      /usr/local/go/src/testing/testing.go:827 +0x162

Previous read at 0x00c0001f3558 by goroutine 16:
  github.com/ledgerwatch/turbo-geth/trie.(*Trie).NeedResolution()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/trie/trie.go:167 +0x25e
  github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).resolveAccountTouches()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:438 +0x142
  github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).computeTrieRoots()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:490 +0x19d
  github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).ComputeTrieRoots()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:296 +0x6f
  github.com/ledgerwatch/turbo-geth/core/state.TestCopy.func1()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/core/state/statedb_test.go:589 +0x1d5

Goroutine 7 (running) created at:
  testing.(*T).Run()
      /usr/local/go/src/testing/testing.go:878 +0x650
  testing.runTests.func1()
      /usr/local/go/src/testing/testing.go:1119 +0xa8
  testing.tRunner()
      /usr/local/go/src/testing/testing.go:827 +0x162
  testing.runTests()
      /usr/local/go/src/testing/testing.go:1117 +0x4ee
  testing.(*M).Run()
      /usr/local/go/src/testing/testing.go:1034 +0x2ee
  main.main()
      _testmain.go:68 +0x221

Goroutine 16 (running) created at:
  github.com/ledgerwatch/turbo-geth/core/state.TestCopy()
      /home/b00ris/go/src/github.com/ledgerwatch/turbo-geth/core/state/statedb_test.go:582 +0x1688
  testing.tRunner()
      /usr/local/go/src/testing/testing.go:827 +0x162

letter-coded sync modes instead of full, archive etc.

Some database buckets are mandatory for successful sync and block processing. For example, AT (current state of accounts), ST (current state of contract storage), SUFFIX (changesets used for rewinding). Other buckets are optional and only add value for certain users. We would like users to have flexibility to pick and choose those buckets in any combination. Users may specify the mode of operation as a combination of one-letter flags (similar to unix tar command). Here are suggested flags:

  1. r Transaction receipts. Currently turned off by default. If turned on, receipts are written to a special bucket, which is then used to server getTransactionReceipt and getLogs RPC commands. In blockchain.go this is set by EnableReceipts(true)
  2. h History of state. These are buckets hAT (history of state of accounts) and hST (history of state of contract storage). Although these buckets contain the same information as the SUFFIX (change-set) bucket, it is organised for quick access using addressHash | blockNr composite key for accounts, and addressHash | incarnation | blockNr composite key for contract storage. Bucket SUFFIX is organised for quick access for composite keys "hAT" | blockNr and "hST" | blockNr. History of state is used for answering various RPC queries related to account balances, contract storage contents at some point in the past.
  3. t Index allowing to look up block number and transaction index given a transaction hash. It is used by RPC requests such as getTransaction, because it allows retrieving the body of the transaction from the bucket with blocks
  4. p preimages for address hashes and storage key hashes. This information is needed by RPC queries such as getModifiedAccounts and getStorageRangeAt.
  5. Block headers and bodies? In order to perform reorgs, some short history of alternative blocks and headers is required.

Some examples of which modes might be useful for different use-cases:

  • Miners would prefer to turn all the flags off, they only care about the current state and occasional reorgs. Their priority - low latency of block processing, and lowering I/O would definitely help
  • Exchanges would need preimages (p) and history (h) to be able to watch for deposits. If they use logs (events) to figure out token transfers, they might also want to turn on receipts (r) for better performance
  • Block explorers might want to turn everything on
  • Dapp Web sites would want to turn on receipts (r) for better scanning of the logs (lots of dapp use logs as a substitute for storage).

Switch historical trie.Resolver from using hAT & hST bucket to using SUFFIX (changeset) bucket

Currently trie.Resolver supports historical mode, where it reads the range of keys from the history "As Of" given block number. The efficiently of such operation declines with the growth of the historical buckets.

Alternative approach, which should work more efficiently for the cases where the "As Of" point is not too far into the past, is to utilise change-sets in the SUFFIX bucket, and effectively rewind the range of the current state while resolving. This should help answer queries related to snapshot sync Red-Queen/Firehose protocol, as well as light server support in the future.

Issue #123 provides a way to perform such partial rewind efficiently because binary search on the change sets becomes possible

Fix mining code

Mining code is currently disabled, because mining logic needs special handling. Since the state in Turbo-Geth is singleton, miner has to share it with the block processing, rather than relying on the thread-safe copy. The solution should be similar to #111, essentially computing the state root of the mined block without modifying the state trie.

Unstable TestWatchNoDir

? github.com/ledgerwatch/turbo-geth/accounts/external [no test files]
--- FAIL: TestWatchNoDir (12.70s)
account_cache_test.go:132:
got []
want [{[126 245 166 19 95 31 214 160 37 147 238 220 134 156 109 65 217 52 174 248] keystore:///tmp/eth-keystore-watch-test-8097-2334823778697134485/aaa}]
FAIL
FAIL github.com/ledgerwatch/turbo-geth/accounts/keystore 20.501s

https://circleci.com/gh/ledgerwatch/turbo-geth/401

Visualisation of database tables after each change of state and highlighting of the changes

Now that we can visualise the database buckets, we would like to do this after each change of state in the initialState1 function, and highlight what changed every time. Highlighting is achieved by using highlighted parameter of the Horizontal function.
In order to highlight the differences, you would probably want to snapshot the database (using MemCopy), and then write a function that compares the current state with the snapshot.

Replace red-black trees with simple maps in ethdb.mutation

Originally, red-black trees were put there to support iteration (Walk function) over the mutation. However, this operation has been removed, and iterations are now only supported over the committed databases (and not mutation). There is no good reason to keep using red black trees instead of simple maps anymore

Misleading Log "Impossible reorg, please file an issue"

INFO [07-17|14:22:40.264] Database size=449657688064 written=22122497
INFO [07-17|14:22:45.825] Rewinding to block=8168731
Rewinding from block 8168739 to block 8168731
reorg 8168739 8168731
ERROR[07-17|14:22:46.017] Impossible reorg, please file an issue oldnum=8168731 oldhash=61753c…32cb12 newnum=8168731 newhash=61753c…32cb12
INFO [07-17|14:22:53.776] Imported new chain segment blocks=11 txs=1725 mgas=87.858 elapsed=8.131s mgasps=10.804 number=8168742 hash=2e878a…7e67f8 batch=14005 age=6m20s

Remove Copy member functions from IntraBlockState

Since there is no efficient way of copying intra-block-state (for the purpose of concurrent access, for example), it makes sense to remove Copy functions rather than leaving them incorrect, but present.
This will also illuminate the need to fix the miner code

Looking ways to collaborate with you

Hi there,

I want to collaborate with this project and I'm looking for some starting issues or ways to collaborate. I'm a programmer at Decentraland and we are interested in helping the development of storage rent and other eth 1x improvements.

I read about the state rent proposal and I understand the major points

Last week I could start syncing the node, but it would fail to validate all blocks on restart. Today it doesn't event start to download blocks. Neither in master nor in the eip-2027 branch. Probably there are different ways for testing the node that do not require to syncing it and you may help me pointing those?

Or maybe you know better how I can help you guys

Thanks,

Fix stCreate2/create2collisionStorage tests

st.fails(`^stCreate2/create2collisionStorage.json/Constantinople/0`, "work in progress")
st.fails(`^stCreate2/create2collisionStorage.json/Constantinople/1`, "work in progress")
st.fails(`^stCreate2/create2collisionStorage.json/Constantinople/1`, "work in progress")
st.fails(`^stCreate2/create2collisionStorage.json/Constantinople/2`, "work in progress")
st.fails(`^stCreate2/create2collisionStorage.json/ConstantinopleFix/0`, "work in progress")
st.fails(`^stCreate2/create2collisionStorage.json/ConstantinopleFix/1`, "work in progress")
st.fails(`^stCreate2/create2collisionStorage.json/ConstantinopleFix/2`, "work in progress")

from https://github.com/ledgerwatch/turbo-geth/blob/master/tests/state_test.go

Make Turbogeth imports compatible with go-ethereum

When rebasing to a new version of geth, a lot of conflicts happen because of different import names.

When a new import added upstream, git gets confused. Resolving these merges is trivial, but annoying and can lead to occasionally importing actual unpatched go-ethereum module.

Not Connecting to Parity Peers

DEBUG[07-04|05:45:50.848] Adding p2p peer name=Parity-Ethereum/v2.5... addr=67.110.215.92:5
6852 peers=15
DEBUG[07-04|05:45:50.848] Ethereum peer connected id=904a58a02a6ce296 conn=inbound name=Parity-Ethereum/v2.5.1-beta-adabd81-20190514/x86_64-linux-gnu/rustc1.34.1
DEBUG[07-04|05:45:50.848] Failed to read Status msg id=904a58a02a6ce296 conn=inbound err=EOF
DEBUG[07-04|05:45:50.848] Ethereum handshake failed id=904a58a02a6ce296 conn=inbound err=EOF
DEBUG[07-04|05:45:50.848] Removing p2p peer id=904a58a02a6ce296 conn=inbound duration=325.273µs peers=14 req=false err=EOF
DEBUG[07-04|05:45:50.849] Failed to send Status msg id=904a58a02a6ce296 conn=inbound err="write tcp 10.128.0.7:30303->67.110.215.92:56852: use of closed network connection"

(built from red-queen @ commit 1b2c836 )

Crash during sync in the trie cleaning

Backtrace

INFO [10-04|12:27:59.317] Memory                                   nodes=4999974 alloc=14091714 sys=25457625 numGC=287
INFO [10-04|12:27:59.317] Database                                 size=142020145152 written=570843137
INFO [10-04|12:27:59.358] Imported new chain segment               blocks=31  txs=3409  mgas=173.164  elapsed=11.556s   mgasps=14.984  number=5104347 hash=c38607…61a26
7 batch=584    age=1y7mo3w
INFO [10-04|12:28:07.370] Imported new chain segment               blocks=130 txs=14839 mgas=804.555  elapsed=8.011s    mgasps=100.422 number=5104477 hash=224281…77f2a
1 batch=84155  age=1y7mo3w
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x18 pc=0x8c090e]
goroutine 53760 [running]:
github.com/ledgerwatch/turbo-geth/trie.(*hasher).hashInternal(0xc2dbe44000, 0x0, 0x0, 0xc0f6d65e00, 0xc07a77fd60, 0x20, 0x20, 0x0, 0x45e491)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/trie/hasher.go:80 +0x8e
github.com/ledgerwatch/turbo-geth/trie.(*hasher).hash(...)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/trie/hasher.go:69
github.com/ledgerwatch/turbo-geth/trie.(*Trie).unload(0xc000ec5480, 0xc28b06e400, 0x40, 0x40, 0xc2dbe44000)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/trie/trie.go:1255 +0x4bd
github.com/ledgerwatch/turbo-geth/trie.pruneMap(0xc000ec5480, 0xc227afef70, 0xc2dbe44000, 0x41)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/trie/trie_pruning.go:160 +0x232
github.com/ledgerwatch/turbo-geth/trie.(*TriePruning).PruneToTimestamp(0xc000554c60, 0xc000ec5480, 0x4d1fcc)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/trie/trie_pruning.go:189 +0x266
github.com/ledgerwatch/turbo-geth/trie.(*TriePruning).PruneTo(0xc000554c60, 0xc000ec5480, 0x4c4b40, 0x1)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/trie/trie_pruning.go:216 +0xbf
github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).PruneTries(0xc0000e2a10, 0x216e7000)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:1043 +0x80
github.com/ledgerwatch/turbo-geth/core.(*BlockChain).insertChain(0xc000170400, 0x1303da0, 0xc187943740, 0xc540418000, 0xb79, 0xb79, 0x1, 0x0, 0x0, 0x0, ...)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/core/blockchain.go:1360 +0x25c0
github.com/ledgerwatch/turbo-geth/core.(*BlockChain).InsertChain(0xc000170400, 0xc540418000, 0xb79, 0xb79, 0x2065648, 0x0, 0x0)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/core/blockchain.go:1068 +0xe04
github.com/ledgerwatch/turbo-geth/eth/downloader.(*Downloader).importBlockResults(0xc00025a6e0, 0xc53ff4e000, 0xb79, 0xb79, 0xb79, 0x0)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/eth/downloader/downloader.go:1471 +0x538
github.com/ledgerwatch/turbo-geth/eth/downloader.(*Downloader).processFullSyncContent(0xc00025a6e0, 0xc0001086c0, 0xc0ecee5798)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/eth/downloader/downloader.go:1445 +0x79
github.com/ledgerwatch/turbo-geth/eth/downloader.(*Downloader).spawnSync.func1(0xc00025a6e0, 0xc244e00d20, 0xc43c514a10)
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/eth/downloader/downloader.go:478 +0x59
created by github.com/ledgerwatch/turbo-geth/eth/downloader.(*Downloader).spawnSync
        /home/akhounov/workspace/src/github.com/ledgerwatch/turbo-geth/eth/downloader/downloader.go:478 +0xaf
(END)

Reorganise the way contract codes are stored in the database, and then remove codeHash from hAT bucket

Presently, the contract codes are stored in the bucket CODE, and it is a simple mapping of code hashes to bytecodes. We would like to change that to a mapping of a composite key contractAddrHash | incarnation to the bytecode. That would allow us to remove codeHash component from the hAT bucket (which is the history of state of accounts), therefore reducing the size of the bucket and write I/O to maintain it. It will also allow extension of pruning to contract codes.

The downside might be potential code duplication. If such downside to be avoided, we can use the indirect mapping contractAddrHash | incarcation to the code hash, and then add some kind of reference counter to the existing mapping of codeHash to bytecode (to be able to remove it when all the contracts using such bytecode have been pruned).

Add visualisations of database tables to the visualisation of state tries

Please use the branch block_witness for reference until this PR is merged: #112

The file https://github.com/ledgerwatch/turbo-geth/blob/block_witness/cmd/pics/state.go contains the code that generates a series of illustrations used in this presentation (still needs to be turned to a documentation): https://github.com/AlexeyAkhunov/papers/blob/master/STARKs_for_Stateless_client_The_beginning.pdf

In order to run them, you can execute

make pics
./build/bin/pics -pic initial_state_1

You would need graphviz to be installed for this to work. If it worked, you will see lots of PNG files in the current directory.

After generating each picture, it would be great if we committed the changes to the database (db.Commit()), and then visualise the buckets in the database. The code that iterates over the buckets can be taken from here: https://github.com/ledgerwatch/turbo-geth/blob/block_witness/ethdb/memory_database.go (function MemCopy), but instead of inserting all the keys and values into a new (copy) database, we would try to visualise them, perhaps using the same primitives (coloured squares) as we use for the trie. Specifically, I would start with visualising the buckets AT (accounts) and ST. In order to see the correspondence between the entries in the database and the content of the state trie, the keys taken from the database need to be converted to HEX or QUAD encoding before being displayed, and perhaps also split into parts (for ST bucket).

Eventually, these new visualisations would become part of the programmer's guide: https://github.com/ledgerwatch/turbo-geth/blob/master/docs/programmers_guide/guide.md

Not to wrap values into RLP when storing the valueNode in the trie

Before we performed state trie unification and therefore created a separate leaf node type accountNode, simple value nodes and account nodes needed different amount of RLP "wrapping" (The quirk of yellow paper). Now that that their are separate, we should remove RLP wrapping from valueNode leaves. This will simplify code in some parts, and also reduce number of allocation due to wrapping/up-wrapping.
Issue #118 was a result of extra wrapping applied erroneously

Linters in CI

Eugene:

кстати, можно от таких штук прикрутить линтеров в CI. А чтоб не бороться со старым кодом, то линтеры можно запускать только для diff с текущим мастеров.

golangci, я думаю.
в make что-то типа
LATEST_COMMIT ?= $(shell git log -n 1 origin/master --pretty=format:"%H")
ifeq ($(LATEST_COMMIT),)
LATEST_COMMIT := $(shell git log -n 1 HEAD~1 --pretty=format:"%H")
endif

lint:
@echo "--> Running linter for code diff versus commit $(LATEST_COMMIT)"
@./build/bin/golangci-lint run --new-from-rev=$(LATEST_COMMIT)

lint-deps:
curl -sfL https://install.goreleaser.com/github.com/golangci/golangci-lint.sh | sh -s -- -b ./build/bin v1.16.0
;

и .golangci.yml:

run:
deadline: 10m

linters:
enable-all: true
disable:
- lll
- dupl
- gochecknoglobals
- gocritic
- maligned
- goimports

linters-settings:
govet:
check-shadowing: true
golint:
min-confidence: 0
maligned:
suggest-new: true
goconst:
min-len: 2
min-occurrences: 2
misspell:
locale: US
gocyclo:
min-complexity: 35
gofmt:
auto-fix: false

месяц назад в форк для работы с geth добавлял :)

Check and fix reorg

  1. for full sync archive node
  2. for full sync -noHistory

archive mode

INFO [09-18|15:46:37.141] Upgrading chain index                    type=bloombits percentage=96
INFO [09-18|15:46:45.351] Imported new chain segment               blocks=3002 txs=1525  mgas=247.621  elapsed=8.001s    mgasps=30.946  number=223069 hash=7aba52…bd44e0 batch=28692  age=2y9mo6d
INFO [09-18|15:46:47.851] Finished upgrading chain index           type=bloombits
INFO [09-18|15:47:01.878] Memory                                   nodes=188300 alloc=725282  sys=1321001 numGC=630
INFO [09-18|15:47:01.878] Database                                 size=1209487360 written=107163649
INFO [09-18|15:47:01.913] Imported new chain segment               blocks=3099 txs=5324  mgas=783.476  elapsed=16.562s   mgasps=47.305  number=226168 hash=4a8b75…af1fe3 batch=276    age=2y9mo6d
INFO [09-18|15:47:05.460] Imported new chain segment               blocks=2091 txs=4035  mgas=596.935  elapsed=3.546s    mgasps=168.307 number=228259 hash=2795d9…b7c1d8 batch=59592  age=2y9mo5d
INFO [09-18|15:47:19.710] Memory                                   nodes=192546 alloc=701734  sys=1456301 numGC=632
INFO [09-18|15:47:19.710] Database                                 size=1209487360 written=72851457
INFO [09-18|15:47:19.742] Upgrading chain index                    type=bloombits percentage=96
INFO [09-18|15:47:28.189] Imported new chain segment               blocks=1827 txs=3326  mgas=409.059  elapsed=8.004s    mgasps=51.101  number=230086 hash=0f8b88…09fed1 batch=35410  age=2y9mo5d
INFO [09-18|15:47:28.209] Finished upgrading chain index           type=bloombits
INFO [09-18|15:47:40.801] Memory                                   nodes=201048 alloc=697643  sys=1456301 numGC=639
INFO [09-18|15:47:40.801] Database                                 size=1249263616 written=104996865
INFO [09-18|15:47:40.802] Imported new chain segment               blocks=676  txs=3503  mgas=661.853  elapsed=12.612s   mgasps=52.475  number=230762 hash=ec974c…e25486 batch=11     age=2y9mo5d
INFO [09-18|15:47:40.955] Rewinding                                to block=228259
ERROR[09-18|15:47:44.342] Incorrect rewinding                      root=833633f1876d00cba03887e9b7bbbd11cfb583d7d0e0b70adbefef9921c263b1 expected=fe568ba6a8fb7b300774993e60465ce8ced71a583fc0b68faf56444544cebc9c
WARN [09-18|15:47:44.345] Synchronisation failed, dropping peer    peer=774a4679069c07a4 err="retrieved hash chain is invalid"
INFO [09-18|15:47:47.708] Creating IntraBlockState from latest state block=230881
WARN [09-18|15:47:47.810] Synchronisation failed, dropping peer    peer=92570a250f771747 err="retrieved hash chain is invalid"
INFO [09-18|15:53:50.693] Rewinding                                to block=230881
ERROR[09-18|15:53:50.719] Incorrect rewinding                      root=d7f66d90a145bcdee58ca1d36f2be53e731ef303aea8bc7ee8358725db857dd5 expected=fe568ba6a8fb7b300774993e60465ce8ced71a583fc0b68faf56444544cebc9c
WARN [09-18|15:53:50.719] Synchronisation failed, dropping peer    peer=190b5d40eb03cd59 err="retrieved hash chain is invalid"

and then it stuck.

Fix concurrent map writes error


goroutine 48 [running]:
runtime.throw(0x72f8b2, 0x15)
	/usr/local/go/src/runtime/panic.go:617 +0x72 fp=0xc00011a758 sp=0xc00011a728 pc=0x42f862
runtime.mapassign_fast64(0x6cffe0, 0xc000013050, 0x0, 0x5dfd9e)
	/usr/local/go/src/runtime/map_fast64.go:101 +0x35f fp=0xc00011a798 sp=0xc00011a758 pc=0x4135cf
github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).leftGeneration(...)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:686
github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).leftGeneration-fm(0x0)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:684 +0x49 fp=0xc00011a7c8 sp=0xc00011a798 pc=0x6747d9
github.com/ledgerwatch/turbo-geth/trie.(*duoNode).updateT(0xc0005269c0, 0x1, 0xc0005120d0, 0xc0005120e0)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/trie/node.go:334 +0x48 fp=0xc00011a7e0 sp=0xc00011a7c8 pc=0x5dfc68
github.com/ledgerwatch/turbo-geth/trie.(*Trie).insert(0xc000108100, 0x7a8d00, 0xc0005269c0, 0xc00082acd0, 0x41, 0x41, 0x2, 0x7a8f40, 0xc000820c20, 0xc00011ae90, ...)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/trie/trie.go:1227 +0xa43 fp=0xc00011a9e8 sp=0xc00011a7e0 pc=0x5fb103
github.com/ledgerwatch/turbo-geth/trie.(*Trie).insert(0xc000108100, 0x7a8d60, 0xc0005d6c60, 0xc00082acd0, 0x41, 0x41, 0x1, 0x7a8f40, 0xc000820c20, 0xc00011ae90, ...)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/trie/trie.go:1329 +0x633 fp=0xc00011abf0 sp=0xc00011a9e8 pc=0x5facf3
github.com/ledgerwatch/turbo-geth/trie.(*Trie).insert(0xc000108100, 0x7a8d60, 0xc000511760, 0xc00082acd0, 0x41, 0x41, 0x0, 0x7a8f40, 0xc000820c20, 0xc00011ae90, ...)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/trie/trie.go:1329 +0x633 fp=0xc00011adf8 sp=0xc00011abf0 pc=0x5facf3
github.com/ledgerwatch/turbo-geth/trie.(*Trie).Update(0xc000108100, 0xc00011b200, 0x20, 0x20, 0xc00082ac80, 0x48, 0x48, 0x1)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/trie/trie.go:808 +0x178 fp=0xc00011af30 sp=0xc00011adf8 pc=0x5f7c28
github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).computeTrieRoots(0xc000718c60, 0xc000501901, 0x0, 0x3e, 0x8060104, 0x3e, 0xffffffffffffffff)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:498 +0x1a57 fp=0xc000131c10 sp=0xc00011af30 pc=0x656977
github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).ComputeTrieRoots(0xc000718c60, 0xc000010201, 0x7a7e80, 0xc000010200, 0x0, 0x0)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:286 +0x34 fp=0xc000131c58 sp=0xc000131c10 pc=0x654b84
github.com/ledgerwatch/turbo-geth/core/state.TestCopy(0xc00010a300)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/core/state/statedb_test.go:190 +0x98f fp=0xc000131fa8 sp=0xc000131c58 pc=0x66c71f
testing.tRunner(0xc00010a300, 0x73aec0)
	/usr/local/go/src/testing/testing.go:865 +0xc0 fp=0xc000131fd0 sp=0xc000131fa8 pc=0x4e2250
runtime.goexit()
	/usr/local/go/src/runtime/asm_amd64.s:1337 +0x1 fp=0xc000131fd8 sp=0xc000131fd0 pc=0x45da11
created by testing.(*T).Run
	/usr/local/go/src/testing/testing.go:916 +0x35a

goroutine 1 [chan receive]:
testing.(*T).Run(0xc00010a300, 0x72bbd6, 0x8, 0x73aec0, 0x479101)
	/usr/local/go/src/testing/testing.go:917 +0x381
testing.runTests.func1(0xc00010a200)
	/usr/local/go/src/testing/testing.go:1157 +0x78
testing.tRunner(0xc00010a200, 0xc0000e1e30)
	/usr/local/go/src/testing/testing.go:865 +0xc0
testing.runTests(0xc00000f120, 0xbad040, 0xc, 0xc, 0x0)
	/usr/local/go/src/testing/testing.go:1155 +0x2a9
testing.(*M).Run(0xc000108180, 0x0)
	/usr/local/go/src/testing/testing.go:1072 +0x162
main.main()
	_testmain.go:64 +0x13e

goroutine 49 [runnable]:
github.com/ledgerwatch/turbo-geth/rlp.(*encbuf).encodeString(0xc00082c000, 0xc0004a867e, 0x2, 0x2)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/rlp/encode.go:201 +0x1ba
github.com/ledgerwatch/turbo-geth/rlp.writeBigInt(0xc00038c7c0, 0xc00082c000, 0x196, 0x40e801)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/rlp/encode.go:435 +0x199
github.com/ledgerwatch/turbo-geth/rlp.writeBigIntPtr(0x728320, 0xc000297e20, 0x196, 0xc00082c000, 0x728320, 0xc000297e20)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/rlp/encode.go:421 +0x10d
github.com/ledgerwatch/turbo-geth/rlp.makeStructWriter.func1(0x6fc7c0, 0xc000297e18, 0x199, 0xc00082c000, 0xc000297e18, 0x199)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/rlp/encode.go:535 +0x149
github.com/ledgerwatch/turbo-geth/rlp.makePtrWriter.func4(0x6ab120, 0xc000297e18, 0x16, 0xc00082c000, 0x0, 0x0)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/rlp/encode.go:578 +0xb1
github.com/ledgerwatch/turbo-geth/rlp.(*encbuf).encode(0xc00082c000, 0x6ab120, 0xc000297e18, 0x706b00, 0xc00082c000)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/rlp/encode.go:187 +0x135
github.com/ledgerwatch/turbo-geth/rlp.EncodeToBytes(0x6ab120, 0xc000297e18, 0x0, 0x0, 0x0, 0x0, 0x0)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/rlp/encode.go:101 +0xe4
github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).computeTrieRoots(0xc000184000, 0xc0004f2001, 0x0, 0x13, 0x7060107, 0x13, 0xffffffffffffffff)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:494 +0x19ee
github.com/ledgerwatch/turbo-geth/core/state.(*TrieDbState).ComputeTrieRoots(0xc000184000, 0xc000398001, 0x7a7e80, 0xc000398000, 0x0, 0x0)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/core/state/database.go:286 +0x34
github.com/ledgerwatch/turbo-geth/core/state.TestCopy.func1(0xc000506000, 0xc000184000, 0xc0003f6120)
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/core/state/statedb_test.go:184 +0x7e
created by github.com/ledgerwatch/turbo-geth/core/state.TestCopy
	/go/src/github.com/ledgerwatch/turbo-geth/build/_workspace/src/github.com/ledgerwatch/turbo-geth/core/state/statedb_test.go:182 +0x929

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.