# Running in production ## Logging Default logging level (`main:info,state:info,*:`) should suffice for normal operation mode. Read [this post](https://blog.cosmos.network/one-of-the-exciting-new-features-in-0-10-0-release-is-smart-log-level-flag-e2506b4ab756) for details on how to configure `log_level` config variable. Some of the modules can be found [here](./how-to-read-logs.md#list-of-modules). If you're trying to debug Tendermint or asked to provide logs with debug logging level, you can do so by running tendermint with `--log_level="*:debug"`. ## DOS Exposure and Mitigation Validators are supposed to setup [Sentry Node Architecture](https://blog.cosmos.network/tendermint-explained-bringing-bft-based-pos-to-the-public-blockchain-domain-f22e274a0fdb) to prevent Denial-of-service attacks. You can read more about it [here](../interviews/tendermint-bft.md). ### P2P The core of the Tendermint peer-to-peer system is `MConnection`. Each connection has `MaxPacketMsgPayloadSize`, which is the maximum packet size and bounded send & receive queues. One can impose restrictions on send & receive rate per connection (`SendRate`, `RecvRate`). ### RPC Endpoints returning multiple entries are limited by default to return 30 elements (100 max). See [here](./rpc.md) for more information about the RPC. Rate-limiting and authentication are another key aspects to help protect against DOS attacks. While in the future we may implement these features, for now, validators are supposed to use external tools like [NGINX](https://www.nginx.com/blog/rate-limiting-nginx/) or [traefik](https://docs.traefik.io/configuration/commons/#rate-limiting) to achieve the same things. ## Debugging Tendermint If you ever have to debug Tendermint, the first thing you should probably do is to check out the logs. See [How to read logs](./how-to-read-logs.md), where we explain what certain log statements mean. If, after skimming through the logs, things are not clear still, the next thing to try is query the /status RPC endpoint. It provides the necessary info: whenever the node is syncing or not, what height it is on, etc. ``` curl http(s)://{ip}:{rpcPort}/status ``` `dump_consensus_state` will give you a detailed overview of the consensus state (proposer, lastest validators, peers states). From it, you should be able to figure out why, for example, the network had halted. ``` curl http(s)://{ip}:{rpcPort}/dump_consensus_state ``` There is a reduced version of this endpoint - `consensus_state`, which returns just the votes seen at the current height. - [Github Issues](https://github.com/tendermint/tendermint/issues) - [StackOverflow questions](https://stackoverflow.com/questions/tagged/tendermint) ## Monitoring Tendermint Each Tendermint instance has a standard `/health` RPC endpoint, which responds with 200 (OK) if everything is fine and 500 (or no response) - if something is wrong. Other useful endpoints include mentioned earlier `/status`, `/net_info` and `/validators`. We have a small tool, called `tm-monitor`, which outputs information from the endpoints above plus some statistics. The tool can be found [here](https://github.com/tendermint/tendermint/tree/master/tools/tm-monitor). Tendermint also can report and serve Prometheus metrics. See [Metrics](./metrics.md). ## What happens when my app dies? You are supposed to run Tendermint under a [process supervisor](https://en.wikipedia.org/wiki/Process_supervision) (like systemd or runit). It will ensure Tendermint is always running (despite possible errors). Getting back to the original question, if your application dies, Tendermint will panic. After a process supervisor restarts your application, Tendermint should be able to reconnect successfully. The order of restart does not matter for it. ## Signal handling We catch SIGINT and SIGTERM and try to clean up nicely. For other signals we use the default behaviour in Go: [Default behavior of signals in Go programs](https://golang.org/pkg/os/signal/#hdr-Default_behavior_of_signals_in_Go_programs). ## Corruption **NOTE:** Make sure you have a backup of the Tendermint data directory. ### Possible causes Remember that most corruption is caused by hardware issues: - RAID controllers with faulty / worn out battery backup, and an unexpected power loss - Hard disk drives with write-back cache enabled, and an unexpected power loss - Cheap SSDs with insufficient power-loss protection, and an unexpected power-loss - Defective RAM - Defective or overheating CPU(s) Other causes can be: - Database systems configured with fsync=off and an OS crash or power loss - Filesystems configured to use write barriers plus a storage layer that ignores write barriers. LVM is a particular culprit. - Tendermint bugs - Operating system bugs - Admin error (e.g., directly modifying Tendermint data-directory contents) (Source: https://wiki.postgresql.org/wiki/Corruption) ### WAL Corruption If consensus WAL is corrupted at the lastest height and you are trying to start Tendermint, replay will fail with panic. Recovering from data corruption can be hard and time-consuming. Here are two approaches you can take: 1. Delete the WAL file and restart Tendermint. It will attempt to sync with other peers. 2. Try to repair the WAL file manually: 1) Create a backup of the corrupted WAL file: ``` cp "$TMHOME/data/cs.wal/wal" > /tmp/corrupted_wal_backup ``` 2. Use `./scripts/wal2json` to create a human-readable version ``` ./scripts/wal2json/wal2json "$TMHOME/data/cs.wal/wal" > /tmp/corrupted_wal ``` 3. Search for a "CORRUPTED MESSAGE" line. 4. By looking at the previous message and the message after the corrupted one and looking at the logs, try to rebuild the message. If the consequent messages are marked as corrupted too (this may happen if length header got corrupted or some writes did not make it to the WAL ~ truncation), then remove all the lines starting from the corrupted one and restart Tendermint. ``` $EDITOR /tmp/corrupted_wal ``` 5. After editing, convert this file back into binary form by running: ``` ./scripts/json2wal/json2wal /tmp/corrupted_wal $TMHOME/data/cs.wal/wal ``` ## Hardware ### Processor and Memory While actual specs vary depending on the load and validators count, minimal requirements are: - 1GB RAM - 25GB of disk space - 1.4 GHz CPU SSD disks are preferable for applications with high transaction throughput. Recommended: - 2GB RAM - 100GB SSD - x64 2.0 GHz 2v CPU While for now, Tendermint stores all the history and it may require significant disk space over time, we are planning to implement state syncing (See [this issue](https://github.com/tendermint/tendermint/issues/828)). So, storing all the past blocks will not be necessary. ### Operating Systems Tendermint can be compiled for a wide range of operating systems thanks to Go language (the list of \$OS/\$ARCH pairs can be found [here](https://golang.org/doc/install/source#environment)). While we do not favor any operation system, more secure and stable Linux server distributions (like Centos) should be preferred over desktop operation systems (like Mac OS). ### Miscellaneous NOTE: if you are going to use Tendermint in a public domain, make sure you read [hardware recommendations](https://cosmos.network/validators) for a validator in the Cosmos network. ## Configuration parameters - `p2p.flush_throttle_timeout` - `p2p.max_packet_msg_payload_size` - `p2p.send_rate` - `p2p.recv_rate` If you are going to use Tendermint in a private domain and you have a private high-speed network among your peers, it makes sense to lower flush throttle timeout and increase other params. ``` [p2p] send_rate=20000000 # 2MB/s recv_rate=20000000 # 2MB/s flush_throttle_timeout=10 max_packet_msg_payload_size=10240 # 10KB ``` - `mempool.recheck` After every block, Tendermint rechecks every transaction left in the mempool to see if transactions committed in that block affected the application state, so some of the transactions left may become invalid. If that does not apply to your application, you can disable it by setting `mempool.recheck=false`. - `mempool.broadcast` Setting this to false will stop the mempool from relaying transactions to other peers until they are included in a block. It means only the peer you send the tx to will see it until it is included in a block. - `consensus.skip_timeout_commit` We want `skip_timeout_commit=false` when there is economics on the line because proposers should wait to hear for more votes. But if you don't care about that and want the fastest consensus, you can skip it. It will be kept false by default for public deployments (e.g. [Cosmos Hub](https://cosmos.network/intro/hub)) while for enterprise applications, setting it to true is not a problem. - `consensus.peer_gossip_sleep_duration` You can try to reduce the time your node sleeps before checking if theres something to send its peers. - `consensus.timeout_commit` You can also try lowering `timeout_commit` (time we sleep before proposing the next block). - `p2p.addr_book_strict` By default, Tendermint checks whenever a peer's address is routable before saving it to the address book. The address is considered as routable if the IP is [valid and within allowed ranges](https://github.com/tendermint/tendermint/blob/27bd1deabe4ba6a2d9b463b8f3e3f1e31b993e61/p2p/netaddress.go#L209). This may not be the case for private or local networks, where your IP range is usually strictly limited and private. If that case, you need to set `addr_book_strict` to `false` (turn it off). - `rpc.max_open_connections` By default, the number of simultaneous connections is limited because most OS give you limited number of file descriptors. If you want to accept greater number of connections, you will need to increase these limits. [Sysctls to tune the system to be able to open more connections](https://github.com/satori-com/tcpkali/blob/master/doc/tcpkali.man.md#sysctls-to-tune-the-system-to-be-able-to-open-more-connections) ...for N connections, such as 50k: ``` kern.maxfiles=10000+2*N # BSD kern.maxfilesperproc=100+2*N # BSD kern.ipc.maxsockets=10000+2*N # BSD fs.file-max=10000+2*N # Linux net.ipv4.tcp_max_orphans=N # Linux # For load-generating clients. net.ipv4.ip_local_port_range="10000 65535" # Linux. net.inet.ip.portrange.first=10000 # BSD/Mac. net.inet.ip.portrange.last=65535 # (Enough for N < 55535) net.ipv4.tcp_tw_reuse=1 # Linux net.inet.tcp.maxtcptw=2*N # BSD # If using netfilter on Linux: net.netfilter.nf_conntrack_max=N echo $((N/8)) > /sys/module/nf_conntrack/parameters/hashsize ``` The similar option exists for limiting the number of gRPC connections - `rpc.grpc_max_open_connections`.