2024-09-03 09:17:18 +00:00
# testnet-nitro-node
## Prerequisites
* laconic-so: see [installation ](https://git.vdb.to/cerc-io/stack-orchestrator#install )
* Ansible: see [installation ](https://git.vdb.to/cerc-io/testnet-ops#installation )
2024-09-05 11:38:25 +00:00
* yq: see [installation ](https://github.com/mikefarah/yq#install )
2024-09-03 09:17:18 +00:00
* Check versions to verify installation:
```bash
laconic-so version
ansible --version
2024-09-05 11:38:25 +00:00
yq --version
2024-09-03 09:17:18 +00:00
```
## Setup
* Clone the `cerc-io/testnet-ops` repository:
```bash
git clone git@git.vdb.to:cerc-io/testnet-ops.git
cd testnet-ops/nitro-node-setup
```
* Fetch the required Nitro node config:
```bash
wget -O nitro-vars.yml https://git.vdb.to/cerc-io/testnet-laconicd-stack/raw/branch/main/ops/stage2/nitro-node-config.yml
2024-09-05 07:06:41 +00:00
# Expected variables in the fetched config file:
# nitro_l1_chain_url: ""
# nitro_l2_chain_url: ""
# na_address: ""
# ca_address: ""
2024-09-05 11:38:25 +00:00
# vpa_address: ""
# l1_asset_address: ""
2024-09-05 07:06:41 +00:00
# bridge_contract_address: ""
# bridge_nitro_address: ""
# nitro_l1_bridge_multiaddr: ""
# nitro_l2_bridge_multiaddr: ""
2024-09-03 09:17:18 +00:00
```
* TODO: Get Laconic tokens on your address
2024-09-05 07:06:41 +00:00
* Edit `nitro-vars.yml` and add the following variables:
2024-09-03 09:17:18 +00:00
```bash
# Private key for your Nitro address
nitro_sc_pk: ""
# Private key for a funded account on L1
# This account should have Laconic tokens for funding your Nitro channels
nitro_chain_pk: ""
2024-09-05 07:06:41 +00:00
# Multiaddr with publically accessible IP address / DNS for your L1 nitro node
# Use port 3007
# Example: "/ip4/192.168.x.y/tcp/3007"
# Example: "/dns4/example.com/tcp/3007"
nitro_l1_ext_multiaddr: ""
# Multiaddr with publically accessible IP address / DNS for your L2 nitro node
# Use port 3009
2024-09-03 09:17:18 +00:00
# Example: "/ip4/192.168.x.y/tcp/3009"
# Example: "/dns4/example.com/tcp/3009"
2024-09-05 07:06:41 +00:00
nitro_l2_ext_multiaddr: ""
2024-09-03 09:17:18 +00:00
```
* Update the target dir in `setup-vars.yml` :
```bash
2024-09-03 10:10:29 +00:00
# Set path to desired deployments dir
2024-09-03 09:17:18 +00:00
DEPLOYMENTS_DIR=< path-to-deployments-dir >
2024-09-03 10:10:29 +00:00
2024-09-03 09:17:18 +00:00
sed -i "s|^nitro_directory:.*|nitro_directory: $DEPLOYMENTS_DIR/nitro-node|" setup-vars.yml
# Will create deployments at $DEPLOYMENTS_DIR/nitro-node/l1-nitro-deployment and $DEPLOYMENTS_DIR/nitro-node/l2-nitro-deployment
```
## Run
* Setup and run a Nitro node (L1+L2) by executing the `run-nitro-node.yml` Ansible playbook:
```bash
LANG=en_US.utf8 ansible-playbook -i localhost, --connection=local run-nitro-node.yml --extra-vars='{ "target_host": "localhost"}' --user $USER
```
* For skipping container build, run with `"skip_container_build" : true` :
```bash
LANG=en_US.utf8 ansible-playbook -i localhost, --connection=local run-nitro-node.yml --extra-vars='{ "target_host": "localhost", "skip_container_build": true }' --user $USER
```
2024-09-05 11:38:25 +00:00
* Check status:
```bash
cd $DEPLOYMENTS_DIR
# Check the logs, ensure that the nodes are running
laconic-so deployment --dir l1-nitro-deployment logs nitro-node -f
laconic-so deployment --dir l2-nitro-deployment logs nitro-node -f
```
## Create Channels
* Create a ledger channel with the bridge on L1 that gets mirrored on L2
* Set required variables:
```bash
cd $DEPLOYMENTS_DIR
# Fetch the required Nitro node config
wget https://git.vdb.to/cerc-io/testnet-laconicd-stack/raw/branch/main/ops/stage2/nitro-node-config.yml
export BRIDGE_NITRO_ADDRESS=$(yq eval '.bridge_contract_address' nitro-node-config.yml)
export L1_ASSET_ADDRESS=$(yq eval '.l1_asset_address' nitro-node-config.yml)
```
* Check that check that you have no existing channels on L1 or L2:
```bash
laconic-so deployment --dir l1-nitro-deployment exec nitro-rpc-client "nitro-rpc-client get-all-ledger-channels -p 4005 -h nitro-node"
laconic-so deployment --dir l2-nitro-deployment exec nitro-rpc-client "nitro-rpc-client get-all-ledger-channels -p 4005 -h nitro-node"
# Expected output:
# []
```
* Create a ledger channel between your L1 Nitro node and Bridge with custom asset:
```bash
# Set amount to ledger
LEDGER_AMOUNT=1000000
laconic-so deployment --dir l1-nitro-deployment exec nitro-rpc-client "nitro-rpc-client direct-fund $BRIDGE_NITRO_ADDRESS --assetAddress $L1_ASSET_ADDRESS --alphaAmount $LEDGER_AMOUNT --betaAmount $LEDGER_AMOUNT -p 4005 -h nitro-node"
# Follow your L1 Nitro node logs for progress
# Expected Output:
# Objective started DirectFunding-0x161d289a50222caa781db215bb82a3ede4f557217742245525b8e8cbff04ec21
# Channel Open 0x161d289a50222caa781db215bb82a3ede4f557217742245525b8e8cbff04ec21
# Set the resulting ledger channel id in a variable
export LEDGER_CHANNEL_ID=
```
* Check the [Troubleshooting ](#troubleshooting ) section if command to create a ledger channel fails or gets stuck
* Once direct-fund objective is complete, the bridge will create a mirrored channel on L2
* Check L2 Nitro node's logs to see that a bridged-fund objective completed:
```bash
laconic-so deployment --dir l2-nitro-deployment logs nitro-node -f --tail 30
# Expected Output:
# nitro-node-1 | 5:01AM INF INFO Objective cranked address=0xaaa6628ec44a8a742987ef3a114ddfe2d4f7adce objective-id=bridgedfunding-0x6a9f5ccf1fa802525d794f4a899897f947615f6acc7141e61e056a8bfca29179 waiting-for=WaitingForNothing
# nitro-node-1 | 5:01AM INF INFO Objective is complete & returned to API address=0xaaa6628ec44a8a742987ef3a114ddfe2d4f7adce objective-id=bridgedfunding-0x6a9f5ccf1fa802525d794f4a899897f947615f6acc7141e61e056a8bfca29179
```
* Check status of L1 ledger channel with the bridge using channel id:
```bash
laconic-so deployment --dir l1-nitro-deployment exec nitro-rpc-client "nitro-rpc-client get-ledger-channel $LEDGER_CHANNEL_ID -p 4005 -h nitro-node"
# Expected output:
# {
# ID: '0x161d289a50222caa781db215bb82a3ede4f557217742245525b8e8cbff04ec21',
# Status: 'Open',
# Balance: {
# AssetAddress: '< l1-asset-address > ',
# Me: '< your-nitro-address > ',
# Them: '< bridge-nitro-address > ',
# MyBalance: 1000000n,
# TheirBalance: 1000000n
# },
# ChannelMode: 'Open'
# }
```
* Check status of the mirrored channel on L2:
```bash
laconic-so deployment --dir l2-nitro-deployment exec nitro-rpc-client "nitro-rpc-client get-all-ledger-channels -p 4005 -h nitro-node"
# Expected output:
# {
# ID: '0x6a9f5ccf1fa802525d794f4a899897f947615f6acc7141e61e056a8bfca29179',
# Status: 'Open',
# Balance: {
# AssetAddress: '< l2-asset-address > ',
# Me: '< your-nitro-address > ',
# Them: '< bridge-nitro-address > ',
# MyBalance: 1000000n,
# TheirBalance: 1000000n
# },
# ChannelMode: 'Open'
# }
```
## Clean up
2024-09-03 09:17:18 +00:00
* Switch to deployments dir:
```bash
cd $DEPLOYMENTS_DIR/nitro-node
```
* Stop all Nitro services running in the background:
```bash
laconic-so deployment --dir l1-nitro-deployment stop
laconic-so deployment --dir l2-nitro-deployment stop
```
* To stop all services and also delete data:
```bash
laconic-so deployment --dir l1-nitro-deployment stop --delete-volumes
laconic-so deployment --dir l2-nitro-deployment stop --delete-volumes
# Remove deployment directories (deployments will have to be recreated for a re-run)
sudo rm -r l1-nitro-deployment
sudo rm -r l2-nitro-deployment
```