Add instructions to setup and run testnet L2 and Nitro deployments (#26)
Part of [Create a public laconicd testnet](https://www.notion.so/Create-a-public-laconicd-testnet-896a11bdd8094eff8f1b49c0be0ca3b8) Co-authored-by: Shreerang Kale <shreerangkale@gmail.com> Co-authored-by: Neeraj <neeraj.rtly@gmail.com> Reviewed-on: #26 Co-authored-by: Prathamesh Musale <prathamesh.musale0@gmail.com> Co-committed-by: Prathamesh Musale <prathamesh.musale0@gmail.com>
This commit is contained in:
parent
8c672c0966
commit
9d92a20ab9
@ -15,3 +15,7 @@ Stacks to run a node for laconic testnet
|
||||
## Join LORO testnet
|
||||
|
||||
Follow steps in [testnet-onboarding-validator.md](./testnet-onboarding-validator.md) to onboard your participant and join as a validator on the LORO testnet
|
||||
|
||||
## Run testnet Nitro Node
|
||||
|
||||
Follow steps in [testnet-nitro-node.md](./testnet-nitro-node.md) to run you Nitro node for the testnet
|
||||
|
@ -10,6 +10,515 @@
|
||||
cd /srv
|
||||
```
|
||||
|
||||
## Prerequisites
|
||||
|
||||
* Local:
|
||||
|
||||
* Clone the `cerc-io/testnet-ops` repository:
|
||||
|
||||
```bash
|
||||
git clone git@git.vdb.to:cerc-io/testnet-ops.git
|
||||
```
|
||||
|
||||
* Ansible: see [installation](https://git.vdb.to/cerc-io/testnet-ops#installation)
|
||||
|
||||
* On deployments VM(s):
|
||||
|
||||
* laconic-so: see [installation](https://git.vdb.to/cerc-io/testnet-ops/src/branch/main/stack-orchestrator-setup/README.md)
|
||||
|
||||
<details open>
|
||||
<summary>L2 Optimism</summary>
|
||||
|
||||
## L2 Optimism
|
||||
|
||||
* Stack: <https://git.vdb.to/cerc-io/fixturenet-optimism-stack>
|
||||
|
||||
* Source repos:
|
||||
* <github.com/ethereum-optimism/optimism@v1.7.7>
|
||||
* <github.com/ethereum-optimism/op-geth@v1.101315.2>
|
||||
|
||||
* Target dir: `/srv/op-sepolia/optimism-deployment`
|
||||
|
||||
* Cleanup an existing deployment on VM if required:
|
||||
|
||||
```bash
|
||||
cd /srv/op-sepolia
|
||||
|
||||
# Stop the deployment
|
||||
laconic-so deployment --dir optimism-deployment stop --delete-volumes
|
||||
|
||||
# Remove the deployment dir
|
||||
sudo rm -rf optimism-deployment
|
||||
```
|
||||
|
||||
### Setup
|
||||
|
||||
* Switch to `testnet-ops/l2-setup` directory on your local machine:
|
||||
|
||||
```bash
|
||||
cd testnet-ops/l2-setup
|
||||
```
|
||||
|
||||
* Copy the `l2-vars-example.yml` vars file:
|
||||
|
||||
```bash
|
||||
cp l2-vars-example.yml l2-vars.yml
|
||||
```
|
||||
|
||||
* Edit `l2-vars.yml` with required values:
|
||||
|
||||
```bash
|
||||
# L1 chain ID (Sepolia: 11155111)
|
||||
l1_chain_id: "11155111"
|
||||
|
||||
# L1 RPC endpoint
|
||||
l1_rpc: "http://host.docker.internal:8545"
|
||||
|
||||
# L1 RPC endpoint host or IP address
|
||||
l1_host: "host.docker.internal"
|
||||
|
||||
# L1 RPC endpoint port number
|
||||
l1_port: "8545"
|
||||
|
||||
# L1 Beacon endpoint
|
||||
l1_beacon: "http://host.docker.internal:8001"
|
||||
|
||||
# Account credentials for the Admin account
|
||||
# Used for Optimism contracts deployment and funding other generated accounts
|
||||
l1_address: ""
|
||||
l1_priv_key: ""
|
||||
```
|
||||
|
||||
* Update the target dir in `setup-vars.yml`:
|
||||
|
||||
```bash
|
||||
sed -i 's|^l2_directory:.*|l2_directory: /srv/op-sepolia|' setup-vars.yml
|
||||
|
||||
# Will create deployment at /srv/op-sepolia/optimism-deployment
|
||||
```
|
||||
|
||||
### Run
|
||||
|
||||
* Set up and run L2 on remote host by executing `run-optimism.yml` Ansible playbook on your local machine:
|
||||
|
||||
* Create a new `hosts.ini` file:
|
||||
|
||||
```bash
|
||||
cp ../hosts.example.ini hosts.ini
|
||||
```
|
||||
|
||||
* Edit the [`hosts.ini`](./hosts.ini) file to run the playbook on a remote machine:
|
||||
|
||||
```ini
|
||||
[l2_host]
|
||||
<host_name> ansible_host=<target_ip> ansible_user=<ssh_user> ansible_ssh_common_args='-o ForwardAgent=yes'
|
||||
```
|
||||
|
||||
- Replace `<host_name>` with the alias of your choice
|
||||
- Replace `<target_ip>` with the IP address or hostname of the target machine
|
||||
- Replace `<ssh_user>` with the SSH username (e.g., dev, ubuntu)
|
||||
|
||||
* Verify that you are able to connect to the host using the following command:
|
||||
|
||||
```bash
|
||||
ansible all -m ping -i hosts.ini -k
|
||||
|
||||
# Expected output:
|
||||
# <host_name> | SUCCESS => {
|
||||
# "ansible_facts": {
|
||||
# "discovered_interpreter_python": "/usr/bin/python3.10"
|
||||
# },
|
||||
# "changed": false,
|
||||
# "ping": "pong"
|
||||
# }
|
||||
```
|
||||
|
||||
* Execute the `run-optimism.yml` Ansible playbook for remote deployment:
|
||||
|
||||
```bash
|
||||
LANG=en_US.utf8 ansible-playbook -i hosts.ini run-optimism.yml --extra-vars='{ "target_host": "l2_host"}' --user $USER -kK
|
||||
```
|
||||
|
||||
* Bridge funds on L2:
|
||||
|
||||
* On the deployment VM, set the following variables:
|
||||
|
||||
```bash
|
||||
cd /srv/op-sepolia
|
||||
|
||||
L1_RPC=http://host.docker.internal:8545
|
||||
L2_RPC=http://host.docker.internal:9545
|
||||
|
||||
NETWORK=$(grep 'cluster-id' optimism-deployment/deployment.yml | sed 's/cluster-id: //')_default
|
||||
|
||||
DEPLOYMENT_CONTEXT=11155111
|
||||
ACCOUNT=<admin-account-address>
|
||||
```
|
||||
|
||||
* Read the bridge contract address from the L1 deployment records in the `op-node` container:
|
||||
|
||||
```bash
|
||||
BRIDGE=$(laconic-so deployment --dir optimism-deployment exec op-node "cat /l1-deployment/$DEPLOYMENT_CONTEXT-deploy.json" | jq -r .L1StandardBridgeProxy)
|
||||
|
||||
# Get the funded account's pk
|
||||
ACCOUNT_PK=$(laconic-so deployment --dir optimism-deployment exec op-node "jq -r '.AdminKey' /l2-accounts/accounts.json")
|
||||
```
|
||||
|
||||
* Check that the starting balance for account on L2 is 0:
|
||||
|
||||
```bash
|
||||
docker run --rm --network $NETWORK cerc/optimism-contracts:local "cast balance $ACCOUNT --rpc-url $L2_RPC"
|
||||
|
||||
# 0
|
||||
```
|
||||
|
||||
* Use cast to send ETH to the bridge contract:
|
||||
|
||||
```bash
|
||||
docker run --rm cerc/optimism-contracts:local "cast send --from $ACCOUNT --value 1ether $BRIDGE --rpc-url $L1_RPC --private-key $ACCOUNT_PK"
|
||||
```
|
||||
|
||||
* Allow a couple minutes for the bridge to complete
|
||||
|
||||
* Check balance on L2:
|
||||
|
||||
```bash
|
||||
docker run --rm --network $NETWORK cerc/optimism-contracts:local "cast balance $ACCOUNT --rpc-url $L2_RPC"
|
||||
|
||||
# 100000000000000000
|
||||
```
|
||||
|
||||
</details>
|
||||
|
||||
<details open>
|
||||
<summary>L1 Nitro Contracts Deployment</summary>
|
||||
|
||||
## L1 Nitro Contracts Deployment
|
||||
|
||||
* Stack: <https://git.vdb.to/cerc-io/nitro-stack/src/branch/main/stack-orchestrator/stacks/nitro-contracts>
|
||||
|
||||
* Source repo: <https://github.com/cerc-io/go-nitro>
|
||||
|
||||
* Target dir: `/srv/bridge/nitro-contracts-deployment`
|
||||
|
||||
* Cleanup an existing deployment on VM if required:
|
||||
|
||||
```bash
|
||||
cd /srv/bridge
|
||||
|
||||
# Stop the deployment
|
||||
laconic-so deployment --dir nitro-contracts-deployment stop --delete-volumes
|
||||
|
||||
# Remove the deployment dir
|
||||
sudo rm -rf nitro-contracts-deployment
|
||||
|
||||
### Setup
|
||||
|
||||
* Switch to `testnet-ops/nitro-contracts-setup` directory on your local machine:
|
||||
|
||||
```bash
|
||||
cd testnet-ops/nitro-contracts-setup
|
||||
```
|
||||
|
||||
* Copy the `contract-vars-example.yml` vars file:
|
||||
|
||||
```bash
|
||||
cp contract-vars.example.yml contract-vars.yml
|
||||
```
|
||||
|
||||
* Edit [`contract-vars.yml`](./contract-vars.yml) and fill in the following values:
|
||||
|
||||
```bash
|
||||
# L1 RPC endpoint
|
||||
geth_url: "https://sepolia.laconic.com"
|
||||
|
||||
# L1 chain ID (Sepolia: 11155111)
|
||||
geth_chain_id: "11155111"
|
||||
|
||||
# Private key for a funded L1 account, to be used for contract deployment on L1
|
||||
# Must also be funded on L2 for deploying contracts
|
||||
# Required since this private key will be utilized by both L1 and L2 nodes of the bridge
|
||||
geth_deployer_pk: ""
|
||||
|
||||
# Custom L1 token to be deployed
|
||||
token_name: ""
|
||||
token_symbol: ""
|
||||
intial_token_supply: ""
|
||||
```
|
||||
|
||||
* Update the target dir in `setup-vars.yml`:
|
||||
|
||||
```bash
|
||||
sed -i 's|^nitro_directory:.*|nitro_directory: /srv/bridge|' setup-vars.yml
|
||||
|
||||
# Will create deployment at /srv/bridge/nitro-contracts-deployment
|
||||
```
|
||||
|
||||
### Run
|
||||
|
||||
* Deploy nitro contracts on remote host by executing `deploy-contracts.yml` Ansible playbook on your local machine:
|
||||
|
||||
* Create a new `hosts.ini` file:
|
||||
|
||||
```bash
|
||||
cp ../hosts.example.ini hosts.ini
|
||||
```
|
||||
|
||||
* Edit the [`hosts.ini`](./hosts.ini) file to run the playbook on a remote machine:
|
||||
|
||||
```ini
|
||||
[nitro_host]
|
||||
<host_name> ansible_host=<target_ip> ansible_user=<ssh_user> ansible_ssh_common_args='-o ForwardAgent=yes'
|
||||
```
|
||||
|
||||
- Replace `<host_name>` with the alias of your choice
|
||||
- Replace `<target_ip>` with the IP address or hostname of the target machine
|
||||
- Replace `<ssh_user>` with the SSH username (e.g., dev, ubuntu)
|
||||
|
||||
* Verify that you are able to connect to the host using the following command:
|
||||
|
||||
```bash
|
||||
ansible all -m ping -i hosts.ini -k
|
||||
|
||||
# Expected output:
|
||||
# <host_name> | SUCCESS => {
|
||||
# "ansible_facts": {
|
||||
# "discovered_interpreter_python": "/usr/bin/python3.10"
|
||||
# },
|
||||
# "changed": false,
|
||||
# "ping": "pong"
|
||||
# }
|
||||
```
|
||||
|
||||
* Execute the `deploy-contracts.yml` Ansible playbook for remote deployment:
|
||||
|
||||
```bash
|
||||
LANG=en_US.utf8 ansible-playbook -i hosts.ini deploy-contracts.yml --extra-vars='{ "target_host": "nitro_host"}' --user $USER -kK
|
||||
```
|
||||
|
||||
* Check logs for deployment on the virtual machine:
|
||||
|
||||
```bash
|
||||
cd /srv/bridge
|
||||
|
||||
# Check the l1 nitro contract deployments
|
||||
laconic-so deployment --dir nitro-contracts-deployment logs nitro-contracts -f
|
||||
```
|
||||
|
||||
</details>
|
||||
|
||||
<details open>
|
||||
<summary>Nitro Bridge</summary>
|
||||
|
||||
## Nitro Bridge
|
||||
|
||||
* Stack: <https://git.vdb.to/cerc-io/nitro-stack/src/branch/main/stack-orchestrator/stacks/bridge>
|
||||
|
||||
* Source repo: <https://github.com/cerc-io/go-nitro>
|
||||
|
||||
* Target dir: `/srv/bridge/bridge-deployment`
|
||||
|
||||
* Cleanup an existing deployment on VM if required:
|
||||
|
||||
```bash
|
||||
cd /srv/bridge
|
||||
|
||||
# Stop the deployment
|
||||
laconic-so deployment --dir bridge-deployment stop --delete-volumes
|
||||
|
||||
# Remove the deployment dir
|
||||
sudo rm -rf bridge-deployment
|
||||
```
|
||||
|
||||
### Setup
|
||||
|
||||
* Execute the command on the deployment VM to get the deployed L1 Nitro contract addresses along with the L1 asset address:
|
||||
|
||||
```bash
|
||||
cd /srv/bridge
|
||||
|
||||
laconic-so deployment --dir nitro-contracts-deployment exec nitro-contracts "cat /app/deployment/nitro-addresses.json"
|
||||
|
||||
# Expected output:
|
||||
# {
|
||||
# "11155111": [
|
||||
# {
|
||||
# "name": "geth",
|
||||
# "chainId": "11155111",
|
||||
# "contracts": {
|
||||
# "ConsensusApp": {
|
||||
# "address": "0xC98aD0B41B9224dad0605be32A9241dB9c67E2e8"
|
||||
# },
|
||||
# "NitroAdjudicator": {
|
||||
# "address": "0x7C22fdA703Cdf09eB8D3B5Adc81F723526713D0e"
|
||||
# },
|
||||
# "VirtualPaymentApp": {
|
||||
# "address": "0x778e4e6297E8BF04C67a20Ec989618d72eB4a19E"
|
||||
# },
|
||||
# "Token": {
|
||||
# "address": "0x02ebfB2706527C7310F2a7d9098b2BC61014C5F2"
|
||||
# }
|
||||
# }
|
||||
# }
|
||||
# ]
|
||||
# }
|
||||
```
|
||||
|
||||
* Switch to `testnet-ops/nitro-bridge-setup` directory on your local machine:
|
||||
|
||||
```bash
|
||||
cd testnet-ops/nitro-bridge-setup
|
||||
```
|
||||
|
||||
* Create the required vars file:
|
||||
|
||||
```bash
|
||||
cp bridge-vars.example.yml bridge-vars.yml
|
||||
```
|
||||
|
||||
* Edit `bridge-vars.yml` with required values:
|
||||
|
||||
```bash
|
||||
# L1 WS endpoint
|
||||
nitro_l1_chain_url: "wss://sepolia.laconic.com"
|
||||
|
||||
# L2 WS endpoint
|
||||
nitro_l2_chain_url: "wss://optimism.laconic.com"
|
||||
|
||||
# Private key for bridge Nitro address
|
||||
nitro_sc_pk: ""
|
||||
|
||||
# Private key should correspond to a funded account on both L1 and L2, and this account must own the Nitro contracts on L1
|
||||
# It also needs to hold L1 tokens to fund Nitro channels and will be used for deploying contracts on L2
|
||||
nitro_chain_pk: ""
|
||||
|
||||
# L2 chain ID (Optimism: 42069)
|
||||
optimism_chain_id: "42069"
|
||||
|
||||
# L2 RPC endpoint
|
||||
optimism_url: "https://optimism.laconic.com"
|
||||
|
||||
# Custom L2 token to be deployed
|
||||
token_name: ""
|
||||
token_symbol: ""
|
||||
intial_token_supply: ""
|
||||
|
||||
# Deployed L1 Nitro contract addresses
|
||||
na_address: ""
|
||||
vpa_address: ""
|
||||
ca_address: ""
|
||||
|
||||
# Deployed L1 token address
|
||||
l1_asset_address: ""
|
||||
```
|
||||
|
||||
* Update the target dir in `setup-vars.yml`:
|
||||
|
||||
```bash
|
||||
sed -i 's|^nitro_directory:.*|nitro_directory: /srv/bridge|' setup-vars.yml
|
||||
|
||||
# Will create deployment at /srv/bridge/nitro-contracts-deployment and /srv/bridge/bridge-deployment
|
||||
```
|
||||
|
||||
### Run
|
||||
|
||||
* Deploy L2 contracts and start the bridge on remote host by executing `run-nitro-bridge.yml` Ansible playbook on your local machine:
|
||||
|
||||
* Create a new `hosts.ini` file:
|
||||
|
||||
```bash
|
||||
cp ../hosts.example.ini hosts.ini
|
||||
```
|
||||
|
||||
* Edit the [`hosts.ini`](./hosts.ini) file to run the playbook on a remote machine:
|
||||
|
||||
```ini
|
||||
[nitro_host]
|
||||
<host_name> ansible_host=<target_ip> ansible_user=<ssh_user> ansible_ssh_common_args='-o ForwardAgent=yes'
|
||||
```
|
||||
|
||||
- Replace `<host_name>` with the alias of your choice
|
||||
- Replace `<target_ip>` with the IP address or hostname of the target machine
|
||||
- Replace `<ssh_user>` with the SSH username (e.g., dev, ubuntu)
|
||||
|
||||
* Verify that you are able to connect to the host using the following command:
|
||||
|
||||
```bash
|
||||
ansible all -m ping -i hosts.ini -k
|
||||
|
||||
# Expected output:
|
||||
# <host_name> | SUCCESS => {
|
||||
# "ansible_facts": {
|
||||
# "discovered_interpreter_python": "/usr/bin/python3.10"
|
||||
# },
|
||||
# "changed": false,
|
||||
# "ping": "pong"
|
||||
# }
|
||||
```
|
||||
|
||||
* Execute the `run-nitro-bridge.yml` Ansible playbook for remote deployment:
|
||||
|
||||
```bash
|
||||
LANG=en_US.utf8 ansible-playbook -i hosts.ini run-nitro-bridge.yml --extra-vars='{ "target_host": "nitro_host"}' --user $USER -kK
|
||||
```
|
||||
|
||||
* Check logs for deployments on the virtual machine:
|
||||
|
||||
```bash
|
||||
cd /srv/bridge
|
||||
|
||||
# Check the l2 nitro contract deployments
|
||||
laconic-so deployment --dir bridge-deployment logs l2-nitro-contracts -f
|
||||
|
||||
# Check bridge logs, ensure that the node is running
|
||||
laconic-so deployment --dir bridge-deployment logs nitro-bridge -f
|
||||
```
|
||||
|
||||
* Create Nitro node config for users:
|
||||
|
||||
```bash
|
||||
cd /srv/bridge
|
||||
|
||||
# Create required variables
|
||||
GETH_CHAIN_ID="11155111"
|
||||
OPTIMISM_CHAIN_ID="42069"
|
||||
|
||||
export NA_ADDRESS=$(laconic-so deployment --dir nitro-contracts-deployment exec nitro-contracts "jq -r '.\"$GETH_CHAIN_ID\"[0].contracts.NitroAdjudicator.address' /app/deployment/nitro-addresses.json")
|
||||
export CA_ADDRESS=$(laconic-so deployment --dir nitro-contracts-deployment exec nitro-contracts "jq -r '.\"$GETH_CHAIN_ID\"[0].contracts.ConsensusApp.address' /app/deployment/nitro-addresses.json")
|
||||
export VPA_ADDRESS=$(laconic-so deployment --dir nitro-contracts-deployment exec nitro-contracts "jq -r '.\"$GETH_CHAIN_ID\"[0].contracts.VirtualPaymentApp.address' /app/deployment/nitro-addresses.json")
|
||||
|
||||
export L1_ASSET_ADDRESS=$(laconic-so deployment --dir nitro-contracts-deployment exec nitro-contracts "jq -r '.\"$GETH_CHAIN_ID\"[0].contracts.Token.address' /app/deployment/nitro-addresses.json")
|
||||
|
||||
export BRIDGE_CONTRACT_ADDRESS=$(laconic-so deployment --dir bridge-deployment exec nitro-bridge "jq -r '.\"$OPTIMISM_CHAIN_ID\"[0].contracts.Bridge.address' /app/deployment/nitro-addresses.json")
|
||||
|
||||
export BRIDGE_NITRO_ADDRESS=$(laconic-so deployment --dir bridge-deployment exec nitro-rpc-client "nitro-rpc-client get-node-info -p 4006 -h nitro-bridge" | jq -r '.SCAddress')
|
||||
|
||||
export BRIDGE_PEER_ID=$(laconic-so deployment --dir bridge-deployment exec nitro-rpc-client "nitro-rpc-client get-node-info -p 4006 -h nitro-bridge" | jq -r '.MessageServicePeerId')
|
||||
|
||||
export L1_BRIDGE_MULTIADDR="/dns4/bridge.laconic.com/tcp/3005/p2p/$BRIDGE_PEER_ID"
|
||||
export L2_BRIDGE_MULTIADDR="/dns4/bridge.laconic.com/tcp/3006/p2p/$BRIDGE_PEER_ID"
|
||||
|
||||
# Create the required config file
|
||||
cat <<EOF > nitro-node-config.yml
|
||||
nitro_l1_chain_url: "wss://sepolia.laconic.com"
|
||||
nitro_l2_chain_url: "wss://optimism.laconic.com"
|
||||
na_address: "$NA_ADDRESS"
|
||||
ca_address: "$CA_ADDRESS"
|
||||
vpa_address: "$VPA_ADDRESS"
|
||||
l1_asset_address: "${L1_ASSET_ADDRESS}"
|
||||
bridge_contract_address: "$BRIDGE_CONTRACT_ADDRESS"
|
||||
bridge_nitro_address: "$BRIDGE_NITRO_ADDRESS"
|
||||
nitro_l1_bridge_multiaddr: "$L1_BRIDGE_MULTIADDR"
|
||||
nitro_l2_bridge_multiaddr: "$L2_BRIDGE_MULTIADDR"
|
||||
EOF
|
||||
```
|
||||
|
||||
* The required config file should be generated at `/srv/bridge/nitro-node-config.yml`
|
||||
|
||||
* Check in the generated file at location `ops/stage2/nitro-node-config.yml` within this repository
|
||||
|
||||
</details>
|
||||
|
||||
<details open>
|
||||
<summary>stage0 laconicd</summary>
|
||||
|
||||
@ -669,10 +1178,22 @@
|
||||
## Domains / Port Mappings
|
||||
|
||||
```bash
|
||||
laconicd.laconic.com -> 26657
|
||||
laconicd.laconic.com/api -> 9473/api
|
||||
faucet.laconic.com -> 4000
|
||||
loro-signup.laconic.com -> 3000
|
||||
wallet.laconic.com -> 5000
|
||||
loro-console.laconic.com -> 4001
|
||||
# Machine 1
|
||||
https://laconicd.laconic.com -> 26657
|
||||
https://laconicd.laconic.com/api -> 9473/api
|
||||
https://faucet.laconic.com -> 4000
|
||||
https://loro-signup.laconic.com -> 3000
|
||||
https://wallet.laconic.com -> 5000
|
||||
https://loro-console.laconic.com -> 4001
|
||||
|
||||
# Machine 2
|
||||
https://sepolia.laconic.com -> 8545
|
||||
wss://sepolia.laconic.com -> 8546
|
||||
https://optimism.laconic.com -> 9545
|
||||
wss://optimism.laconic.com -> 9546
|
||||
|
||||
bridge.laconic.com
|
||||
Open ports:
|
||||
3005 (L1 side)
|
||||
3006 (L2 side)
|
||||
```
|
||||
|
@ -22,7 +22,8 @@
|
||||
},
|
||||
"scripts": {
|
||||
"build": "tsc",
|
||||
"map-subscribers-to-participants": "node dist/map-subscribers-to-participants.js"
|
||||
"map-subscribers-to-participants": "node dist/map-subscribers-to-participants.js",
|
||||
"participants-with-filtered-validators": "node dist/participants-with-filtered-validators.js"
|
||||
},
|
||||
"packageManager": "yarn@1.22.19+sha1.4ba7fc5c6e704fce2066ecbfb0b0d8976fe62447"
|
||||
}
|
||||
|
305
testnet-nitro-node.md
Normal file
305
testnet-nitro-node.md
Normal file
@ -0,0 +1,305 @@
|
||||
# testnet-nitro-node
|
||||
|
||||
## Prerequisites
|
||||
|
||||
* Ansible: see [installation](https://git.vdb.to/cerc-io/testnet-ops#installation)
|
||||
|
||||
* yq: see [installation](https://github.com/mikefarah/yq#install)
|
||||
|
||||
* laconic-so: see [installation](https://git.vdb.to/cerc-io/testnet-ops/src/branch/main/stack-orchestrator-setup/README.md)
|
||||
|
||||
* Check versions to verify installation:
|
||||
|
||||
```bash
|
||||
laconic-so version
|
||||
|
||||
ansible --version
|
||||
|
||||
yq --version
|
||||
```
|
||||
|
||||
## Setup
|
||||
|
||||
* Clone the `cerc-io/testnet-ops` repository:
|
||||
|
||||
```bash
|
||||
git clone git@git.vdb.to:cerc-io/testnet-ops.git
|
||||
|
||||
cd testnet-ops/nitro-node-setup
|
||||
```
|
||||
|
||||
* Fetch the required Nitro node config:
|
||||
|
||||
```bash
|
||||
wget -O nitro-vars.yml https://git.vdb.to/cerc-io/testnet-laconicd-stack/raw/branch/main/ops/stage2/nitro-node-config.yml
|
||||
|
||||
# Expected variables in the fetched config file:
|
||||
|
||||
# nitro_l1_chain_url: ""
|
||||
# nitro_l2_chain_url: ""
|
||||
# na_address: ""
|
||||
# ca_address: ""
|
||||
# vpa_address: ""
|
||||
# l1_asset_address: ""
|
||||
# bridge_contract_address: ""
|
||||
# bridge_nitro_address: ""
|
||||
# nitro_l1_bridge_multiaddr: ""
|
||||
# nitro_l2_bridge_multiaddr: ""
|
||||
```
|
||||
|
||||
* TODO: Get L1 tokens on your address
|
||||
|
||||
* Edit `nitro-vars.yml` and add the following variables:
|
||||
|
||||
```bash
|
||||
# Private key for your Nitro account (same as the one used in stage0 onboarding)
|
||||
# Export the key from Laconic wallet (https://wallet.laconic.com)
|
||||
nitro_sc_pk: ""
|
||||
|
||||
# Private key for a funded account on L1
|
||||
# This account should have L1 tokens for funding your Nitro channels
|
||||
nitro_chain_pk: ""
|
||||
|
||||
# Multiaddr with publically accessible IP address / DNS for your L1 nitro node
|
||||
# Use port 3007
|
||||
# Example: "/ip4/192.168.x.y/tcp/3007"
|
||||
# Example: "/dns4/example.com/tcp/3007"
|
||||
nitro_l1_ext_multiaddr: ""
|
||||
|
||||
# Multiaddr with publically accessible IP address / DNS for your L2 nitro node
|
||||
# Use port 3009
|
||||
# Example: "/ip4/192.168.x.y/tcp/3009"
|
||||
# Example: "/dns4/example.com/tcp/3009"
|
||||
nitro_l2_ext_multiaddr: ""
|
||||
```
|
||||
|
||||
* Update the target dir in `setup-vars.yml`:
|
||||
|
||||
```bash
|
||||
# Set path to desired deployments dir
|
||||
DEPLOYMENTS_DIR=<path-to-deployments-dir>
|
||||
|
||||
sed -i "s|^nitro_directory:.*|nitro_directory: $DEPLOYMENTS_DIR/nitro-node|" setup-vars.yml
|
||||
|
||||
# Will create deployments at $DEPLOYMENTS_DIR/nitro-node/l1-nitro-deployment and $DEPLOYMENTS_DIR/nitro-node/l2-nitro-deployment
|
||||
```
|
||||
|
||||
## Run Nitro Nodes
|
||||
|
||||
### On Local Host
|
||||
|
||||
* Setup and run a Nitro node (L1+L2) by executing the `run-nitro-nodes.yml` Ansible playbook:
|
||||
|
||||
```bash
|
||||
LANG=en_US.utf8 ansible-playbook -i localhost, --connection=local run-nitro-nodes.yml --extra-vars='{ "target_host": "localhost"}' --user $USER
|
||||
```
|
||||
|
||||
### On Remote Host (optional)
|
||||
|
||||
* Create a new `hosts.ini` file:
|
||||
|
||||
```bash
|
||||
cp ../hosts.example.ini hosts.ini
|
||||
```
|
||||
|
||||
* Edit the [`hosts.ini`](./hosts.ini) file to run the playbook on a remote machine:
|
||||
|
||||
```ini
|
||||
[nitro_host]
|
||||
<host_name> ansible_host=<target_ip> ansible_user=<ssh_user> ansible_ssh_common_args='-o ForwardAgent=yes'
|
||||
```
|
||||
|
||||
- Replace `<host_name>` with the alias of your choice
|
||||
- Replace `<target_ip>` with the IP address or hostname of the target machine
|
||||
- Replace `<ssh_user>` with the SSH username (e.g., dev, ubuntu)
|
||||
|
||||
* Verify that you are able to connect to the host using the following command
|
||||
|
||||
```bash
|
||||
ansible all -m ping -i hosts.ini -k
|
||||
|
||||
# Expected output:
|
||||
|
||||
# <host_name> | SUCCESS => {
|
||||
# "ansible_facts": {
|
||||
# "discovered_interpreter_python": "/usr/bin/python3.10"
|
||||
# },
|
||||
# "changed": false,
|
||||
# "ping": "pong"
|
||||
# }
|
||||
```
|
||||
|
||||
* Execute the `run-nitro-nodes.yml` Ansible playbook for remote deployment:
|
||||
|
||||
```bash
|
||||
LANG=en_US.utf8 ansible-playbook -i hosts.ini run-nitro-nodes.yml --extra-vars='{ "target_host": "nitro_host"}' --user $USER -kK
|
||||
```
|
||||
|
||||
### Check Deployment Status
|
||||
|
||||
* Run the following command in the directory where the deployments are created:
|
||||
|
||||
```bash
|
||||
cd $DEPLOYMENTS_DIR/nitro-node
|
||||
|
||||
# Check the logs, ensure that the nodes are running
|
||||
laconic-so deployment --dir l1-nitro-deployment logs nitro-node -f
|
||||
laconic-so deployment --dir l2-nitro-deployment logs nitro-node -f
|
||||
```
|
||||
|
||||
## Create Channels
|
||||
Create a ledger channel with the bridge on L1 which is mirrored on L2
|
||||
|
||||
* Run the following commands from the directory where the deployments are created
|
||||
|
||||
* Set required variables:
|
||||
|
||||
```bash
|
||||
cd $DEPLOYMENTS_DIR/nitro-node
|
||||
|
||||
# Fetch the required Nitro node config
|
||||
wget https://git.vdb.to/cerc-io/testnet-laconicd-stack/raw/branch/main/ops/stage2/nitro-node-config.yml
|
||||
|
||||
export BRIDGE_NITRO_ADDRESS=$(yq eval '.bridge_nitro_address' nitro-node-config.yml)
|
||||
export L1_ASSET_ADDRESS=$(yq eval '.l1_asset_address' nitro-node-config.yml)
|
||||
```
|
||||
|
||||
* Check that check that you have no existing channels on L1 or L2:
|
||||
|
||||
```bash
|
||||
laconic-so deployment --dir l1-nitro-deployment exec nitro-rpc-client "nitro-rpc-client get-all-ledger-channels -p 4005 -h nitro-node"
|
||||
laconic-so deployment --dir l2-nitro-deployment exec nitro-rpc-client "nitro-rpc-client get-all-ledger-channels -p 4005 -h nitro-node"
|
||||
|
||||
# Expected output:
|
||||
# []
|
||||
```
|
||||
|
||||
* Create a ledger channel between your L1 Nitro node and Bridge with custom asset:
|
||||
|
||||
```bash
|
||||
# Set amount to ledger
|
||||
LEDGER_AMOUNT=1000000
|
||||
|
||||
laconic-so deployment --dir l1-nitro-deployment exec nitro-rpc-client "nitro-rpc-client direct-fund $BRIDGE_NITRO_ADDRESS --assetAddress $L1_ASSET_ADDRESS --alphaAmount $LEDGER_AMOUNT --betaAmount $LEDGER_AMOUNT -p 4005 -h nitro-node"
|
||||
|
||||
# Follow your L1 Nitro node logs for progress
|
||||
|
||||
# Expected Output:
|
||||
# Objective started DirectFunding-0x161d289a50222caa781db215bb82a3ede4f557217742245525b8e8cbff04ec21
|
||||
# Channel Open 0x161d289a50222caa781db215bb82a3ede4f557217742245525b8e8cbff04ec21
|
||||
|
||||
# Set the resulting ledger channel id in a variable
|
||||
export LEDGER_CHANNEL_ID=
|
||||
```
|
||||
|
||||
* Check the [Troubleshooting](#troubleshooting) section if command to create a ledger channel fails or gets stuck
|
||||
|
||||
* Once direct-fund objective is complete, the bridge will create a mirrored channel on L2
|
||||
|
||||
* Check L2 Nitro node's logs to see that a bridged-fund objective completed:
|
||||
|
||||
```bash
|
||||
laconic-so deployment --dir l2-nitro-deployment logs nitro-node -f --tail 30
|
||||
|
||||
# Expected Output:
|
||||
# nitro-node-1 | 5:01AM INF INFO Objective cranked address=0xaaa6628ec44a8a742987ef3a114ddfe2d4f7adce objective-id=bridgedfunding-0x6a9f5ccf1fa802525d794f4a899897f947615f6acc7141e61e056a8bfca29179 waiting-for=WaitingForNothing
|
||||
# nitro-node-1 | 5:01AM INF INFO Objective is complete & returned to API address=0xaaa6628ec44a8a742987ef3a114ddfe2d4f7adce objective-id=bridgedfunding-0x6a9f5ccf1fa802525d794f4a899897f947615f6acc7141e61e056a8bfca29179
|
||||
```
|
||||
|
||||
* Check status of L1 ledger channel with the bridge using channel id:
|
||||
|
||||
```bash
|
||||
laconic-so deployment --dir l1-nitro-deployment exec nitro-rpc-client "nitro-rpc-client get-ledger-channel $LEDGER_CHANNEL_ID -p 4005 -h nitro-node"
|
||||
|
||||
# Expected output:
|
||||
# {
|
||||
# ID: '0x161d289a50222caa781db215bb82a3ede4f557217742245525b8e8cbff04ec21',
|
||||
# Status: 'Open',
|
||||
# Balance: {
|
||||
# AssetAddress: '<l1-asset-address>',
|
||||
# Me: '<your-nitro-address>',
|
||||
# Them: '<bridge-nitro-address>',
|
||||
# MyBalance: <ledger-amount>n,
|
||||
# TheirBalance: <ledger-amount>n
|
||||
# },
|
||||
# ChannelMode: 'Open'
|
||||
# }
|
||||
```
|
||||
|
||||
* Check status of the mirrored channel on L2:
|
||||
|
||||
```bash
|
||||
laconic-so deployment --dir l2-nitro-deployment exec nitro-rpc-client "nitro-rpc-client get-all-ledger-channels -p 4005 -h nitro-node"
|
||||
|
||||
# Expected output:
|
||||
# [
|
||||
# {
|
||||
# "ID": "0x6a9f5ccf1fa802525d794f4a899897f947615f6acc7141e61e056a8bfca29179",
|
||||
# "Status": "Open",
|
||||
# "Balance": {
|
||||
# "AssetAddress": "<l2-asset-address>",
|
||||
# "Me": "<your-nitro-address>",
|
||||
# "Them": "<bridge-nitro-address>",
|
||||
# "MyBalance": <ledger-amount>n,
|
||||
# "TheirBalance": <ledger-amount>n
|
||||
# },
|
||||
# "ChannelMode": "Open"
|
||||
# }
|
||||
# ]
|
||||
```
|
||||
|
||||
## Clean up
|
||||
|
||||
* Switch to deployments dir:
|
||||
|
||||
```bash
|
||||
cd $DEPLOYMENTS_DIR/nitro-node
|
||||
```
|
||||
|
||||
* Stop all Nitro services running in the background:
|
||||
|
||||
```bash
|
||||
laconic-so deployment --dir l1-nitro-deployment stop
|
||||
laconic-so deployment --dir l2-nitro-deployment stop
|
||||
```
|
||||
|
||||
* To stop all services and also delete data:
|
||||
|
||||
```bash
|
||||
laconic-so deployment --dir l1-nitro-deployment stop --delete-volumes
|
||||
laconic-so deployment --dir l2-nitro-deployment stop --delete-volumes
|
||||
|
||||
# Remove deployment directories (deployments will have to be recreated for a re-run)
|
||||
sudo rm -r l1-nitro-deployment
|
||||
sudo rm -r l2-nitro-deployment
|
||||
```
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
* Stop (`Ctrl+C`) the direct-fund command if it is stuck
|
||||
|
||||
* Restart the L1 Nitro node:
|
||||
|
||||
* Stop the deployment:
|
||||
|
||||
```bash
|
||||
cd $DEPLOYMENTS_DIR/nitro-node
|
||||
|
||||
laconic-so deployment --dir l1-nitro-deployment stop
|
||||
```
|
||||
|
||||
* Reset the node's durable store:
|
||||
|
||||
```bash
|
||||
sudo rm -rf l1-nitro-deployment/data/nitro_node_data
|
||||
|
||||
mkdir l1-nitro-deployment/data/nitro_node_data
|
||||
```
|
||||
|
||||
* Restart the deployment:
|
||||
|
||||
```bash
|
||||
laconic-so deployment --dir l1-nitro-deployment start
|
||||
```
|
||||
|
||||
* Retry the ledger channel creation command
|
Loading…
Reference in New Issue
Block a user