Compare commits

...

7 Commits

Author SHA1 Message Date
40428cdaa3 Add Ansible task to fetch nitro node config file (#8)
Part of [Automate testnet nitro deployments using Ansible](https://www.notion.so/Automate-testnet-nitro-deployments-using-Ansible-0d15579430204b8daba9a8aa31e07568)
- Added ansible tasks to:
  - Install yq
  - Download nitro node config file
- Modified playbooks to handle prompts when cloning repositories

Co-authored-by: Adw8 <adwaitgharpure@gmail.com>
Reviewed-on: cerc-io/testnet-ops#8
2024-09-17 13:55:18 +00:00
bcbd175f00 Update playbook for Nitro bridge setup (#7)
Part of [Automate testnet nitro deployments using Ansible](https://www.notion.so/Automate-testnet-nitro-deployments-using-Ansible-0d15579430204b8daba9a8aa31e07568)

Co-authored-by: Neeraj <neeraj.rtly@gmail.com>
Reviewed-on: cerc-io/testnet-ops#7
Co-authored-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
Co-committed-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
2024-09-16 06:49:46 +00:00
cd314d2fdf Add Ansible playbooks to setup Docker and stack orchestrator (#6)
Part of [Automate testnet nitro deployments using Ansible](https://www.notion.so/Automate-testnet-nitro-deployments-using-Ansible-0d15579430204b8daba9a8aa31e07568)

Add Ansible playbooks to:
- Setup Docker if not present
- Setup Stack Orchestrator if not present

Co-authored-by: Adw8 <adwaitgharpure@gmail.com>
Co-authored-by: Neeraj <neeraj.rtly@gmail.com>
Reviewed-on: cerc-io/testnet-ops#6
Co-authored-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
Co-committed-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
2024-09-12 04:43:53 +00:00
23f3a4c8ed Add instructions to run Ansible playbooks on remote machines (#5)
Part of [Automate testnet nitro deployments using Ansible](https://www.notion.so/Automate-testnet-nitro-deployments-using-Ansible-0d15579430204b8daba9a8aa31e07568)

Co-authored-by: Adw8 <adwaitgharpure@gmail.com>
Reviewed-on: cerc-io/testnet-ops#5
Co-authored-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
Co-committed-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
2024-09-09 13:37:41 +00:00
88e0b48540 Add ansible playbook to setup and run Nitro bridge (#4)
Part of [Automate testnet nitro deployments using Ansible](https://www.notion.so/Automate-testnet-nitro-deployments-using-Ansible-0d15579430204b8daba9a8aa31e07568)
Implement Ansible playbook to:
  - Deploy nitro contracts on L1
  - Deploy bridge contract on L2
  - Setup and run nitro bridge

Co-authored-by: Adw8 <adwaitgharpure@gmail.com>
Reviewed-on: cerc-io/testnet-ops#4
Co-authored-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
Co-committed-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
2024-09-09 06:11:31 +00:00
e142357428 Add ansible playbook to setup and run L2 deployment (#3)
Part of [Automate testnet nitro deployments using Ansible](https://www.notion.so/Automate-testnet-nitro-deployments-using-Ansible-0d15579430204b8daba9a8aa31e07568)

Co-authored-by: Adw8 <adwaitgharpure@gmail.com>
Reviewed-on: cerc-io/testnet-ops#3
Co-authored-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
Co-committed-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
2024-09-05 07:58:00 +00:00
a7af33a45b Add ansible playbooks to setup and run nitro nodes (#2)
Part of [Automate testnet nitro deployments using Ansible](https://www.notion.so/Automate-testnet-nitro-deployments-using-Ansible-0d15579430204b8daba9a8aa31e07568)
- Add playbooks to setup and run L1 and L2 nitro nodes

Co-authored-by: Adw8 <adwaitgharpure@gmail.com>
Reviewed-on: cerc-io/testnet-ops#2
Co-authored-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
Co-committed-by: Prathamesh Musale <prathamesh@noreply.git.vdb.to>
2024-09-05 07:27:35 +00:00
35 changed files with 1432 additions and 0 deletions

1
.gitignore vendored Normal file
View File

@ -0,0 +1 @@
hosts.ini

View File

@ -1,2 +1,45 @@
# testnet-ops
## Installation
- Install [Ansible](https://docs.ansible.com/ansible/latest/installation_guide/intro_installation.html#installing-and-upgrading-ansible-with-pip)
- Add location of the directory containing the ansible binary to your `PATH`
- Set Locale Encoding to `UTF-8`
Ansible requires the locale encoding to be `UTF-8`. You can either use the `LANG` prefix when running Ansible commands or set the system-wide locale
- Option 1: Use `LANG` Prefix in Commands
If you prefer not to change the system-wide locale, you can use the `LANG` prefix when running Ansible commands:
```bash
LANG=en_US.UTF-8 ansible-playbook your_playbook.yml
```
- Option 2: Set System-Wide Locale
- Edit the `/etc/default/locale` file:
```bash
sudo nano /etc/default/locale
```
- Set the `LANG` variable to en_US.UTF-8:
```
LANG="en_US.UTF-8"
```
- Reboot your system or log out and log back in to apply the changes
- Reference: <https://udhayakumarc.medium.com/error-ansible-requires-the-locale-encoding-to-be-utf-8-detected-iso8859-1-6da808387f7d>
## Playbooks
- [stack-orchestrator-setup](./stack-orchestrator-setup/README.md)
- [l2-setup](./l2-setup/README.md)
- [nitro-node-setup](./nitro-nodes-setup/README.md)
- [nitro-bridge-setup](./nitro-bridge-setup/README.md)
- [nitro-contracts-setup](./nitro-contracts-setup/README.md)

2
hosts.example.ini Normal file
View File

@ -0,0 +1,2 @@
[deployment_host]
<host_name> ansible_host=<target_ip> ansible_user=<ssh_user> ansible_ssh_common_args='-o ForwardAgent=yes'

3
l2-setup/.gitignore vendored Normal file
View File

@ -0,0 +1,3 @@
out
l2-vars.yml
hosts.ini

147
l2-setup/README.md Normal file
View File

@ -0,0 +1,147 @@
# l2-setup
## Setup Ansible
To get started, follow the [installation](../README.md#installation) guide to setup ansible on your machine
## Setup
The following commands have to be executed in the [`l2-setup`](./) directory:
- Copy the `l2-vars.example.yml` vars file:
```bash
cp l2-vars.example.yml l2-vars.yml
```
- Edit [`l2-vars.yml`](./l2-vars.yml) with the required values:
```yaml
# L1 chain ID
l1_chain_id: ""
# L1 RPC endpoint
l1_rpc: ""
# L1 RPC endpoint host or IP address
l1_host: ""
# L1 RPC endpoint port number
l1_port: ""
# L1 Beacon endpoint
l1_beacon: ""
# Address of the funded account on L1
# Used for optimism contracts deployment
l1_address: ""
# Private key of the funded account on L1
l1_priv_key: ""
```
## Run L2
### On Local Host
- To setup and run L2 locally, execute the `run-optimism.yml` Ansible playbook:
```bash
LANG=en_US.utf8 ansible-playbook run-optimism.yml --extra-vars='{ "target_host": "localhost"}' --user $USER -kK
```
NOTE: By default, deployments are created in an `out` directory. To change this location, update the `l2_directory` variable in the [setup-vars.yml](./setup-vars.yml) file
- For skipping container build, set `"skip_container_build" : true` in the `--extra-vars` parameter:
```bash
LANG=en_US.utf8 ansible-playbook run-optimism.yml --extra-vars='{"target_host" : "localhost", "skip_container_build": true}' --user $USER -kK
```
- To run using existing contracts deployment:
- Update `artifact_path` in [`setup-vars.yml`](./setup-vars.yml) file with path to data directory of the existing deployment
- Run the ansible playbook with `"existing_contracts_deployment": true` in the `--extra-vars` parameter:
```bash
LANG=en_US.utf8 ansible-playbook run-optimism.yml --extra-vars='{"target_host" : "localhost", "existing_contracts_deployment": true}' --user $USER -kK
```
### On Remote Host
To run the playbook on a remote host:
- Create a new `hosts.ini` file:
```bash
cp ../hosts.example.ini hosts.ini
```
- Edit the [`hosts.ini`](./hosts.ini) file to run the playbook on a remote machine:
```ini
[<deployment_host>]
<host_name> ansible_host=<target_ip> ansible_user=<ssh_user> ansible_ssh_common_args='-o ForwardAgent=yes'
```
- Replace `<deployment_host>` with `l2_host`
- Replace `<host_name>` with the alias of your choice
- Replace `<target_ip>` with the IP address or hostname of the target machine
- Replace `<ssh_user>` with the SSH username (e.g., dev, ubuntu)
- Verify that you are able to connect to the host using the following command
```bash
ansible all -m ping -i hosts.ini -k
# Expected output:
# <host_name> | SUCCESS => {
# "ansible_facts": {
# "discovered_interpreter_python": "/usr/bin/python3.10"
# },
# "changed": false,
# "ping": "pong"
# }
```
- Execute the `run-optimism.yml` Ansible playbook for remote deployment:
```bash
LANG=en_US.utf8 ansible-playbook -i hosts.ini run-optimism.yml --extra-vars='{ "target_host": "l2_host"}' --user $USER -kK
```
- For skipping container build, set `"skip_container_build" : true` in the `--extra-vars` parameter:
```bash
LANG=en_US.utf8 ansible-playbook -i hosts.ini run-optimism.yml --extra-vars='{"target_host" : "l2_host", "skip_container_build": true}' --user $USER -kK
```
- To run using existing contracts deployment:
- Update `artifact_path` in [`setup-vars.yml`](./setup-vars.yml) file with path to data directory of the existing deployment
- Run the ansible playbook with `"existing_contracts_deployment": true` in the `--extra-vars` parameter:
```bash
LANG=en_US.utf8 ansible-playbook -i hosts.ini run-optimism.yml --extra-vars='{"target_host" : "l2_host", "existing_contracts_deployment": true}' --user $USER -kK
```
## Check Deployment Status
Run the following commands in the directory where the optimism-deployment is created:
- Follow optimism contracts deployment logs:
```bash
laconic-so deployment --dir optimism-deployment logs -f fixturenet-optimism-contracts
```
- Check L2 logs:
```bash
laconic-so deployment --dir optimism-deployment logs -f op-geth
# Ensure new blocks are getting created
```

View File

@ -0,0 +1,9 @@
l1_chain_id: ""
l1_rpc: ""
l1_host: ""
l1_port: ""
l1_beacon: ""
l1_address: ""
l1_priv_key: ""
proposer_amount: "0.2ether"
batcher_amount: "0.1ether"

97
l2-setup/run-optimism.yml Normal file
View File

@ -0,0 +1,97 @@
- name: Setup L2 on host
hosts: "{{ target_host }}"
vars_files:
- setup-vars.yml
- l2-vars.yml
environment:
PATH: "{{ ansible_env.PATH }}:/home/{{ansible_user}}/bin"
tasks:
- name: Create directory for L2
file:
path: "{{ l2_directory }}"
state: directory
- name: Change owner of l2-directory
file:
path: "{{ l2_directory }}"
owner: "{{ansible_user}}"
group: "{{ansible_user}}"
state: directory
recurse: yes
- name: Clone fixturenet-optimism-stack
expect:
command: laconic-so fetch-stack git.vdb.to/cerc-io/fixturenet-optimism-stack --pull
responses:
"Are you sure you want to continue connecting \\(yes/no/\\[fingerprint\\]\\)\\?": "yes"
timeout: 300
ignore_errors: yes
- name: Clone required repositories for fixturenet-optimism
expect:
command: laconic-so --stack {{ ansible_env.HOME }}/cerc/fixturenet-optimism-stack/stack/fixturenet-optimism setup-repositories --pull
responses:
"Are you sure you want to continue connecting \\(yes/no/\\[fingerprint\\]\\)\\?": "yes"
timeout: 300
ignore_errors: yes
- name: Build container images for L2
command: laconic-so --stack ~/cerc/fixturenet-optimism-stack/stack/fixturenet-optimism build-containers --force-rebuild
when: not skip_container_build
- name: Generate spec file for L2 deployment
template:
src: "./templates/specs/l2-spec.yml.j2"
dest: "{{ l2_directory }}/optimism-spec.yml"
- name: Check if the deployment directory exists for L2
stat:
path: "{{ l2_directory }}/optimism-deployment"
register: l2_deployment_dir
- name: Create a deployment from the spec file for L2
command: laconic-so --stack ~/cerc/fixturenet-optimism-stack/stack/fixturenet-optimism deploy create --spec-file optimism-spec.yml --deployment-dir optimism-deployment
args:
chdir: "{{ l2_directory }}"
when: not l2_deployment_dir.stat.exists
- name: Generate config.env for L2 deployment
template:
src: "./templates/configs/l2-config.env.j2"
dest: "{{ l2_directory }}/optimism-deployment/config.env"
- name: Copy deployed contract addresses and configuration files
block:
- name: Copy l1 deployment file
copy:
src: "{{ artifact_path }}/l1_deployment/{{ l1_chain_id }}-deploy.json"
dest: "{{ l2_directory }}/optimism-deployment/data/l1_deployment"
remote_src: "{{ target_host != 'localhost' }}"
- name: Copy l2 configuration file
copy:
src: "{{ artifact_path }}/l2_config/{{ l1_chain_id }}.json"
dest: "{{ l2_directory }}/optimism-deployment/data/l2_config"
remote_src: "{{ target_host != 'localhost' }}"
- name: Copy allocs-l2 file
copy:
src: "{{ artifact_path }}/l2_config/allocs-l2.json"
dest: "{{ l2_directory }}/optimism-deployment/data/l2_config"
remote_src: "{{ target_host != 'localhost' }}"
- name: Copy l2 accounts file
copy:
src: "{{ artifact_path }}/l2_accounts/accounts.json"
dest: "{{ l2_directory }}/optimism-deployment/data/l2_accounts"
remote_src: "{{ target_host != 'localhost' }}"
when: existing_contracts_deployment
- name: Start L2-deployment
command: laconic-so deployment --dir optimism-deployment start
args:
chdir: "{{ l2_directory }}"

4
l2-setup/setup-vars.yml Normal file
View File

@ -0,0 +1,4 @@
skip_container_build: false
l2_directory: "out"
existing_contracts_deployment: false
artifact_path: ""

View File

@ -0,0 +1,9 @@
CERC_L1_CHAIN_ID={{ l1_chain_id }}
CERC_L1_RPC={{ l1_rpc }}
CERC_L1_HOST={{ l1_host }}
CERC_L1_PORT={{ l1_port }}
CERC_L1_BEACON={{ l1_beacon }}
CERC_L1_ADDRESS={{ l1_address }}
CERC_L1_PRIV_KEY={{ l1_priv_key }}
CERC_PROPOSER_AMOUNT={{ proposer_amount }}
CERC_BATCHER_AMOUNT={{ batcher_amount }}

View File

@ -0,0 +1,18 @@
stack: /home/{{ansible_user}}/cerc/fixturenet-optimism-stack/stack/fixturenet-optimism
deploy-to: compose
network:
ports:
op-geth:
- '9545:8545'
- '9546:8546'
op-node:
- '8547'
op-batcher:
- '8548'
op-proposer:
- '8560'
volumes:
l1_deployment: ./data/l1_deployment
l2_accounts: ./data/l2_accounts
l2_config: ./data/l2_config
l2_geth_data: ./data/l2_geth_data

3
nitro-bridge-setup/.gitignore vendored Normal file
View File

@ -0,0 +1,3 @@
out
bridge-vars.yml
hosts.ini

View File

@ -0,0 +1,140 @@
# nitro-bridge-setup
## Setup Ansible
To get started, follow the [installation](../README.md#installation) guide to setup ansible on your machine
## Setup
The following commands have to be executed in the [`nitro-bridge-setup`](./) directory:
- Copy the `bridge-vars.example.yml` vars file:
```bash
cp bridge-vars.example.yml bridge-vars.yml
```
- Edit [`bridge-vars.yml`](./bridge-vars.yml) with the required values:
```yaml
# L1 WS endpoint
nitro_l1_chain_url: ""
# L2 WS endpoint
nitro_l2_chain_url: ""
# Private key for the bridge's nitro address
nitro_sc_pk: ""
# Private key for a funded account on L1
# This account should have tokens for funding Nitro channels
nitro_chain_pk: ""
# L2 chain ID
optimism_chain_id: ""
# L2 RPC endpoint
optimism_url: ""
# Custom L2 token to be deployed
token_name: "LaconicNetworkToken"
token_symbol: "LNT"
initial_token_supply: "129600"
# Addresses of the deployed nitro contracts
na_address: ""
vpa_address: ""
ca_address: ""
# Address of deployed custom L1 token
l1_asset_address: ""
```
## Run Nitro Bridge
### On Local Host
- To setup and run nitro bridge locally, execute the `run-nitro-bridge.yml` Ansible playbook:
```bash
LANG=en_US.utf8 ansible-playbook run-nitro-bridge.yml --extra-vars='{ "target_host": "localhost"}' --user $USER -kK
```
NOTE: By default, deployments are created in an `out` directory. To change this location, update the `nitro_directory` variable in the [setup-vars.yml](./setup-vars.yml) file
- For skipping container build, set `"skip_container_build" : true` in the `--extra-vars` parameter:
```bash
LANG=en_US.utf8 ansible-playbook run-nitro-bridge.yml --extra-vars='{"target_host" : "localhost", "skip_container_build": true}' --user $USER -kK
```
### On Remote Host
To run the playbook on a remote host:
- Create a new `hosts.ini` file:
```bash
cp ../hosts.example.ini hosts.ini
```
- Edit the [`hosts.ini`](./hosts.ini) file to run the playbook on a remote machine:
```ini
[<deployment_host>]
<host_name> ansible_host=<target_ip> ansible_user=<ssh_user> ansible_ssh_common_args='-o ForwardAgent=yes'
```
- Replace `<deployment_host>` with `nitro_host`
- Replace `<host_name>` with the alias of your choice
- Replace `<target_ip>` with the IP address or hostname of the target machine
- Replace `<ssh_user>` with the SSH username (e.g., dev, ubuntu)
- Verify that you are able to connect to the host using the following command
```bash
ansible all -m ping -i hosts.ini -k
# Expected output:
# <host_name> | SUCCESS => {
# "ansible_facts": {
# "discovered_interpreter_python": "/usr/bin/python3.10"
# },
# "changed": false,
# "ping": "pong"
# }
```
- Execute the `run-nitro-bridge.yml` Ansible playbook for remote deployment:
```bash
LANG=en_US.utf8 ansible-playbook -i hosts.ini run-nitro-bridge.yml --extra-vars='{ "target_host": "nitro_host"}' --user $USER -kK
```
- For skipping container build, run with `"skip_container_build" : true` in the `--extra-vars` parameter:
```bash
LANG=en_US.utf8 ansible-playbook -i hosts.ini run-nitro-bridge.yml --extra-vars='{ "target_host": "nitro_host", "skip_container_build": true }' --user $USER -kK
```
## Check Deployment Status
- Run the following command in the directory where the bridge-deployment is created:
- Check logs for deployments:
```bash
# Check the bridge deployment logs, ensure that the node is running
laconic-so deployment --dir bridge-deployment logs nitro-bridge -f
```
## Get Contract Addresses
- Run the following commands in the directory where the deployments are created:
- Get addresses of L2 nitro contracts:
```bash
laconic-so deployment --dir bridge-deployment exec nitro-bridge "cat /app/deployment/nitro-addresses.json"
```

View File

@ -0,0 +1,13 @@
nitro_l1_chain_url: ""
nitro_l2_chain_url: ""
nitro_chain_pk: ""
nitro_sc_pk: ""
optimism_chain_id: ""
optimism_url: ""
token_name: "LaconicNetworkToken"
token_symbol: "LNT"
initial_token_supply: "129600"
na_address: ""
vpa_address: ""
ca_address: ""
l1_asset_address: ""

View File

@ -0,0 +1,79 @@
- name: Setup go-nitro on host
hosts: "{{ target_host }}"
vars_files:
- setup-vars.yml
- bridge-vars.yml
environment:
PATH: "{{ ansible_env.PATH }}:/home/{{ansible_user}}/bin"
tasks:
- name: Create directory for nitro bridge deployment
file:
path: "{{ nitro_directory }}"
state: directory
- name: Change owner of nitro-directory
file:
path: "{{ nitro_directory }}"
owner: "{{ansible_user}}"
group: "{{ansible_user}}"
state: directory
recurse: yes
- name: Clone go-nitro stack repo
expect:
command: laconic-so fetch-stack git.vdb.to/cerc-io/nitro-stack --git-ssh --pull
responses:
"Are you sure you want to continue connecting \\(yes/no/\\[fingerprint\\]\\)\\?": "yes"
timeout: 300
ignore_errors: yes
- name: Clone repositories required for nitro-stack
expect:
command: laconic-so --stack {{ ansible_env.HOME }}/cerc/nitro-stack/stack-orchestrator/stacks/bridge setup-repositories --git-ssh --pull
responses:
"Are you sure you want to continue connecting \\(yes/no/\\[fingerprint\\]\\)\\?": "yes"
timeout: 300
ignore_errors: yes
- name: Build containers
command: laconic-so --stack ~/cerc/nitro-stack/stack-orchestrator/stacks/bridge build-containers --force-rebuild
when: not skip_container_build
- name: Check if deployment exists for bridge node
stat:
path: "{{ nitro_directory }}/bridge-deployment"
register: bridge_deployment
- name: Generate spec file for bridge deployment
template:
src: "./templates/specs/bridge-nitro-spec.yml.j2"
dest: "{{ nitro_directory }}/bridge-nitro-spec.yml"
when: not bridge_deployment.stat.exists
- name: Create a deployment for the bridge node
command: laconic-so --stack ~/cerc/nitro-stack/stack-orchestrator/stacks/bridge deploy create --spec-file bridge-nitro-spec.yml --deployment-dir bridge-deployment
args:
chdir: "{{ nitro_directory }}"
when: not bridge_deployment.stat.exists
- name: Start the nitro bridge
command: laconic-so deployment --dir bridge-deployment start
args:
chdir: "{{ nitro_directory }}"
- name: Wait for nitro-addresses.json to be created
wait_for:
path: "{{ nitro_directory }}/bridge-deployment/data/nitro_deployment/nitro-addresses.json"
state: present
timeout: 180
- name: Print BRIDGE_ADDRESS
shell: laconic-so deployment --dir bridge-deployment exec nitro-bridge "jq -r '.\"{{ optimism_chain_id }}\"[0].contracts.Bridge.address' /app/deployment/nitro-addresses.json"
args:
chdir: "{{ nitro_directory }}"
register: bridge_address
- debug:
msg: "BRIDGE_ADDRESS: {{ bridge_address.stdout }}"

View File

@ -0,0 +1,3 @@
target_host: "localhost"
nitro_directory: out
skip_container_build: false

View File

@ -0,0 +1,28 @@
stack: /home/{{ ansible_user }}/cerc/nitro-stack/stack-orchestrator/stacks/bridge
deploy-to: compose
config:
NITRO_L1_CHAIN_URL: {{ nitro_l1_chain_url }}
NITRO_L2_CHAIN_URL: {{ nitro_l2_chain_url }}
NITRO_CHAIN_PK: {{ nitro_chain_pk }}
NITRO_SC_PK: {{ nitro_sc_pk }}
OPTIMISM_CHAIN_ID: {{ optimism_chain_id }}
OPTIMISM_URL: {{ optimism_url }}
OPTIMISM_DEPLOYER_PK: {{ nitro_chain_pk }}
TOKEN_NAME: {{ token_name }}
TOKEN_SYMBOL: {{ token_symbol }}
INITIAL_TOKEN_SUPPLY: {{ initial_token_supply }}
NA_ADDRESS: "{{ na_address }}"
VPA_ADDRESS: "{{ vpa_address }}"
CA_ADDRESS: "{{ ca_address }}"
L1_ASSET_ADDRESS: "{{ l1_asset_address }}"
network:
ports:
nitro-bridge:
- 0.0.0.0:3005:3005
- 0.0.0.0:3006:3006
- 0.0.0.0:4006:4006
volumes:
nitro_bridge_data: ./data/nitro_bridge_data
nitro_bridge_tls: ./data/nitro_bridge_tls
nitro_node_caroot: ./data/nitro_node_caroot
nitro_deployment: ./data/nitro_deployment

3
nitro-contracts-setup/.gitignore vendored Normal file
View File

@ -0,0 +1,3 @@
out
contract-vars.yml
hosts.ini

View File

@ -0,0 +1,122 @@
# nitro-contracts-setup
## Setup Ansible
To get started, follow the [installation](../README.md#installation) guide to setup ansible on your machine
## Setup
The following commands have to be executed in the [`nitro-contracts-setup`](./) directory:
- Copy the `contract-vars.example.yml` vars file
```bash
cp contract-vars.example.yml contract-vars.yml
```
- Edit [`contract-vars.yml`](./contract-vars.yml) and fill in the following values
```bash
# L1 RPC endpoint
geth_url: ""
# L1 chain ID
geth_chain_id: ""
# Private key for a funded account on L1 to use for contracts deployment on L1
geth_deployer_pk: ""
# Custom L1 token to be deployed
token_name: "LaconicNetworkToken"
token_symbol: "LNT"
initial_token_supply: "129600"
```
## Deploy Contracts
### On Local Host
- To deploy nitro contracts locally, execute the `deploy-contracts.yml` Ansible playbook:
```bash
LANG=en_US.utf8 ansible-playbook deploy-contracts.yml --extra-vars='{ "target_host": "localhost"}' --user $USER -kK
```
NOTE: By default, deployments are created in an `out` directory. To change this location, update the `nitro_directory` variable in the [setup-vars.yml](./setup-vars.yml) file
- For skipping container build, set `"skip_container_build" : true` in the `--extra-vars` parameter:
```bash
LANG=en_US.utf8 ansible-playbook deploy-contracts.yml --extra-vars='{"target_host" : "localhost", "skip_container_build": true}' --user $USER -kK
```
### On Remote Host
To run the playbook on a remote host:
- Create a new `hosts.ini` file:
```bash
cp ../hosts.example.ini hosts.ini
```
- Edit the [`hosts.ini`](./hosts.ini) file to run the playbook on a remote machine:
```ini
[<deployment_host>]
<host_name> ansible_host=<target_ip> ansible_user=<ssh_user> ansible_ssh_common_args='-o ForwardAgent=yes'
```
- Replace `<deployment_host>` with `nitro_host`
- Replace `<host_name>` with the alias of your choice
- Replace `<target_ip>` with the IP address or hostname of the target machine
- Replace `<ssh_user>` with the SSH username (e.g., dev, ubuntu)
- Verify that you are able to connect to the host using the following command
```bash
ansible all -m ping -i hosts.ini -k
# Expected output:
# <host_name> | SUCCESS => {
# "ansible_facts": {
# "discovered_interpreter_python": "/usr/bin/python3.10"
# },
# "changed": false,
# "ping": "pong"
# }
```
- Execute the `deploy-contracts.yml` Ansible playbook for remote deployment:
```bash
LANG=en_US.utf8 ansible-playbook -i hosts.ini deploy-contracts.yml --extra-vars='{ "target_host": "nitro_host"}' --user $USER -kK
```
- For skipping container build, run with `"skip_container_build" : true` in the `--extra-vars` parameter:
```bash
LANG=en_US.utf8 ansible-playbook -i hosts.ini deploy-contracts.yml --extra-vars='{ "target_host": "nitro_host", "skip_container_build": true }' --user $USER -kK
```
## Check Deployment Status
- Run the following command in the directory where the nitro-contracts-deployment is created:
- Check logs for deployments:
```bash
# Check the L2 nitro contract deployment logs
laconic-so deployment --dir nitro-contracts-deployment logs l2-nitro-contracts -f
```
## Get Contract Addresses
- Run the following commands in the directory where the deployments are created:
- Get addresses of L1 nitro contracts:
```bash
laconic-so deployment --dir nitro-contracts-deployment exec nitro-contracts "cat /app/deployment/nitro-addresses.json"
```

View File

@ -0,0 +1,6 @@
geth_url: ""
geth_chain_id: ""
geth_deployer_pk: ""
token_name: ""
token_symbol: ""
initial_token_supply: ""

View File

@ -0,0 +1,102 @@
- name: Deploy nitro contracts
hosts: "{{ target_host }}"
vars_files:
- setup-vars.yml
- contract-vars.yml
environment:
PATH: "{{ ansible_env.PATH }}:/home/{{ansible_user}}/bin"
tasks:
- name: Create directory for nitro bridge deployment
file:
path: "{{ nitro_directory }}"
state: directory
- name: Change owner of nitro-directory
file:
path: "{{ nitro_directory }}"
owner: "{{ansible_user}}"
group: "{{ansible_user}}"
state: directory
recurse: yes
- name: Clone go-nitro stack repo
expect:
command: laconic-so fetch-stack git.vdb.to/cerc-io/nitro-stack --git-ssh --pull
responses:
"Are you sure you want to continue connecting \\(yes/no/\\[fingerprint\\]\\)\\?": "yes"
timeout: 300
ignore_errors: yes
- name: Clone repositories required for nitro-stack
expect:
command: laconic-so --stack ~/cerc/nitro-stack/stack-orchestrator/stacks/bridge setup-repositories --git-ssh --pull
responses:
"Are you sure you want to continue connecting \\(yes/no/\\[fingerprint\\]\\)\\?": "yes"
timeout: 300
ignore_errors: yes
- name: Build containers
command: laconic-so --stack ~/cerc/nitro-stack/stack-orchestrator/stacks/bridge build-containers --force-rebuild
when: not skip_container_build
- name: Generate spec file for nitro contracts deployment
template:
src: "./templates/specs/nitro-contracts-spec.yml.j2"
dest: "{{ nitro_directory }}/nitro-contracts-spec.yml"
- name: Check if deployment exists for nitro contracts
stat:
path: "{{ nitro_directory }}/nitro-contracts-deployment"
register: nitro_contracts_deployment
- name: Create a deployment for nitro contracts
command: laconic-so --stack ~/cerc/nitro-stack/stack-orchestrator/stacks/nitro-contracts deploy create --spec-file nitro-contracts-spec.yml --deployment-dir nitro-contracts-deployment
args:
chdir: "{{ nitro_directory }}"
when: not nitro_contracts_deployment.stat.exists
- name: Start deployment for nitro-contracts
command: laconic-so deployment --dir nitro-contracts-deployment start
args:
chdir: "{{ nitro_directory }}"
- name: Wait for the contracts to be deployed
wait_for:
path: "{{ nitro_directory }}/nitro-contracts-deployment/data/nitro_deployment/nitro-addresses.json"
timeout: 300
- name: Export NA_ADDRESS
shell: laconic-so deployment --dir nitro-contracts-deployment exec nitro-contracts "jq -r '.\"{{ geth_chain_id }}\"[0].contracts.NitroAdjudicator.address' /app/deployment/nitro-addresses.json"
args:
chdir: "{{ nitro_directory }}"
register: na_address
- debug:
msg: "NA_ADDRESS: {{ na_address.stdout }}"
- name: Export CA_ADDRESS
shell: laconic-so deployment --dir nitro-contracts-deployment exec nitro-contracts "jq -r '.\"{{ geth_chain_id }}\"[0].contracts.ConsensusApp.address' /app/deployment/nitro-addresses.json"
args:
chdir: "{{ nitro_directory }}"
register: ca_address
- debug:
msg: "CA_ADDRESS: {{ ca_address.stdout }}"
- name: Export VPA_ADDRESS
shell: laconic-so deployment --dir nitro-contracts-deployment exec nitro-contracts "jq -r '.\"{{ geth_chain_id }}\"[0].contracts.VirtualPaymentApp.address' /app/deployment/nitro-addresses.json"
args:
chdir: "{{ nitro_directory }}"
register: vpa_address
- debug:
msg: "VPA_ADDRESS: {{ vpa_address.stdout }}"
- name: Export L1_ASSET_ADDRESS
shell: laconic-so deployment --dir nitro-contracts-deployment exec nitro-contracts "jq -r '.\"{{ geth_chain_id }}\"[0].contracts.Token.address' /app/deployment/nitro-addresses.json"
args:
chdir: "{{ nitro_directory }}"
register: l1_asset_address
- debug:
msg: "L1_ASSET_ADDRESS: {{ l1_asset_address.stdout }}"

View File

@ -0,0 +1,3 @@
target_host: "localhost"
nitro_directory: out
skip_container_build: false

View File

@ -0,0 +1,13 @@
stack: /home/{{ ansible_user }}/cerc/nitro-stack/stack-orchestrator/stacks/nitro-contracts
deploy-to: compose
config:
GETH_URL: {{ geth_url }}
GETH_CHAIN_ID: {{ geth_chain_id }}
GETH_DEPLOYER_PK: {{ geth_deployer_pk }}
TOKEN_NAME: {{ token_name }}
TOKEN_SYMBOL: {{ token_symbol }}
INITIAL_TOKEN_SUPPLY: {{ initial_token_supply }}
network:
ports: {}
volumes:
nitro_deployment: ./data/nitro_deployment

3
nitro-nodes-setup/.gitignore vendored Normal file
View File

@ -0,0 +1,3 @@
out
nitro-vars.yml
hosts.ini

152
nitro-nodes-setup/README.md Normal file
View File

@ -0,0 +1,152 @@
# nitro-nodes-setup
## Setup Ansible
To get started, follow the [installation](../README.md#installation) guide to setup ansible on your machine
## Setup for Remote Host
To run the playbook on a remote host:
- Follow steps from [setup remote hosts](../README.md#setup-remote-hosts)
- Update / append the [`hosts.ini`](../hosts.ini) file for your remote host with `<deployment_host>` set as `nitro_host`
## Setup
The following commands have to be executed in [`nitro-nodes-setup`](./) directory
- Copy the `nitro-vars.example.yml` vars file
```bash
cp nitro-vars.example.yml nitro-vars.yml
```
- Edit [`nitro-vars.yml`](./nitro-vars.yml) and fill in the following values
```bash
# L1 WS endpoint
nitro_l1_chain_url: ""
# L2 WS endpoint
nitro_l2_chain_url: ""
# Private key for your nitro address
nitro_sc_pk: ""
# Private key of the account on chain that is used for funding channels in Nitro node
nitro_chain_pk: ""
# Contract address of NitroAdjudicator
na_address: ""
# Contract address of VirtualPaymentApp
vpa_address: ""
# Contract address of ConsensusApp
ca_address: ""
# Address of the bridge node
bridge_contract_address: ""
# Multiaddr of the L1 bridge node
nitro_l1_bridge_multiaddr: ""
# Multiaddr of the L2 bridge node
nitro_l2_bridge_multiaddr: ""
# Multiaddr with publically accessible IP address / DNS for your L1 nitro node
# Example: "/ip4/192.168.x.y/tcp/3009"
# Example: "/dns4/example.com/tcp/3009"
nitro_l1_ext_multiaddr: ""
# Multiaddr with publically accessible IP address / DNS for your L2 nitro node
nitro_l2_ext_multiaddr: ""
```
## Run Nitro Node
### On Local Host
- To run a nitro node, execute the `run-nitro-nodes.yml` Ansible playbook by running the following command:
```bash
LANG=en_US.utf8 ansible-playbook run-nitro-nodes.yml --extra-vars='{ "target_host": "localhost"}' --user $USER -kK
```
NOTE: By default, deployments are created in a `out` directory. To change this location, update the `nitro_directory` variable in the [setup-vars.yml](./setup-vars.yml) file
- For skipping container build, run with `"skip_container_build" : true` in the `--extra-vars` parameter:
```bash
LANG=en_US.utf8 ansible-playbook -i hosts.ini run-nitro-nodes.yml --extra-vars='{ "target_host": "nitro_host", "skip_container_build": true }' --user $USER -kK
```
### On Remote Host
To run the playbook on a remote host:
- Create a new `hosts.ini` file:
```bash
cp ../hosts.example.ini hosts.ini
```
- Edit the [`hosts.ini`](./hosts.ini) file to run the playbook on a remote machine:
```ini
[<deployment_host>]
<host_name> ansible_host=<target_ip> ansible_user=<ssh_user> ansible_ssh_common_args='-o ForwardAgent=yes'
```
- Replace `<deployment_host>` with `nitro_host`
- Replace `<host_name>` with the alias of your choice
- Replace `<target_ip>` with the IP address or hostname of the target machine
- Replace `<ssh_user>` with the SSH username (e.g., dev, ubuntu)
- Verify that you are able to connect to the host using the following command
```bash
ansible all -m ping -i hosts.ini -k
# Expected output:
# <host_name> | SUCCESS => {
# "ansible_facts": {
# "discovered_interpreter_python": "/usr/bin/python3.10"
# },
# "changed": false,
# "ping": "pong"
# }
```
- Copy and edit the [`nitro-vars.yml`](./nitro-vars.yml) file as described in the [local setup](./README.md#run-nitro-node-on-local-host) section
- Execute the `run-nitro-nodes.yml` Ansible playbook for remote deployment:
```bash
LANG=en_US.utf8 ansible-playbook -i hosts.ini run-nitro-nodes.yml --extra-vars='{ "target_host": "nitro_host"}' --user $USER -kK
```
- For skipping container build, run with `"skip_container_build" : true` in the `--extra-vars` parameter:
```bash
LANG=en_US.utf8 ansible-playbook -i hosts.ini run-nitro-nodes.yml --extra-vars='{ "target_host": "nitro_host", "skip_container_build": true }' --user $USER -kK
```
## Check Deployment Status
- Run the following command in the directory where the deployments are created
- Check L1 nitro node logs:
```bash
laconic-so deployment --dir l1-nitro-deployment logs nitro-node -f
```
- Check L2 nitro node logs:
```bash
laconic-so deployment --dir l2-nitro-deployment logs nitro-node -f
```

View File

@ -0,0 +1,12 @@
nitro_l1_chain_url: ""
nitro_l2_chain_url: ""
nitro_sc_pk: ""
nitro_chain_pk: ""
na_address: ""
vpa_address: ""
ca_address: ""
bridge_contract_address: ""
nitro_l1_bridge_multiaddr: ""
nitro_l2_bridge_multiaddr: ""
nitro_l1_ext_multiaddr: ""
nitro_l2_ext_multiaddr: ""

View File

@ -0,0 +1,125 @@
- name: Setup and run nitro nodes
hosts: "{{ target_host }}"
vars_files:
- setup-vars.yml
- nitro-vars.yml
environment:
PATH: "{{ ansible_env.PATH }}:/home/{{ansible_user}}/bin"
tasks:
- name: Install yq
get_url:
url: https://github.com/mikefarah/yq/releases/latest/download/yq_linux_amd64
dest: /usr/bin/yq
mode: '0755'
become: yes
- name: Create directory for nitro-stack
file:
path: "{{ nitro_directory }}"
state: directory
- name: Change owner of nitro-directory
file:
path: "{{ nitro_directory }}"
owner: "{{ansible_user}}"
group: "{{ansible_user}}"
state: directory
recurse: yes
- name: Clone go-nitro stack repo
expect:
command: laconic-so fetch-stack git.vdb.to/cerc-io/nitro-stack --git-ssh --pull
responses:
"Are you sure you want to continue connecting \\(yes/no/\\[fingerprint\\]\\)\\?": "yes"
timeout: 300
ignore_errors: yes
- name: Clone repositories required for nitro-stack
expect:
command: laconic-so --stack {{ ansible_env.HOME }}/cerc/nitro-stack/stack-orchestrator/stacks/nitro-node setup-repositories --git-ssh --pull
responses:
"Are you sure you want to continue connecting \\(yes/no/\\[fingerprint\\]\\)\\?": "yes"
timeout: 300
ignore_errors: yes
- name: Build containers
command: laconic-so --stack {{ ansible_env.HOME }}/cerc/nitro-stack/stack-orchestrator/stacks/nitro-node build-containers --force-rebuild
when: not skip_container_build
- name: Generate spec file for L1 nitro node
template:
src: "./templates/specs/l1-nitro-spec.yml.j2"
dest: "{{ nitro_directory }}/l1-nitro-spec.yml"
- name: Generate spec file for L2 nitro node
template:
src: "./templates/specs/l2-nitro-spec.yml.j2"
dest: "{{ nitro_directory }}/l2-nitro-spec.yml"
- name: Check if deployment exists for L1 nitro node
stat:
path: "{{ nitro_directory }}/l1-nitro-deployment"
register: l1_deployment
- name: Create a deployment for L1 nitro node
command: laconic-so --stack ~/cerc/nitro-stack/stack-orchestrator/stacks/nitro-node deploy create --spec-file l1-nitro-spec.yml --deployment-dir l1-nitro-deployment
args:
chdir: "{{ nitro_directory }}"
when: not l1_deployment.stat.exists
- name: Check if deployment exists for L2 nitro node
stat:
path: "{{ nitro_directory }}/l2-nitro-deployment"
register: l2_deployment
- name: Create a deployment for L2 nitro node
command: laconic-so --stack ~/cerc/nitro-stack/stack-orchestrator/stacks/nitro-node deploy create --spec-file l2-nitro-spec.yml --deployment-dir l2-nitro-deployment
args:
chdir: "{{ nitro_directory }}"
when: not l2_deployment.stat.exists
- name: Check if the config.env file is empty for L1 deployment
shell: "[ -s {{ nitro_directory }}/l1-nitro-deployment/config.env ] && echo 'File is not empty' || echo 'File is empty'"
register: file_check_result
- name: Display the result of the file check
debug:
msg: "{{ file_check_result.stdout }}"
- name: Copy config.env for L1 nitro deployment if it is empty
template:
src: "./templates/configs/l1-nitro-config.env.j2"
dest: "{{ nitro_directory }}/l1-nitro-deployment/config.env"
when: file_check_result.stdout == 'File is empty'
- name: Check if the config.env file is empty for L2 deployment
shell: "[ -s {{ nitro_directory }}/l2-nitro-deployment/config.env ] && echo 'File is not empty' || echo 'File is empty'"
register: file_check_result
- name: Display the result of the file check
debug:
msg: "{{ file_check_result.stdout }}"
- name: Copy config.env for L2 nitro deployment if it is empty
template:
src: "./templates/configs/l2-nitro-config.env.j2"
dest: "{{ nitro_directory }}/l2-nitro-deployment/config.env"
when: file_check_result.stdout == 'File is empty'
- name: Start deployment for L1 nitro node
command: laconic-so deployment --dir l1-nitro-deployment start
args:
chdir: "{{ nitro_directory }}"
- name: Start deployment for L2 nitro node
command: laconic-so deployment --dir l2-nitro-deployment start
args:
chdir: "{{ nitro_directory }}"
- name: Fetch the nitro-node-config file
get_url:
url: https://git.vdb.to/cerc-io/testnet-laconicd-stack/raw/branch/main/ops/stage2/nitro-node-config.yml
dest: "{{ nitro_directory }}"

View File

@ -0,0 +1,3 @@
target_host: "localhost"
nitro_directory: out
skip_container_build: false

View File

@ -0,0 +1,8 @@
NITRO_CHAIN_URL={{ nitro_l1_chain_url }}
NITRO_SC_PK={{ nitro_sc_pk }}
NITRO_CHAIN_PK={{ nitro_chain_pk }}
NA_ADDRESS="{{ na_address }}"
VPA_ADDRESS="{{ vpa_address }}"
CA_ADDRESS="{{ ca_address }}"
NITRO_BOOTPEERS={{ nitro_l1_bridge_multiaddr }}
NITRO_EXT_MULTIADDR={{ nitro_l1_ext_multiaddr }}

View File

@ -0,0 +1,10 @@
NITRO_CHAIN_URL={{ nitro_l2_chain_url }}
NITRO_SC_PK={{ nitro_sc_pk }}
NITRO_CHAIN_PK={{ nitro_chain_pk }}
NA_ADDRESS="{{ na_address }}"
VPA_ADDRESS="{{ vpa_address }}"
CA_ADDRESS="{{ ca_address }}"
BRIDGE_ADDRESS="{{ bridge_contract_address }}"
NITRO_BOOTPEERS={{ nitro_l2_bridge_multiaddr }}
NITRO_EXT_MULTIADDR={{ nitro_l2_ext_multiaddr }}
NITRO_L2=true

View File

@ -0,0 +1,11 @@
stack: /home/{{ansible_user}}/cerc/nitro-stack/stack-orchestrator/stacks/nitro-node
deploy-to: compose
network:
ports:
nitro-node:
- 3007:3005
- 4007:4005
volumes:
nitro_node_data: ./data/nitro_node_data
nitro_node_tls: ./data/nitro_node_tls
nitro_node_caroot: ./data/nitro_node_caroot

View File

@ -0,0 +1,11 @@
stack: /home/{{ansible_user}}/cerc/nitro-stack/stack-orchestrator/stacks/nitro-node
deploy-to: compose
network:
ports:
nitro-node:
- 3009:3005
- 4009:4005
volumes:
nitro_node_data: ./data/nitro_node_data
nitro_node_tls: ./data/nitro_node_tls
nitro_node_caroot: ./data/nitro_node_caroot

1
stack-orchestrator-setup/.gitignore vendored Normal file
View File

@ -0,0 +1 @@
hosts.ini

View File

@ -0,0 +1,100 @@
# stack-orchestrator-setup
## Setup Ansible
To get started, follow the [installation](../README.md#installation) guide to setup ansible on your machine.
## Setup Stack Orchestrator
This playbook will install Docker and Stack Orchestrator (laconic-so) on the machine if they aren't already present.
Run the following commands in the [`stack-orchestrator-setup`](./) directory.
### On Local Host
To setup stack orchestrator and docker locally, execute the `setup-laconic-so.yml` Ansible playbook:
```bash
LANG=en_US.utf8 ansible-playbook setup-laconic-so.yml --user $USER -kK
```
### On Remote Host
To run the playbook on a remote host:
- Create a new `hosts.ini` file:
```bash
cp ../hosts.example.ini hosts.ini
```
- Edit the [`hosts.ini`](./hosts.ini) file to run the playbook on a remote machine:
```ini
[deployment_host]
<host_name> ansible_host=<target_ip> ansible_user=<ssh_user> ansible_ssh_common_args='-o ForwardAgent=yes'
```
- Replace `<host_name>` with the alias of your choice
- Replace `<target_ip>` with the IP address or hostname of the target machine
- Replace `<ssh_user>` with the SSH username (e.g., dev, ubuntu)
- Verify that you are able to connect to the host using the following command
```bash
ansible all -m ping -i hosts.ini -k
# Expected output:
# <host_name> | SUCCESS => {
# "ansible_facts": {
# "discovered_interpreter_python": "/usr/bin/python3.10"
# },
# "changed": false,
# "ping": "pong"
# }
```
- Execute the `setup-laconic-so.yml` Ansible playbook for setting up stack orchestrator and docker on a remote machine:
```bash
LANG=en_US.utf8 ansible-playbook setup-laconic-so.yml -i hosts.ini --extra-vars='{ "target_host": "deployment_host"}' --user $USER -kK
```
## Verify Installation
- After the installation is complete, verify if `$HOME/bin` is already included in your PATH by running:
```bash
echo $PATH | grep -q "$HOME/bin" && echo "$HOME/bin is already in PATH" || echo "$HOME/bin is not in PATH"
```
If the command outputs `"$HOME/bin is not in PATH"`, you'll need to add it to your `PATH`.
- To add `$HOME/bin` to your PATH, run the following command:
```bash
export PATH="$HOME/bin:$PATH"
```
- To make this change permanent, add the following line to your shell configuration file (`~/.bashrc` or `~/.zshrc`, depending on your shell):
```bash
# For bash users
echo 'export PATH="$HOME/bin:$PATH"' >> ~/.bashrc
# For zsh users
echo 'export PATH="$HOME/bin:$PATH"' >> ~/.zshrc
```
- Once the PATH is set, verify the installation by running the following commands:
```bash
# Check version of docker
docker --version
# Check version of docker compose
docker compose version
# Check version of Stack Orchestrator
laconic-so version
```

View File

@ -0,0 +1,91 @@
- name: Set up Docker
hosts: "{{ target_host }}"
become: yes
vars:
target_host: "localhost"
docker_gpg_key_url: "https://download.docker.com/linux/ubuntu/gpg"
docker_gpg_key_path: "/etc/apt/keyrings/docker.asc"
tasks:
- name: Check if Docker is installed
command: which docker
register: is_docker_present
ignore_errors: yes
- block:
- name: Exit if docker is present
debug:
msg: "Docker already on host, ending play"
- meta: end_play
when: is_docker_present.rc == 0
- name: Update apt cache
apt:
update_cache: yes
- name: Install prerequisites
apt:
name:
- ca-certificates
- curl
state: present
- name: Ensure keyrings directory exists
file:
path: "/etc/apt/keyrings"
state: directory
mode: '0755'
- name: Download Docker GPG key
get_url:
url: "{{ docker_gpg_key_url }}"
dest: "{{ docker_gpg_key_path }}"
mode: '0644'
- name: Get system architecture
shell: dpkg --print-architecture
register: system_arch
- name: Add Docker repository
apt_repository:
repo: "deb [arch={{ system_arch.stdout }} signed-by={{ docker_gpg_key_path }}] https://download.docker.com/linux/ubuntu {{ ansible_lsb.codename }} stable"
state: present
filename: docker
- name: Update apt cache
apt:
update_cache: yes
- name: Install Docker packages
apt:
name:
- docker-ce
- docker-ce-cli
- containerd.io
- docker-buildx-plugin
- docker-compose-plugin
state: latest
notify: Restart Docker
- name: Add user to docker group
user:
name: "{{ ansible_user }}"
groups: docker
append: true
- name: Verify Docker installation by running the hello world image
command: docker run hello-world
register: hello_world_output
ignore_errors: true
become: no
- name: Display hello-world output
debug:
var: hello_world_output.stdout_lines
handlers:
- name: Restart Docker
service:
name: docker
state: restarted

View File

@ -0,0 +1,57 @@
- name: Install Docker if it isn't present
import_playbook: setup-docker.yml
- name: Set up Stack Orchestrator
hosts: "{{ target_host }}"
vars:
target_host: "localhost"
environment:
PATH: "{{ ansible_env.PATH }}:/home/{{ansible_user}}/bin"
tasks:
- name: Check if Stack Orchestrator is installed
shell: which laconic-so
register: is_so_present
ignore_errors: yes
- block:
- name: Exit if Stack Orchestrator is present
debug:
msg: "Stack Orchestrator already on host, ending play"
- meta: end_play
when: is_so_present.rc == 0
- name: Install jq
apt:
name: jq
state: present
update_cache: yes
become: yes
- name: Ensure that directory ~/bin exists and is writable
file:
path: "{{ ansible_env.HOME }}/bin"
state: directory
mode: '0755'
- name: Download the laconic-so binary
get_url:
url: https://git.vdb.to/cerc-io/stack-orchestrator/releases/download/latest/laconic-so
dest: "{{ ansible_env.HOME }}/bin/laconic-so"
mode: '0755'
force: yes
- name: Ensure ~/.laconic-so directory exists
file:
path: "{{ ansible_env.HOME }}/.laconic-so"
state: directory
mode: '0755'
- name: Save the distribution url to ~/.laconic-so directory
copy:
dest: "{{ ansible_env.HOME }}/.laconic-so/config.yml"
content: |
distribution-url: https://git.vdb.to/cerc-io/stack-orchestrator/releases/download/latest/laconic-so
mode: '0644'