diff --git a/app/base.py b/app/base.py index a7721141..abb1fa9d 100644 --- a/app/base.py +++ b/app/base.py @@ -15,7 +15,7 @@ import os from abc import ABC, abstractmethod -from .deploy import get_stack_status +from app.deploy import get_stack_status from decouple import config diff --git a/app/build_containers.py b/app/build_containers.py index 3523dbc3..b3b6295e 100644 --- a/app/build_containers.py +++ b/app/build_containers.py @@ -27,8 +27,8 @@ import subprocess import click import importlib.resources from pathlib import Path -from .util import include_exclude_check, get_parsed_stack_config -from .base import get_npm_registry_url +from app.util import include_exclude_check, get_parsed_stack_config +from app.base import get_npm_registry_url # TODO: find a place for this # epilog="Config provided either in .env or settings.ini or env vars: CERC_REPO_BASE_DIR (defaults to ~/cerc)" @@ -67,7 +67,7 @@ def command(ctx, include, exclude, force_rebuild, extra_build_args): print('Dev root directory doesn\'t exist, creating') # See: https://stackoverflow.com/a/20885799/1701505 - from . import data + from app import data with importlib.resources.open_text(data, "container-image-list.txt") as container_list_file: all_containers = container_list_file.read().splitlines() diff --git a/app/build_npms.py b/app/build_npms.py index d56b2774..6555ba91 100644 --- a/app/build_npms.py +++ b/app/build_npms.py @@ -25,8 +25,8 @@ from decouple import config import click import importlib.resources from python_on_whales import docker, DockerException -from .base import get_stack -from .util import include_exclude_check, get_parsed_stack_config +from app.base import get_stack +from app.util import include_exclude_check, get_parsed_stack_config builder_js_image_name = "cerc/builder-js:local" @@ -81,7 +81,7 @@ def command(ctx, include, exclude, force_rebuild, extra_build_args): os.makedirs(build_root_path) # See: https://stackoverflow.com/a/20885799/1701505 - from . import data + from app import data with importlib.resources.open_text(data, "npm-package-list.txt") as package_list_file: all_packages = package_list_file.read().splitlines() diff --git a/app/data/compose/docker-compose-mainnet-laconicd.yml b/app/data/compose/docker-compose-mainnet-laconicd.yml index 4de68d7b..78d2cd2f 100644 --- a/app/data/compose/docker-compose-mainnet-laconicd.yml +++ b/app/data/compose/docker-compose-mainnet-laconicd.yml @@ -2,7 +2,7 @@ services: laconicd: restart: no image: cerc/laconicd:local - command: ["sh", "/docker-entrypoint-scripts.d/create-fixturenet.sh"] + command: ["/bin/sh", "-c", "while :; do sleep 600; done"] volumes: # The cosmos-sdk node's database directory: - laconicd-data:/root/.laconicd/data diff --git a/app/data/stacks/mainnet-laconic/deploy/commands.py b/app/data/stacks/mainnet-laconic/deploy/commands.py index 0ac4845f..a8a62bd7 100644 --- a/app/data/stacks/mainnet-laconic/deploy/commands.py +++ b/app/data/stacks/mainnet-laconic/deploy/commands.py @@ -13,45 +13,50 @@ # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see . -import click -import os -from shutil import copyfile -import sys -from .util import get_stack_config_filename, get_parsed_deployment_spec +from dataclasses import dataclass +from app.util import get_yaml +from app.stack_state import State -default_spec_file_content = """stack: mainnet-laconic -data_dir: /my/path -node_name: my-node-name +default_spec_file_content = """config: + node_moniker: my-node-name + chain_id: my-chain-id """ +init_help_text = """Add helpful text here on setting config variables. +""" -def make_default_deployment_dir(): - return "deployment-001" - -@click.command() -@click.option("--output", required=True, help="Write yaml spec file here") -@click.pass_context -def init(ctx, output): - with open(output, "w") as output_file: - output_file.write(default_spec_file_content) +@dataclass +class VolumeMapping: + host_path: str + container_path: str -@click.command() -@click.option("--spec-file", required=True, help="Spec file to use to create this deployment") -@click.option("--deployment-dir", help="Create deployment files in this directory") -@click.pass_context -def create(ctx, spec_file, deployment_dir): - # This function fails with a useful error message if the file doens't exist - parsed_spec = get_parsed_deployment_spec(spec_file) - if ctx.debug: - print(f"parsed spec: {parsed_spec}") - if deployment_dir is None: - deployment_dir = make_default_deployment_dir() - if os.path.exists(deployment_dir): - print(f"Error: {deployment_dir} already exists") - sys.exit(1) - os.mkdir(deployment_dir) - # Copy spec file and the stack file into the deployment dir - copyfile(spec_file, os.path.join(deployment_dir, os.path.basename(spec_file))) - stack_file = get_stack_config_filename(parsed_spec.stack) - copyfile(stack_file, os.path.join(deployment_dir, os.path.basename(stack_file))) +# In order to make this, we need the ability to run the stack +# In theory we can make this same way as we would run deploy up +def run_container_command(ctx, ontainer, command, mounts): + deploy_context = ctx.obj + pass + + +def setup(ctx): + node_moniker = "dbdb-node" + chain_id = "laconic_81337-1" + mounts = [ + VolumeMapping("./path", "~/.laconicd") + ] + output, status = run_container_command(ctx, "laconicd", f"laconicd init {node_moniker} --chain-id {chain_id}", mounts) + + +def init(command_context): + print(init_help_text) + yaml = get_yaml() + return yaml.load(default_spec_file_content) + + +def get_state(command_context): + print("Here we get state") + return State.CONFIGURED + + +def change_state(command_context): + pass diff --git a/app/data/stacks/mainnet-laconic/stack.yml b/app/data/stacks/mainnet-laconic/stack.yml index b5e1f16c..51b0b8a2 100644 --- a/app/data/stacks/mainnet-laconic/stack.yml +++ b/app/data/stacks/mainnet-laconic/stack.yml @@ -25,7 +25,4 @@ containers: pods: - mainnet-laconicd - fixturenet-laconic-console -config: - cli: - key: laconicd.mykey - address: laconicd.myaddress + diff --git a/app/deploy.py b/app/deploy.py index 3f769f3e..41af70f6 100644 --- a/app/deploy.py +++ b/app/deploy.py @@ -26,9 +26,10 @@ import subprocess from python_on_whales import DockerClient, DockerException import click from pathlib import Path -from .util import include_exclude_check, get_parsed_stack_config, global_options2 -from .deployment_create import create as deployment_create -from .deployment_create import init as deployment_init +from app.util import include_exclude_check, get_parsed_stack_config, global_options2 +from app.deployment_create import create as deployment_create +from app.deployment_create import init as deployment_init +from app.deployment_create import setup as deployment_setup class DeployCommandContext(object): @@ -263,7 +264,7 @@ def _make_cluster_context(ctx, stack, include, exclude, cluster, env_file): print(f"Using cluster name: {cluster}") # See: https://stackoverflow.com/a/20885799/1701505 - from . import data + from app import data with resources.open_text(data, "pod-list.txt") as pod_list_file: all_pods = pod_list_file.read().splitlines() @@ -420,3 +421,4 @@ def _orchestrate_cluster_config(ctx, cluster_config, docker, container_exec_env) command.add_command(deployment_init) command.add_command(deployment_create) +command.add_command(deployment_setup) diff --git a/app/deployment.py b/app/deployment.py index eef12222..f8702f28 100644 --- a/app/deployment.py +++ b/app/deployment.py @@ -17,8 +17,8 @@ import click from dataclasses import dataclass from pathlib import Path import sys -from .deploy import up_operation, down_operation, ps_operation, port_operation, exec_operation, logs_operation, create_deploy_context -from .util import global_options +from app.deploy import up_operation, down_operation, ps_operation, port_operation, exec_operation, logs_operation, create_deploy_context +from app.util import global_options @dataclass @@ -128,15 +128,3 @@ def logs(ctx, extra_args): @click.pass_context def status(ctx): print(f"Context: {ctx.parent.obj}") - - - -#from importlib import resources, util -# TODO: figure out how to do this dynamically -#stack = "mainnet-laconic" -#module_name = "commands" -#spec = util.spec_from_file_location(module_name, "./app/data/stacks/" + stack + "/deploy/commands.py") -#imported_stack = util.module_from_spec(spec) -#spec.loader.exec_module(imported_stack) -#command.add_command(imported_stack.init) -#command.add_command(imported_stack.create) diff --git a/app/deployment_create.py b/app/deployment_create.py index fb83f39b..e42920cc 100644 --- a/app/deployment_create.py +++ b/app/deployment_create.py @@ -14,20 +14,12 @@ # along with this program. If not, see . import click +from importlib import util import os from pathlib import Path from shutil import copyfile, copytree import sys -import ruamel.yaml -from .util import get_stack_file_path, get_parsed_deployment_spec, get_parsed_stack_config, global_options - - -def _get_yaml(): - # See: https://stackoverflow.com/a/45701840/1701505 - yaml = ruamel.yaml.YAML() - yaml.preserve_quotes = True - yaml.indent(sequence=3, offset=1) - return yaml +from app.util import get_stack_file_path, get_parsed_deployment_spec, get_parsed_stack_config, global_options, get_yaml def _make_default_deployment_dir(): @@ -47,7 +39,7 @@ def _get_named_volumes(stack): named_volumes = [] parsed_stack = get_parsed_stack_config(stack) pods = parsed_stack["pods"] - yaml = _get_yaml() + yaml = get_yaml() for pod in pods: pod_file_path = os.path.join(_get_compose_file_dir(), f"docker-compose-{pod}.yml") parsed_pod_file = yaml.load(open(pod_file_path, "r")) @@ -96,6 +88,29 @@ def _fixup_pod_file(pod, spec, compose_dir): pod["volumes"][volume] = new_volume_spec +def call_stack_deploy_init(stack): + # Link with the python file in the stack + # Call a function in it + # If no function found, return None + python_file_path = get_stack_file_path(stack).parent.joinpath("deploy", "commands.py") + spec = util.spec_from_file_location("commands", python_file_path) + imported_stack = util.module_from_spec(spec) + spec.loader.exec_module(imported_stack) + return imported_stack.init(None) + + +# TODO: fold this with function above +def call_stack_deploy_setup(stack): + # Link with the python file in the stack + # Call a function in it + # If no function found, return None + python_file_path = get_stack_file_path(stack).parent.joinpath("deploy", "commands.py") + spec = util.spec_from_file_location("commands", python_file_path) + imported_stack = util.module_from_spec(spec) + spec.loader.exec_module(imported_stack) + return imported_stack.setup(None) + + # Inspect the pod yaml to find config files referenced in subdirectories # other than the one associated with the pod def _find_extra_config_dirs(parsed_pod_file, pod): @@ -118,17 +133,19 @@ def _find_extra_config_dirs(parsed_pod_file, pod): @click.option("--output", required=True, help="Write yaml spec file here") @click.pass_context def init(ctx, output): - yaml = _get_yaml() + yaml = get_yaml() stack = global_options(ctx).stack verbose = global_options(ctx).verbose + default_spec_file_content = call_stack_deploy_init(stack) spec_file_content = {"stack": stack} + spec_file_content.update(default_spec_file_content) if verbose: print(f"Creating spec file for stack: {stack}") named_volumes = _get_named_volumes(stack) if named_volumes: volume_descriptors = {} for named_volume in named_volumes: - volume_descriptors[named_volume] = f"../data/{named_volume}" + volume_descriptors[named_volume] = f"./data/{named_volume}" spec_file_content["volumes"] = volume_descriptors with open(output, "w") as output_file: yaml.dump(spec_file_content, output_file) @@ -160,7 +177,7 @@ def create(ctx, spec_file, deployment_dir): destination_compose_dir = os.path.join(deployment_dir, "compose") os.mkdir(destination_compose_dir) data_dir = Path(__file__).absolute().parent.joinpath("data") - yaml = _get_yaml() + yaml = get_yaml() for pod in pods: pod_file_path = os.path.join(_get_compose_file_dir(), f"docker-compose-{pod}.yml") parsed_pod_file = yaml.load(open(pod_file_path, "r")) @@ -180,3 +197,16 @@ def create(ctx, spec_file, deployment_dir): # If the same config dir appears in multiple pods, it may already have been copied if not os.path.exists(destination_config_dir): copytree(source_config_dir, destination_config_dir) + + +@click.command() +@click.option("--node-moniker", help="Help goes here") +@click.option("--key-name", help="Help goes here") +@click.option("--initialize-network", is_flag=True, default=False, help="Help goes here") +@click.option("--join-network", is_flag=True, default=False, help="Help goes here") +@click.option("--create-network", is_flag=True, default=False, help="Help goes here") +@click.pass_context +def setup(ctx, node_moniker, key_name, initialize_network, join_network, create_network): + stack = global_options(ctx).stack + call_stack_deploy_setup(stack) + diff --git a/app/setup_repositories.py b/app/setup_repositories.py index db0bd779..d275a986 100644 --- a/app/setup_repositories.py +++ b/app/setup_repositories.py @@ -25,7 +25,7 @@ import click import importlib.resources from pathlib import Path import yaml -from .util import include_exclude_check +from app.util import include_exclude_check class GitProgress(git.RemoteProgress): @@ -227,7 +227,7 @@ def command(ctx, include, exclude, git_ssh, check_only, pull, branches, branches os.makedirs(dev_root_path) # See: https://stackoverflow.com/a/20885799/1701505 - from . import data + from app import data with importlib.resources.open_text(data, "repository-list.txt") as repository_list_file: all_repos = repository_list_file.read().splitlines() diff --git a/app/stack_state.py b/app/stack_state.py new file mode 100644 index 00000000..830a47f7 --- /dev/null +++ b/app/stack_state.py @@ -0,0 +1,22 @@ +# Copyright © 2023 Cerc + +# This program is free software: you can redistribute it and/or modify +# it under the terms of the GNU Affero General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. + +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU Affero General Public License for more details. + +# You should have received a copy of the GNU Affero General Public License +# along with this program. If not, see . + +from enum import Enum + +class State(Enum): + CREATED = 1 + CONFIGURED = 2 + STARTED = 3 + STOPPED = 4 diff --git a/app/util.py b/app/util.py index 69eda4af..2b12cfbc 100644 --- a/app/util.py +++ b/app/util.py @@ -15,7 +15,7 @@ import os.path import sys -import yaml +import ruamel.yaml from pathlib import Path @@ -42,7 +42,7 @@ def get_parsed_stack_config(stack): stack_file_path = stack if isinstance(stack, os.PathLike) else get_stack_file_path(stack) try: with stack_file_path: - stack_config = yaml.safe_load(open(stack_file_path, "r")) + stack_config = get_yaml().load(open(stack_file_path, "r")) return stack_config except FileNotFoundError as error: # We try here to generate a useful diagnostic error @@ -60,7 +60,7 @@ def get_parsed_deployment_spec(spec_file): spec_file_path = Path(spec_file) try: with spec_file_path: - deploy_spec = yaml.safe_load(open(spec_file_path, "r")) + deploy_spec = get_yaml().load(open(spec_file_path, "r")) return deploy_spec except FileNotFoundError as error: # We try here to generate a useful diagnostic error @@ -69,6 +69,14 @@ def get_parsed_deployment_spec(spec_file): sys.exit(1) +def get_yaml(): + # See: https://stackoverflow.com/a/45701840/1701505 + yaml = ruamel.yaml.YAML() + yaml.preserve_quotes = True + yaml.indent(sequence=3, offset=1) + return yaml + + # TODO: this is fragile wrt to the subcommand depth # See also: https://github.com/pallets/click/issues/108 def global_options(ctx): diff --git a/app/version.py b/app/version.py index 4194f24a..7af18dc1 100644 --- a/app/version.py +++ b/app/version.py @@ -22,7 +22,7 @@ def command(ctx): '''print tool version''' # See: https://stackoverflow.com/a/20885799/1701505 - from . import data + from app import data with importlib.resources.open_text(data, "build_tag.txt") as version_file: # TODO: code better version that skips comment lines version_string = version_file.read().splitlines()[1]