From f57b8b4ba0fe7f509a98d554e82ba0fc64281144 Mon Sep 17 00:00:00 2001 From: David Boreham Date: Wed, 8 Mar 2023 17:03:14 -0700 Subject: [PATCH] Add very basic cluster config mechanism Former-commit-id: 9cae493458f15c9528b45bc116d2d7563263ef8f --- app/data/stacks/fixturenet-laconicd/stack.yml | 3 + app/data/version.txt | 2 +- app/deploy_system.py | 56 ++++++++++++++++++- 3 files changed, 57 insertions(+), 4 deletions(-) diff --git a/app/data/stacks/fixturenet-laconicd/stack.yml b/app/data/stacks/fixturenet-laconicd/stack.yml index fca9623..86fb214 100644 --- a/app/data/stacks/fixturenet-laconicd/stack.yml +++ b/app/data/stacks/fixturenet-laconicd/stack.yml @@ -13,3 +13,6 @@ containers: - cerc/laconic-registry-cli pods: - fixturenet-laconicd +config: + cli: + key: laconicd.mykey diff --git a/app/data/version.txt b/app/data/version.txt index 857da7c..e913786 100644 --- a/app/data/version.txt +++ b/app/data/version.txt @@ -1,2 +1,2 @@ # This file should be re-generated running: scripts/update-version-file.sh script -v1.0.24-d80fb5d +v1.0.25-75376d7 diff --git a/app/deploy_system.py b/app/deploy_system.py index d242fe6..951e375 100644 --- a/app/deploy_system.py +++ b/app/deploy_system.py @@ -19,6 +19,7 @@ import hashlib import copy import os import sys +from dataclasses import dataclass from decouple import config import subprocess from python_on_whales import DockerClient @@ -66,6 +67,9 @@ def command(ctx, include, exclude, cluster, command, extra_args): docker.compose.up(detach=True, services=extra_args_list) for post_start_command in cluster_context.post_start_commands: _run_command(ctx.obj, cluster_context.cluster, post_start_command) + + _orchestrate_cluster_config(ctx.obj, cluster_context.config, docker, container_exec_env) + elif command == "down": if verbose: print("Running compose down") @@ -114,7 +118,7 @@ def command(ctx, include, exclude, cluster, command, extra_args): elif command == "logs": if verbose: print("Running compose logs") - logs_output = docker.compose.logs(services=extra_args_list if extra_args_list != None else []) + logs_output = docker.compose.logs(services=extra_args_list if extra_args_list is not None else []) print(logs_output) @@ -219,15 +223,18 @@ def _make_cluster_context(ctx, include, exclude, cluster): if ctx.verbose: print(f"files: {compose_files}") - return cluster_context(cluster, compose_files, pre_start_commands, post_start_commands) + cluster_config = stack_config['config'] if 'config' in stack_config else None + + return cluster_context(cluster, compose_files, pre_start_commands, post_start_commands, cluster_config) class cluster_context: - def __init__(self, cluster, compose_files, pre_start_commands, post_start_commands) -> None: + def __init__(self, cluster, compose_files, pre_start_commands, post_start_commands, config) -> None: self.cluster = cluster self.compose_files = compose_files self.pre_start_commands = pre_start_commands self.post_start_commands = post_start_commands + self.config = config def _convert_to_new_format(old_pod_array): @@ -258,3 +265,46 @@ def _run_command(ctx, cluster_name, command): if command_result.returncode != 0: print(f"FATAL Error running command: {command}") sys.exit(1) + + +def _orchestrate_cluster_config(ctx, cluster_config, docker, container_exec_env): + + @dataclass + class ConfigDirective: + source_container: str + source_variable: str + destination_container: str + destination_variable: str + + + if cluster_config is not None: + for container in cluster_config: + container_config = cluster_config[container] + if ctx.verbose: + print(f"{container} config: {container_config}") + for directive in container_config: + pd = ConfigDirective( + container_config[directive].split(".")[0], + container_config[directive].split(".")[1], + container, + directive + ) + if ctx.verbose: + print(f"Setting {pd.destination_container}.{pd.destination_variable} = {pd.source_container}.{pd.source_variable}") + # TODO: fix the script paths so they're consistent between containers + source_value = docker.compose.execute(pd.source_container, + ["sh", "-c", + f"sh /docker-entrypoint-scripts.d/export-{pd.source_variable}.sh"], + tty=False, + envs=container_exec_env) + # TODO: handle the case that the value is not yet available + if ctx.debug: + print(f"fetched source value: {source_value}") + destination_output = docker.compose.execute(pd.destination_container, + ["sh", "-c", + f"sh /scripts/import-{pd.destination_variable}.sh {pd.source_variable}"], + tty=False, + envs=container_exec_env) + if ctx.debug: + print(f"destination output: {destination_output}") + # TODO: detect errors here