forked from cerc-io/stack-orchestrator
Generate a unique deployment id for each deployment (#680)
* Move cluster name generation into a function * Generate a unique deployment id for each deployment
This commit is contained in:
parent
6bef0c5b2f
commit
15faed00de
@ -13,12 +13,16 @@
|
|||||||
# You should have received a copy of the GNU Affero General Public License
|
# You should have received a copy of the GNU Affero General Public License
|
||||||
# along with this program. If not, see <http:#www.gnu.org/licenses/>.
|
# along with this program. If not, see <http:#www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
cluster_name_prefix = "laconic-"
|
||||||
stack_file_name = "stack.yml"
|
stack_file_name = "stack.yml"
|
||||||
spec_file_name = "spec.yml"
|
spec_file_name = "spec.yml"
|
||||||
config_file_name = "config.env"
|
config_file_name = "config.env"
|
||||||
|
deployment_file_name = "deployment.yml"
|
||||||
|
compose_dir_name = "compose"
|
||||||
compose_deploy_type = "compose"
|
compose_deploy_type = "compose"
|
||||||
k8s_kind_deploy_type = "k8s-kind"
|
k8s_kind_deploy_type = "k8s-kind"
|
||||||
k8s_deploy_type = "k8s"
|
k8s_deploy_type = "k8s"
|
||||||
|
cluster_id_key = "cluster-id"
|
||||||
kube_config_key = "kube-config"
|
kube_config_key = "kube-config"
|
||||||
deploy_to_key = "deploy-to"
|
deploy_to_key = "deploy-to"
|
||||||
network_key = "network"
|
network_key = "network"
|
||||||
|
@ -24,6 +24,8 @@ from importlib import resources
|
|||||||
import subprocess
|
import subprocess
|
||||||
import click
|
import click
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
from stack_orchestrator import constants
|
||||||
|
from stack_orchestrator.opts import opts
|
||||||
from stack_orchestrator.util import include_exclude_check, get_parsed_stack_config, global_options2, get_dev_root_path
|
from stack_orchestrator.util import include_exclude_check, get_parsed_stack_config, global_options2, get_dev_root_path
|
||||||
from stack_orchestrator.deploy.deployer import Deployer, DeployerException
|
from stack_orchestrator.deploy.deployer import Deployer, DeployerException
|
||||||
from stack_orchestrator.deploy.deployer_factory import getDeployer
|
from stack_orchestrator.deploy.deployer_factory import getDeployer
|
||||||
@ -70,6 +72,9 @@ def create_deploy_context(
|
|||||||
cluster,
|
cluster,
|
||||||
env_file,
|
env_file,
|
||||||
deploy_to) -> DeployCommandContext:
|
deploy_to) -> DeployCommandContext:
|
||||||
|
# Extract the cluster name from the deployment, if we have one
|
||||||
|
if deployment_context and cluster is None:
|
||||||
|
cluster = deployment_context.get_cluster_id()
|
||||||
cluster_context = _make_cluster_context(global_context, stack, include, exclude, cluster, env_file)
|
cluster_context = _make_cluster_context(global_context, stack, include, exclude, cluster, env_file)
|
||||||
deployer = getDeployer(deploy_to, deployment_context, compose_files=cluster_context.compose_files,
|
deployer = getDeployer(deploy_to, deployment_context, compose_files=cluster_context.compose_files,
|
||||||
compose_project_name=cluster_context.cluster,
|
compose_project_name=cluster_context.cluster,
|
||||||
@ -253,6 +258,22 @@ def _make_runtime_env(ctx):
|
|||||||
return container_exec_env
|
return container_exec_env
|
||||||
|
|
||||||
|
|
||||||
|
def _make_default_cluster_name(deployment, compose_dir, stack, include, exclude):
|
||||||
|
# Create default unique, stable cluster name from confile file path and stack name if provided
|
||||||
|
if deployment:
|
||||||
|
path = os.path.realpath(os.path.abspath(compose_dir))
|
||||||
|
else:
|
||||||
|
path = "internal"
|
||||||
|
unique_cluster_descriptor = f"{path},{stack},{include},{exclude}"
|
||||||
|
if opts.o.debug:
|
||||||
|
print(f"pre-hash descriptor: {unique_cluster_descriptor}")
|
||||||
|
hash = hashlib.md5(unique_cluster_descriptor.encode()).hexdigest()[:16]
|
||||||
|
cluster = f"{constants.cluster_name_prefix}{hash}"
|
||||||
|
if opts.o.debug:
|
||||||
|
print(f"Using cluster name: {cluster}")
|
||||||
|
return cluster
|
||||||
|
|
||||||
|
|
||||||
# stack has to be either PathLike pointing to a stack yml file, or a string with the name of a known stack
|
# stack has to be either PathLike pointing to a stack yml file, or a string with the name of a known stack
|
||||||
def _make_cluster_context(ctx, stack, include, exclude, cluster, env_file):
|
def _make_cluster_context(ctx, stack, include, exclude, cluster, env_file):
|
||||||
|
|
||||||
@ -270,18 +291,9 @@ def _make_cluster_context(ctx, stack, include, exclude, cluster, env_file):
|
|||||||
compose_dir = Path(__file__).absolute().parent.parent.joinpath("data", "compose")
|
compose_dir = Path(__file__).absolute().parent.parent.joinpath("data", "compose")
|
||||||
|
|
||||||
if cluster is None:
|
if cluster is None:
|
||||||
# Create default unique, stable cluster name from confile file path and stack name if provided
|
cluster = _make_default_cluster_name(deployment, compose_dir, stack, include, exclude)
|
||||||
if deployment:
|
|
||||||
path = os.path.realpath(os.path.abspath(compose_dir))
|
|
||||||
else:
|
else:
|
||||||
path = "internal"
|
_make_default_cluster_name(deployment, compose_dir, stack, include, exclude)
|
||||||
unique_cluster_descriptor = f"{path},{stack},{include},{exclude}"
|
|
||||||
if ctx.debug:
|
|
||||||
print(f"pre-hash descriptor: {unique_cluster_descriptor}")
|
|
||||||
hash = hashlib.md5(unique_cluster_descriptor.encode()).hexdigest()[:16]
|
|
||||||
cluster = f"laconic-{hash}"
|
|
||||||
if ctx.verbose:
|
|
||||||
print(f"Using cluster name: {cluster}")
|
|
||||||
|
|
||||||
# See: https://stackoverflow.com/a/20885799/1701505
|
# See: https://stackoverflow.com/a/20885799/1701505
|
||||||
from stack_orchestrator import data
|
from stack_orchestrator import data
|
||||||
|
@ -52,7 +52,7 @@ def make_deploy_context(ctx) -> DeployCommandContext:
|
|||||||
context: DeploymentContext = ctx.obj
|
context: DeploymentContext = ctx.obj
|
||||||
stack_file_path = context.get_stack_file()
|
stack_file_path = context.get_stack_file()
|
||||||
env_file = context.get_env_file()
|
env_file = context.get_env_file()
|
||||||
cluster_name = context.get_cluster_name()
|
cluster_name = context.get_cluster_id()
|
||||||
if constants.deploy_to_key in context.spec.obj:
|
if constants.deploy_to_key in context.spec.obj:
|
||||||
deployment_type = context.spec.obj[constants.deploy_to_key]
|
deployment_type = context.spec.obj[constants.deploy_to_key]
|
||||||
else:
|
else:
|
||||||
|
@ -14,15 +14,19 @@
|
|||||||
# You should have received a copy of the GNU Affero General Public License
|
# You should have received a copy of the GNU Affero General Public License
|
||||||
# along with this program. If not, see <http:#www.gnu.org/licenses/>.
|
# along with this program. If not, see <http:#www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
import hashlib
|
||||||
|
import os
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
|
||||||
from stack_orchestrator import constants
|
from stack_orchestrator import constants
|
||||||
|
from stack_orchestrator.util import get_yaml
|
||||||
from stack_orchestrator.deploy.stack import Stack
|
from stack_orchestrator.deploy.stack import Stack
|
||||||
from stack_orchestrator.deploy.spec import Spec
|
from stack_orchestrator.deploy.spec import Spec
|
||||||
|
|
||||||
|
|
||||||
class DeploymentContext:
|
class DeploymentContext:
|
||||||
deployment_dir: Path
|
deployment_dir: Path
|
||||||
|
id: str
|
||||||
spec: Spec
|
spec: Spec
|
||||||
stack: Stack
|
stack: Stack
|
||||||
|
|
||||||
@ -35,9 +39,14 @@ class DeploymentContext:
|
|||||||
def get_env_file(self):
|
def get_env_file(self):
|
||||||
return self.deployment_dir.joinpath(constants.config_file_name)
|
return self.deployment_dir.joinpath(constants.config_file_name)
|
||||||
|
|
||||||
# TODO: implement me
|
def get_deployment_file(self):
|
||||||
def get_cluster_name(self):
|
return self.deployment_dir.joinpath(constants.deployment_file_name)
|
||||||
return None
|
|
||||||
|
def get_compose_dir(self):
|
||||||
|
return self.deployment_dir.joinpath(constants.compose_dir_name)
|
||||||
|
|
||||||
|
def get_cluster_id(self):
|
||||||
|
return self.id
|
||||||
|
|
||||||
def init(self, dir):
|
def init(self, dir):
|
||||||
self.deployment_dir = dir
|
self.deployment_dir = dir
|
||||||
@ -45,3 +54,16 @@ class DeploymentContext:
|
|||||||
self.spec.init_from_file(self.get_spec_file())
|
self.spec.init_from_file(self.get_spec_file())
|
||||||
self.stack = Stack(self.spec.obj["stack"])
|
self.stack = Stack(self.spec.obj["stack"])
|
||||||
self.stack.init_from_file(self.get_stack_file())
|
self.stack.init_from_file(self.get_stack_file())
|
||||||
|
deployment_file_path = self.get_deployment_file()
|
||||||
|
if deployment_file_path.exists():
|
||||||
|
with deployment_file_path:
|
||||||
|
obj = get_yaml().load(open(deployment_file_path, "r"))
|
||||||
|
self.id = obj[constants.cluster_id_key]
|
||||||
|
# Handle the case of a legacy deployment with no file
|
||||||
|
# Code below is intended to match the output from _make_default_cluster_name()
|
||||||
|
# TODO: remove when we no longer need to support legacy deployments
|
||||||
|
else:
|
||||||
|
path = os.path.realpath(os.path.abspath(self.get_compose_dir()))
|
||||||
|
unique_cluster_descriptor = f"{path},{self.get_stack_file()},None,None"
|
||||||
|
hash = hashlib.md5(unique_cluster_descriptor.encode()).hexdigest()[:16]
|
||||||
|
self.id = f"{constants.cluster_name_prefix}{hash}"
|
||||||
|
@ -20,6 +20,7 @@ from pathlib import Path
|
|||||||
from typing import List
|
from typing import List
|
||||||
import random
|
import random
|
||||||
from shutil import copy, copyfile, copytree
|
from shutil import copy, copyfile, copytree
|
||||||
|
from secrets import token_hex
|
||||||
import sys
|
import sys
|
||||||
from stack_orchestrator import constants
|
from stack_orchestrator import constants
|
||||||
from stack_orchestrator.opts import opts
|
from stack_orchestrator.opts import opts
|
||||||
@ -276,7 +277,7 @@ def init(ctx, config, config_file, kube_config, image_registry, output, map_port
|
|||||||
# call it from other commands, bypassing the click decoration stuff
|
# call it from other commands, bypassing the click decoration stuff
|
||||||
def init_operation(deploy_command_context, stack, deployer_type, config,
|
def init_operation(deploy_command_context, stack, deployer_type, config,
|
||||||
config_file, kube_config, image_registry, output, map_ports_to_host):
|
config_file, kube_config, image_registry, output, map_ports_to_host):
|
||||||
yaml = get_yaml()
|
|
||||||
default_spec_file_content = call_stack_deploy_init(deploy_command_context)
|
default_spec_file_content = call_stack_deploy_init(deploy_command_context)
|
||||||
spec_file_content = {"stack": stack, constants.deploy_to_key: deployer_type}
|
spec_file_content = {"stack": stack, constants.deploy_to_key: deployer_type}
|
||||||
if deployer_type == "k8s":
|
if deployer_type == "k8s":
|
||||||
@ -311,8 +312,6 @@ def init_operation(deploy_command_context, stack, deployer_type, config,
|
|||||||
new_config = config_file_variables
|
new_config = config_file_variables
|
||||||
merged_config = {**new_config, **orig_config}
|
merged_config = {**new_config, **orig_config}
|
||||||
spec_file_content.update({"config": merged_config})
|
spec_file_content.update({"config": merged_config})
|
||||||
if opts.o.debug:
|
|
||||||
print(f"Creating spec file for stack: {stack} with content: {spec_file_content}")
|
|
||||||
|
|
||||||
ports = _get_mapped_ports(stack, map_ports_to_host)
|
ports = _get_mapped_ports(stack, map_ports_to_host)
|
||||||
spec_file_content.update({"network": {"ports": ports}})
|
spec_file_content.update({"network": {"ports": ports}})
|
||||||
@ -324,8 +323,11 @@ def init_operation(deploy_command_context, stack, deployer_type, config,
|
|||||||
volume_descriptors[named_volume] = f"./data/{named_volume}"
|
volume_descriptors[named_volume] = f"./data/{named_volume}"
|
||||||
spec_file_content["volumes"] = volume_descriptors
|
spec_file_content["volumes"] = volume_descriptors
|
||||||
|
|
||||||
|
if opts.o.debug:
|
||||||
|
print(f"Creating spec file for stack: {stack} with content: {spec_file_content}")
|
||||||
|
|
||||||
with open(output, "w") as output_file:
|
with open(output, "w") as output_file:
|
||||||
yaml.dump(spec_file_content, output_file)
|
get_yaml().dump(spec_file_content, output_file)
|
||||||
|
|
||||||
|
|
||||||
def _write_config_file(spec_file: Path, config_env_file: Path):
|
def _write_config_file(spec_file: Path, config_env_file: Path):
|
||||||
@ -351,6 +353,13 @@ def _copy_files_to_directory(file_paths: List[Path], directory: Path):
|
|||||||
copy(path, os.path.join(directory, os.path.basename(path)))
|
copy(path, os.path.join(directory, os.path.basename(path)))
|
||||||
|
|
||||||
|
|
||||||
|
def _create_deployment_file(deployment_dir: Path):
|
||||||
|
deployment_file_path = deployment_dir.joinpath(constants.deployment_file_name)
|
||||||
|
cluster = f"{constants.cluster_name_prefix}{token_hex(8)}"
|
||||||
|
with open(deployment_file_path, "w") as output_file:
|
||||||
|
output_file.write(f"{constants.cluster_id_key}: {cluster}\n")
|
||||||
|
|
||||||
|
|
||||||
@click.command()
|
@click.command()
|
||||||
@click.option("--spec-file", required=True, help="Spec file to use to create this deployment")
|
@click.option("--spec-file", required=True, help="Spec file to use to create this deployment")
|
||||||
@click.option("--deployment-dir", help="Create deployment files in this directory")
|
@click.option("--deployment-dir", help="Create deployment files in this directory")
|
||||||
@ -383,6 +392,7 @@ def create_operation(deployment_command_context, spec_file, deployment_dir, netw
|
|||||||
# Copy spec file and the stack file into the deployment dir
|
# Copy spec file and the stack file into the deployment dir
|
||||||
copyfile(spec_file, deployment_dir_path.joinpath(constants.spec_file_name))
|
copyfile(spec_file, deployment_dir_path.joinpath(constants.spec_file_name))
|
||||||
copyfile(stack_file, deployment_dir_path.joinpath(os.path.basename(stack_file)))
|
copyfile(stack_file, deployment_dir_path.joinpath(os.path.basename(stack_file)))
|
||||||
|
_create_deployment_file(deployment_dir_path)
|
||||||
# Copy any config varibles from the spec file into an env file suitable for compose
|
# Copy any config varibles from the spec file into an env file suitable for compose
|
||||||
_write_config_file(spec_file, deployment_dir_path.joinpath(constants.config_file_name))
|
_write_config_file(spec_file, deployment_dir_path.joinpath(constants.config_file_name))
|
||||||
# Copy any k8s config file into the deployment dir
|
# Copy any k8s config file into the deployment dir
|
||||||
|
Loading…
Reference in New Issue
Block a user