# Copyright © 2023 Vulcanize # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see . import datetime import hashlib import json import os import random import subprocess import sys import tempfile import uuid import yaml class AttrDict(dict): def __init__(self, *args, **kwargs): super(AttrDict, self).__init__(*args, **kwargs) self.__dict__ = self def __getattribute__(self, attr): __dict__ = super(AttrDict, self).__getattribute__("__dict__") if attr in __dict__: v = super(AttrDict, self).__getattribute__(attr) if isinstance(v, dict): return AttrDict(v) return v class TimedLogger: def __init__(self, id="", file=None): self.start = datetime.datetime.now() self.last = self.start self.id = id self.file = file def log(self, msg, show_step_time=True, show_total_time=False): prefix = f"{datetime.datetime.utcnow()} - {self.id}" if show_step_time: prefix += f" - {datetime.datetime.now() - self.last} (step)" if show_total_time: prefix += f" - {datetime.datetime.now() - self.start} (total)" print(f"{prefix}: {msg}", file=self.file) if self.file: self.file.flush() self.last = datetime.datetime.now() def logged_cmd(log_file, *vargs): result = None try: if log_file: print(" ".join(vargs), file=log_file) result = subprocess.run(vargs, capture_output=True) result.check_returncode() return result.stdout.decode() except Exception as err: if result: print(result.stderr.decode(), file=log_file) else: print(str(err), file=log_file) raise err def match_owner(recordA, *records): for owner in recordA.owners: for otherRecord in records: if owner in otherRecord.owners: return owner return None def is_lrn(name_or_id: str): if name_or_id: return str(name_or_id).startswith("lrn://") return False def is_id(name_or_id: str): return not is_lrn(name_or_id) def confirm_payment(laconic, record, payment_address, min_amount, logger): req_owner = laconic.get_owner(record) if req_owner == payment_address: # No need to confirm payment if the sender and recipient are the same account. return True if not record.attributes.payment: logger.log(f"{record.id}: no payment tx info") return False tx = laconic.get_tx(record.attributes.payment) if not tx: logger.log(f"{record.id}: cannot locate payment tx") return False if tx.code != 0: logger.log( f"{record.id}: payment tx {tx.hash} was not successful - code: {tx.code}, log: {tx.log}" ) return False if tx.sender != req_owner: logger.log( f"{record.id}: payment sender {tx.sender} in tx {tx.hash} does not match deployment " f"request owner {req_owner}" ) return False if tx.recipient != payment_address: logger.log( f"{record.id}: payment recipient {tx.recipient} in tx {tx.hash} does not match {payment_address}" ) return False pay_denom = "".join([i for i in tx.amount if not i.isdigit()]) if pay_denom != "alnt": logger.log( f"{record.id}: {pay_denom} in tx {tx.hash} is not an expected payment denomination" ) return False pay_amount = int("".join([i for i in tx.amount if i.isdigit()])) if pay_amount < min_amount: logger.log( f"{record.id}: payment amount {tx.amount} is less than minimum {min_amount}" ) return False # Check if the payment was already used on a used = laconic.app_deployments( {"by": payment_address, "payment": tx.hash}, all=True ) if len(used): logger.log(f"{record.id}: payment {tx.hash} already used on deployment {used}") return False used = laconic.app_deployment_removals( {"by": payment_address, "payment": tx.hash}, all=True ) if len(used): logger.log( f"{record.id}: payment {tx.hash} already used on deployment removal {used}" ) return False return True class LaconicRegistryClient: def __init__(self, config_file, log_file=None): self.config_file = config_file self.log_file = log_file self.cache = AttrDict( { "name_or_id": {}, "accounts": {}, "txs": {}, } ) def whoami(self, refresh=False): if not refresh and "whoami" in self.cache: return self.cache["whoami"] args = ["laconic", "-c", self.config_file, "registry", "account", "get"] results = [ AttrDict(r) for r in json.loads(logged_cmd(self.log_file, *args)) if r ] if len(results): self.cache["whoami"] = results[0] return results[0] return None def get_owner(self, record, require=False): bond = self.get_bond(record.bondId, require) if bond: return bond.owner return bond def get_account(self, address, refresh=False, require=False): if not refresh and address in self.cache["accounts"]: return self.cache["accounts"][address] args = [ "laconic", "-c", self.config_file, "registry", "account", "get", "--address", address, ] results = [ AttrDict(r) for r in json.loads(logged_cmd(self.log_file, *args)) if r ] if len(results): self.cache["accounts"][address] = results[0] return results[0] if require: raise Exception("Cannot locate account:", address) return None def get_bond(self, id, require=False): if id in self.cache.name_or_id: return self.cache.name_or_id[id] args = [ "laconic", "-c", self.config_file, "registry", "bond", "get", "--id", id, ] results = [ AttrDict(r) for r in json.loads(logged_cmd(self.log_file, *args)) if r ] self._add_to_cache(results) if len(results): return results[0] if require: raise Exception("Cannot locate bond:", id) return None def list_bonds(self): args = ["laconic", "-c", self.config_file, "registry", "bond", "list"] results = [ AttrDict(r) for r in json.loads(logged_cmd(self.log_file, *args)) if r ] self._add_to_cache(results) return results def list_records(self, criteria=None, all=False): if criteria is None: criteria = {} args = ["laconic", "-c", self.config_file, "registry", "record", "list"] if all: args.append("--all") if criteria: for k, v in criteria.items(): args.append("--%s" % k) args.append(str(v)) results = [ AttrDict(r) for r in json.loads(logged_cmd(self.log_file, *args)) if r ] # Most recent records first results.sort(key=lambda r: r.createTime) results.reverse() self._add_to_cache(results) return results def _add_to_cache(self, records): if not records: return for p in records: self.cache["name_or_id"][p.id] = p if p.names: for lrn in p.names: self.cache["name_or_id"][lrn] = p if p.attributes and p.attributes.type: if p.attributes.type not in self.cache: self.cache[p.attributes.type] = [] self.cache[p.attributes.type].append(p) def resolve(self, name): if not name: return None if name in self.cache.name_or_id: return self.cache.name_or_id[name] args = ["laconic", "-c", self.config_file, "registry", "name", "resolve", name] parsed = [ AttrDict(r) for r in json.loads(logged_cmd(self.log_file, *args)) if r ] if parsed: self._add_to_cache(parsed) return parsed[0] return None def get_record(self, name_or_id, require=False): if not name_or_id: if require: raise Exception("Cannot locate record:", name_or_id) return None if name_or_id in self.cache.name_or_id: return self.cache.name_or_id[name_or_id] if is_lrn(name_or_id): return self.resolve(name_or_id) args = [ "laconic", "-c", self.config_file, "registry", "record", "get", "--id", name_or_id, ] parsed = [ AttrDict(r) for r in json.loads(logged_cmd(self.log_file, *args)) if r ] if len(parsed): self._add_to_cache(parsed) return parsed[0] if require: raise Exception("Cannot locate record:", name_or_id) return None def get_tx(self, txHash, require=False): if txHash in self.cache["txs"]: return self.cache["txs"][txHash] args = [ "laconic", "-c", self.config_file, "registry", "tokens", "gettx", "--hash", txHash, ] parsed = None try: parsed = AttrDict(json.loads(logged_cmd(self.log_file, *args))) except: # noqa: E722 pass if parsed: self.cache["txs"][txHash] = parsed return parsed if require: raise Exception("Cannot locate tx:", hash) def app_deployment_requests(self, criteria=None, all=True): if criteria is None: criteria = {} criteria = criteria.copy() criteria["type"] = "ApplicationDeploymentRequest" return self.list_records(criteria, all) def app_deployments(self, criteria=None, all=True): if criteria is None: criteria = {} criteria = criteria.copy() criteria["type"] = "ApplicationDeploymentRecord" return self.list_records(criteria, all) def app_deployment_removal_requests(self, criteria=None, all=True): if criteria is None: criteria = {} criteria = criteria.copy() criteria["type"] = "ApplicationDeploymentRemovalRequest" return self.list_records(criteria, all) def app_deployment_removals(self, criteria=None, all=True): if criteria is None: criteria = {} criteria = criteria.copy() criteria["type"] = "ApplicationDeploymentRemovalRecord" return self.list_records(criteria, all) def publish(self, record, names=None): if names is None: names = [] tmpdir = tempfile.mkdtemp() try: record_fname = os.path.join(tmpdir, "record.yml") record_file = open(record_fname, "w") yaml.dump(record, record_file) record_file.close() print(open(record_fname, "r").read(), file=self.log_file) new_record_id = json.loads( logged_cmd( self.log_file, "laconic", "-c", self.config_file, "registry", "record", "publish", "--filename", record_fname, ) )["id"] for name in names: self.set_name(name, new_record_id) return new_record_id finally: logged_cmd(self.log_file, "rm", "-rf", tmpdir) def set_name(self, name, record_id): logged_cmd( self.log_file, "laconic", "-c", self.config_file, "registry", "name", "set", name, record_id, ) def delete_name(self, name): logged_cmd( self.log_file, "laconic", "-c", self.config_file, "registry", "name", "delete", name, ) def file_hash(filename): return hashlib.sha1(open(filename).read().encode()).hexdigest() def determine_base_container(clone_dir, app_type="webapp"): if not app_type or not app_type.startswith("webapp"): raise Exception(f"Unsupported app_type {app_type}") base_container = "cerc/webapp-base" if app_type == "webapp/next": base_container = "cerc/nextjs-base" elif app_type == "webapp": pkg_json_path = os.path.join(clone_dir, "package.json") if os.path.exists(pkg_json_path): pkg_json = json.load(open(pkg_json_path)) if "next" in pkg_json.get("dependencies", {}): base_container = "cerc/nextjs-base" return base_container def build_container_image(app_record, tag, extra_build_args=None, logger=None): if extra_build_args is None: extra_build_args = [] tmpdir = tempfile.mkdtemp() # TODO: determine if this code could be calling into the Python git library like setup-repositories try: record_id = app_record["id"] ref = app_record.attributes.repository_ref repo = random.choice(app_record.attributes.repository) clone_dir = os.path.join(tmpdir, record_id) logger.log(f"Cloning repository {repo} to {clone_dir} ...") # Set github credentials if present running a command like: # git config --global url."https://${TOKEN}:@github.com/".insteadOf "https://github.com/" github_token = os.environ.get("DEPLOYER_GITHUB_TOKEN") if github_token: logger.log("Github token detected, setting it in the git environment") git_config_args = [ "git", "config", "--global", f"url.https://{github_token}:@github.com/.insteadOf", "https://github.com/", ] result = subprocess.run( git_config_args, stdout=logger.file, stderr=logger.file ) result.check_returncode() if ref: # TODO: Determing branch or hash, and use depth 1 if we can. git_env = dict(os.environ.copy()) # Never prompt git_env["GIT_TERMINAL_PROMPT"] = "0" try: subprocess.check_call( ["git", "clone", repo, clone_dir], env=git_env, stdout=logger.file, stderr=logger.file, ) except Exception as e: logger.log(f"git clone failed. Is the repository {repo} private?") raise e try: subprocess.check_call( ["git", "checkout", ref], cwd=clone_dir, env=git_env, stdout=logger.file, stderr=logger.file, ) except Exception as e: logger.log(f"git checkout failed. Does ref {ref} exist?") raise e else: # TODO: why is this code different vs the branch above (run vs check_call, and no prompt disable)? result = subprocess.run( ["git", "clone", "--depth", "1", repo, clone_dir], stdout=logger.file, stderr=logger.file, ) result.check_returncode() base_container = determine_base_container( clone_dir, app_record.attributes.app_type ) logger.log("Building webapp ...") build_command = [ sys.argv[0], "--verbose", "build-webapp", "--source-repo", clone_dir, "--tag", tag, "--base-container", base_container, ] if extra_build_args: build_command.append("--extra-build-args") build_command.append(" ".join(extra_build_args)) result = subprocess.run(build_command, stdout=logger.file, stderr=logger.file) result.check_returncode() finally: logged_cmd(logger.file, "rm", "-rf", tmpdir) def push_container_image(deployment_dir, logger): logger.log("Pushing images ...") result = subprocess.run( [sys.argv[0], "deployment", "--dir", deployment_dir, "push-images"], stdout=logger.file, stderr=logger.file, ) result.check_returncode() logger.log("Finished pushing images.") def deploy_to_k8s(deploy_record, deployment_dir, recreate, logger): logger.log("Deploying to k8s ...") if recreate: commands_to_run = ["stop", "start"] else: if not deploy_record: commands_to_run = ["start"] else: commands_to_run = ["update"] for command in commands_to_run: logger.log(f"Running {command} command on deployment dir: {deployment_dir}") result = subprocess.run( [sys.argv[0], "deployment", "--dir", deployment_dir, command], stdout=logger.file, stderr=logger.file, ) result.check_returncode() logger.log(f"Finished {command} command on deployment dir: {deployment_dir}") logger.log("Finished deploying to k8s.") def publish_deployment( laconic: LaconicRegistryClient, app_record, deploy_record, deployment_lrn, dns_record, dns_lrn, deployment_dir, app_deployment_request=None, payment_address=None, logger=None, ): if not deploy_record: deploy_ver = "0.0.1" else: deploy_ver = "0.0.%d" % ( int(deploy_record.attributes.version.split(".")[-1]) + 1 ) if not dns_record: dns_ver = "0.0.1" else: dns_ver = "0.0.%d" % (int(dns_record.attributes.version.split(".")[-1]) + 1) spec = yaml.full_load(open(os.path.join(deployment_dir, "spec.yml"))) fqdn = spec["network"]["http-proxy"][0]["host-name"] uniq = uuid.uuid4() new_dns_record = { "record": { "type": "DnsRecord", "version": dns_ver, "name": fqdn, "resource_type": "A", "meta": {"so": uniq.hex}, } } if app_deployment_request: new_dns_record["record"]["request"] = app_deployment_request.id if logger: logger.log("Publishing DnsRecord.") dns_id = laconic.publish(new_dns_record, [dns_lrn]) new_deployment_record = { "record": { "type": "ApplicationDeploymentRecord", "version": deploy_ver, "url": f"https://{fqdn}", "name": app_record.attributes.name, "application": app_record.id, "dns": dns_id, "meta": { "config": file_hash(os.path.join(deployment_dir, "config.env")), "so": uniq.hex, }, } } if app_deployment_request: new_deployment_record["record"]["request"] = app_deployment_request.id if app_deployment_request.attributes.payment: new_deployment_record["record"][ "payment" ] = app_deployment_request.attributes.payment if payment_address: new_deployment_record["record"]["by"] = payment_address if logger: logger.log("Publishing ApplicationDeploymentRecord.") deployment_id = laconic.publish(new_deployment_record, [deployment_lrn]) return {"dns": dns_id, "deployment": deployment_id} def hostname_for_deployment_request(app_deployment_request, laconic): dns_name = app_deployment_request.attributes.dns if not dns_name: app = laconic.get_record( app_deployment_request.attributes.application, require=True ) dns_name = generate_hostname_for_app(app) elif dns_name.startswith("lrn://"): record = laconic.get_record(dns_name, require=True) dns_name = record.attributes.name return dns_name def generate_hostname_for_app(app): last_part = app.attributes.name.split("/")[-1] m = hashlib.sha256() m.update(app.attributes.name.encode()) m.update(b"|") if isinstance(app.attributes.repository, list): m.update(app.attributes.repository[0].encode()) else: m.update(app.attributes.repository.encode()) return "%s-%s" % (last_part, m.hexdigest()[0:10]) def skip_by_tag(r, include_tags, exclude_tags): for tag in exclude_tags: if r.attributes.tags and tag in r.attributes.tags: return True if include_tags: for tag in include_tags: if r.attributes.tags and tag in r.attributes.tags: return False return True return False