exit on failure, add per-class logging
All checks were successful
Check Nix flake / Build nix outputs (ubuntu-latest) (pull_request) Successful in 3m30s
Check flake.lock / Check health of `flake.lock` (pull_request) Successful in 14s
Check Nix flake / Perform Nix flake checks (ubuntu-latest) (pull_request) Successful in 4m30s
Check Nix formatting / Perform Nix format checks (pull_request) Successful in 2m56s

Signed-off-by: ahuston-0 <aliceghuston@gmail.com>
This commit is contained in:
ahuston-0 2025-03-26 11:22:29 -04:00
parent 1c3314a6df
commit 1f8d589c1b
No known key found for this signature in database
GPG Key ID: 47940175096C1330
5 changed files with 42 additions and 31 deletions

View File

@ -11,6 +11,7 @@ from flupdt.common import bash_wrapper
drv_re = re.compile(r".*(/nix/store/.*\.drv).*") drv_re = re.compile(r".*(/nix/store/.*\.drv).*")
OUTPUT_DIR = mkdtemp(prefix="flupdt-outputs-") OUTPUT_DIR = mkdtemp(prefix="flupdt-outputs-")
logger = logging.getLogger(__name__)
def build_output(path: str, output: str) -> str | None: def build_output(path: str, output: str) -> str | None:
@ -20,16 +21,16 @@ def build_output(path: str, output: str) -> str | None:
:param output: flake output to be built :param output: flake output to be built
:returns the .drv path on success or None on failure :returns the .drv path on success or None on failure
""" """
logging.info(f"build {output}") logger.info(f"build {output}")
logging.debug(f"outputting to {OUTPUT_DIR}/{output}.nixoutput") logger.debug(f"outputting to {OUTPUT_DIR}/{output}.nixoutput")
out = bash_wrapper(f"nix build {path}#{output} -o {OUTPUT_DIR}/{output}.nixoutput") out = bash_wrapper(f"nix build {path}#{output} -o {OUTPUT_DIR}/{output}.nixoutput")
logging.debug("output") logger.debug("output")
logging.debug(out[0]) logger.debug(out[0])
logging.debug("error") logger.debug("error")
logging.debug(out[1]) logger.debug(out[1])
logging.debug("statuscode") logger.debug("statuscode")
logging.debug(out[2]) logger.debug(out[2])
if out[2] != 0: if out[2] != 0:
logging.warning(f"output {output} did not build correctly") logger.warning(f"output {output} did not build correctly")
return None return None
return "" return ""

View File

@ -5,6 +5,8 @@ import shutil
from flupdt.common import bash_wrapper from flupdt.common import bash_wrapper
logger = logging.getLogger(__name__)
def compare_derivations( def compare_derivations(
path_to_flake: str, path_to_pre_drv: str, path_to_post_drv: str, *, soft_failure: bool = True path_to_flake: str, path_to_pre_drv: str, path_to_post_drv: str, *, soft_failure: bool = True
@ -24,10 +26,10 @@ def compare_derivations(
) )
if returncode != 0: if returncode != 0:
log_func = logging.error log_func = logger.error
if soft_failure: if soft_failure:
log_func = logging.warning log_func = logger.warning
logging.warning(f"diff failed for {path_to_pre_drv} and {path_to_post_drv}") logger.warning(f"diff failed for {path_to_pre_drv} and {path_to_post_drv}")
log_func(stderr) log_func(stderr)
return stdout return stdout

View File

@ -9,6 +9,8 @@ from flupdt.common import bash_wrapper
drv_re = re.compile(r".*(/nix/store/.*\.drv).*") drv_re = re.compile(r".*(/nix/store/.*\.drv).*")
logger = logging.getLogger(__name__)
def evaluate_output(path: str, output: str) -> str | None: def evaluate_output(path: str, output: str) -> str | None:
"""Evaluates a given output in a flake. """Evaluates a given output in a flake.
@ -18,19 +20,19 @@ def evaluate_output(path: str, output: str) -> str | None:
:returns the .drv path on success or None on failure :returns the .drv path on success or None on failure
:raises RuntimeError: evaluation succeeds but no derivation is found :raises RuntimeError: evaluation succeeds but no derivation is found
""" """
logging.info(f"evaluating {output}") logger.info(f"evaluating {output}")
out = bash_wrapper(f"nix eval {path}#{output}") out = bash_wrapper(f"nix eval {path}#{output}")
logging.debug(out[0]) logger.debug(out[0])
logging.debug(out[1]) logger.debug(out[1])
logging.debug(out[2]) logger.debug(out[2])
if out[2] != 0: if out[2] != 0:
logging.warning(f"output {output} did not evaluate correctly") logger.warning(f"output {output} did not evaluate correctly")
return None return None
drv_match = drv_re.match(out[0]) drv_match = drv_re.match(out[0])
if drv_match is None: if drv_match is None:
out_msg = "derivation succeeded but output derivation does not contain a derivation" out_msg = "derivation succeeded but output derivation does not contain a derivation"
raise RuntimeError(out_msg) raise RuntimeError(out_msg)
drv = drv_match.group(1) drv = drv_match.group(1)
logging.debug(f"derivation evaluated to {drv}") logger.debug(f"derivation evaluated to {drv}")
return drv return drv

View File

@ -12,6 +12,7 @@ output_regexes = [
re.compile(r"checking derivation (.*)..."), re.compile(r"checking derivation (.*)..."),
re.compile(r"checking NixOS configuration \'(nixosConfigurations.*)\'\.\.\."), re.compile(r"checking NixOS configuration \'(nixosConfigurations.*)\'\.\.\."),
] ]
logger = logging.getLogger(__name__)
def traverse_json_base(json_dict: dict[str, typing.Any], path: list[str]) -> list[str]: def traverse_json_base(json_dict: dict[str, typing.Any], path: list[str]) -> list[str]:
@ -54,7 +55,7 @@ def get_derivations_from_check(nix_path: str, path_to_flake: str) -> list[str]:
""" """
flake_check = bash_wrapper(f"{nix_path} flake check --verbose --keep-going", path=path_to_flake) flake_check = bash_wrapper(f"{nix_path} flake check --verbose --keep-going", path=path_to_flake)
if flake_check[2] != 0: if flake_check[2] != 0:
logging.warning( logger.warning(
"nix flake check returned non-zero exit code, collecting all available outputs" "nix flake check returned non-zero exit code, collecting all available outputs"
) )
error_out = flake_check[1].split("\n") error_out = flake_check[1].split("\n")
@ -62,10 +63,10 @@ def get_derivations_from_check(nix_path: str, path_to_flake: str) -> list[str]:
derivations = [] derivations = []
for output in possible_outputs: for output in possible_outputs:
for r in output_regexes: for r in output_regexes:
logging.debug(f"{output} {r.pattern}") logger.debug(f"{output} {r.pattern}")
match = r.match(output) match = r.match(output)
if match is not None: if match is not None:
logging.debug(match.groups()) logger.debug(match.groups())
derivations += [match.groups()[0]] derivations += [match.groups()[0]]
return derivations return derivations
@ -84,8 +85,8 @@ def get_derivations(path_to_flake: str) -> list[str]:
raise RuntimeError(status_msg) raise RuntimeError(status_msg)
flake_show = bash_wrapper(f"{nix_path} flake show --json", path=path_to_flake) flake_show = bash_wrapper(f"{nix_path} flake show --json", path=path_to_flake)
if flake_show[2] != 0: if flake_show[2] != 0:
logging.error("flake show returned non-zero exit code") logger.error("flake show returned non-zero exit code")
logging.warning("falling back to full evaluation via nix flake check") logger.warning("falling back to full evaluation via nix flake check")
derivations = get_derivations_from_check(nix_path, path_to_flake) derivations = get_derivations_from_check(nix_path, path_to_flake)
else: else:
flake_show_json = json.loads(flake_show[0]) flake_show_json = json.loads(flake_show[0])

View File

@ -4,6 +4,7 @@
import logging import logging
import os import os
import sys
from argparse import Namespace from argparse import Namespace
from pathlib import Path from pathlib import Path
@ -14,6 +15,8 @@ from flupdt.flake_diff import compare_derivations
from flupdt.flake_eval import evaluate_output from flupdt.flake_eval import evaluate_output
from flupdt.flake_show import get_derivations from flupdt.flake_show import get_derivations
logger = logging.getLogger(__name__)
def batch_eval(args: Namespace, flake_path: str, derivations: list[str]) -> None: def batch_eval(args: Namespace, flake_path: str, derivations: list[str]) -> None:
"""Bulk run evaluations or builds on a derivation set. """Bulk run evaluations or builds on a derivation set.
@ -28,12 +31,14 @@ def batch_eval(args: Namespace, flake_path: str, derivations: list[str]) -> None
if args.evaluate: if args.evaluate:
drv_map[d] = evaluate_output(flake_path, d) drv_map[d] = evaluate_output(flake_path, d)
if args.build: if args.build:
build_output(flake_path, d) drv_map[d] = build_output(flake_path, d)
if args.json: if args.json:
with Path.open(args.json, "w+") as f: with Path.open(args.json, "w+") as f:
from json import dump from json import dump
dump(drv_map, f) dump(drv_map, f)
if any(x is None for x in drv_map.values()):
sys.exit(1)
def compare_drvs(args: Namespace) -> None: def compare_drvs(args: Namespace) -> None:
@ -52,8 +57,8 @@ def compare_drvs(args: Namespace) -> None:
pre_json_dict = load(pre) pre_json_dict = load(pre)
post_json_dict = load(post) post_json_dict = load(post)
logging.debug(f"pre-snapshot derivations: {pre_json_dict}") logger.debug(f"pre-snapshot derivations: {pre_json_dict}")
logging.debug(f"post-snapshot derivations: {post_json_dict}") logger.debug(f"post-snapshot derivations: {post_json_dict}")
pre_json_keys = set(pre_json_dict.keys()) pre_json_keys = set(pre_json_dict.keys())
post_json_keys = set(post_json_dict.keys()) post_json_keys = set(post_json_dict.keys())
@ -64,11 +69,11 @@ def compare_drvs(args: Namespace) -> None:
missing_pre_keys = post_json_keys.difference(common_keys_to_eval) missing_pre_keys = post_json_keys.difference(common_keys_to_eval)
if missing_pre_keys: if missing_pre_keys:
logging.warning(f"Following outputs are missing from pre-snapshot: {missing_pre_keys}") logger.warning(f"Following outputs are missing from pre-snapshot: {missing_pre_keys}")
if missing_post_keys: if missing_post_keys:
logging.warning(f"Following outputs are missing from post-snapshot: {missing_post_keys}") logger.warning(f"Following outputs are missing from post-snapshot: {missing_post_keys}")
logging.info(f"Evaluating the following outputs for differences: {common_keys_to_eval}") logger.info(f"Evaluating the following outputs for differences: {common_keys_to_eval}")
out_file: str = os.devnull out_file: str = os.devnull
if args.compare_output_to_file: if args.compare_output_to_file:
@ -97,11 +102,11 @@ def build_or_eval(args: Namespace) -> None:
lambda s: s.startswith("hydraJobs"), get_derivations(flake_path) lambda s: s.startswith("hydraJobs"), get_derivations(flake_path)
) )
derivations, hydra_jobs = list(derivations), list(hydra_jobs) derivations, hydra_jobs = list(derivations), list(hydra_jobs)
logging.info(f"derivations: {list(derivations)}") logger.info(f"derivations: {list(derivations)}")
batch_eval(args, flake_path, derivations) batch_eval(args, flake_path, derivations)
if not args.keep_hydra: if not args.keep_hydra:
logging.info("--keep-hydra flag is not specified, removing Hydra jobs") logger.info("--keep-hydra flag is not specified, removing Hydra jobs")
else: else:
batch_eval(args, flake_path, hydra_jobs) batch_eval(args, flake_path, hydra_jobs)