# Logic for generating Buildkite pipelines from Nix build targets read # by //nix/readTree. # # It outputs a "YAML" (actually JSON) file which is evaluated and # submitted to Buildkite at the start of each build. # # The structure of the file that is being created is documented here: # https://buildkite.com/docs/pipelines/defining-steps { pkgs, ... }: let inherit (builtins) attrValues concatMap concatStringsSep filter foldl' getEnv hasAttr length listToAttrs mapAttrs pathExists toJSON unsafeDiscardStringContext; inherit (pkgs) lib runCommandNoCC writeText; in rec { # Creates a Nix expression that yields the target at the specified # location in the repository. # # This makes a distinction between normal targets (which physically # exist in the repository) and subtargets (which are "virtual" # targets exposed by a physical one) to make it clear in the build # output which is which. mkBuildExpr = target: let descend = expr: attr: "builtins.getAttr \"${attr}\" (${expr})"; targetExpr = foldl' descend "import ./. {}" target.__readTree; subtargetExpr = descend targetExpr target.__subtarget; in if target ? __subtarget then subtargetExpr else targetExpr; # Create a pipeline label from the target's tree location. mkLabel = target: let label = concatStringsSep "/" target.__readTree; in if target ? __subtarget then "${label}:${target.__subtarget}" else label; # Determine whether to skip a target if it has not diverged from the # HEAD branch. shouldSkip = parentTargetMap: label: drvPath: if (hasAttr label parentTargetMap) && parentTargetMap."${label}".drvPath == drvPath then "Target has not changed." else false; # Create a pipeline step from a single target. mkStep = headBranch: parentTargetMap: target: let label = mkLabel target; drvPath = unsafeDiscardStringContext target.drvPath; shouldSkip' = shouldSkip parentTargetMap; in { label = ":nix: " + label; skip = shouldSkip' label drvPath; command = concatStringsSep " " [ # First try to realise the drvPath of the target so we don't evaluate twice. # Nix has no concept of depending on a derivation file without depending on # at least one of its `outPath`s, so we need to discard the string context # if we don't want to build everything during pipeline construction. "nix-store --realise '${drvPath}'" # Since we don't gcroot the derivation files, they may be deleted by the # garbage collector. In that case we can reevaluate and build the attribute # using nix-build. "|| (test ! -f '${drvPath}' && nix-build -E '${mkBuildExpr target}' --show-trace)" ]; # Add a dependency on the initial static pipeline step which # always runs. This allows build steps uploaded in batches to # start running before all batches have been uploaded. depends_on = ":init:"; }; # Helper function to inelegantly divide a list into chunks of at # most n elements. # # This works by assigning each element a chunk ID based on its # index, and then grouping all elements by their chunk ID. chunksOf = n: list: let chunkId = idx: toString (idx / n + 1); assigned = lib.imap1 (idx: value: { inherit value ; chunk = chunkId idx; }) list; unchunk = mapAttrs (_: elements: map (e: e.value) elements); in unchunk (lib.groupBy (e: e.chunk) assigned); # Define a build pipeline chunk as a JSON file, using the pipeline # format documented on # https://buildkite.com/docs/pipelines/defining-steps. makePipelineChunk = chunkId: chunk: rec { filename = "chunk-${chunkId}.json"; path = writeText filename (toJSON { steps = chunk; }); }; # Split the pipeline into chunks of at most 256 steps at once, which # are uploaded sequentially. This is because of a limitation in the # Buildkite backend which struggles to process more than a specific # number of chunks at once. pipelineChunks = steps: attrValues (mapAttrs makePipelineChunk (chunksOf 256 steps)); # Create a pipeline structure for the given targets. mkPipeline = { # HEAD branch of the repository on which release steps, GC # anchoring and other "mainline only" steps should run. headBranch, # List of derivations as read by readTree (in most cases just the # output of readTree.gather) that should be built in Buildkite. # # These are scheduled as the first build steps and run as fast as # possible, in order, without any concurrency restrictions. drvTargets, # Derivation map of a parent commit. Only targets which no longer # correspond to the content of this map will be built. Passing an # empty map will always build all targets. parentTargetMap ? {}, # A list of plain Buildkite step structures to run alongside the # build for all drvTargets, but before proceeding with any # post-build actions such as status reporting. # # Can be used for things like code formatting checks. additionalSteps ? [], # A list of plain Buildkite step structures to run after all # previous steps succeeded. # # Can be used for status reporting steps and the like. postBuildSteps ? [] }: let mkStep' = mkStep headBranch parentTargetMap; steps = # Add build steps for each derivation target. (map mkStep' drvTargets) # Add additional steps (if set). ++ additionalSteps # Wait for all previous checks to complete ++ [({ wait = null; continue_on_failure = true; })] # Run post-build steps for status reporting and co. ++ postBuildSteps; chunks = pipelineChunks steps; in runCommandNoCC "buildkite-pipeline" {} '' mkdir $out echo "Generated ${toString (length chunks)} pipeline chunks" ${ lib.concatMapStringsSep "\n" (chunk: "cp ${chunk.path} $out/${chunk.filename}") chunks } ''; # Create a drvmap structure for the given targets, containing the # mapping of all target paths to their derivations. The mapping can # be persisted for future use. mkDrvmap = drvTargets: writeText "drvmap.json" (toJSON (listToAttrs (map (target: { name = mkLabel target; value = { drvPath = unsafeDiscardStringContext target.drvPath; # Include the attrPath in the output to reconstruct the drv # without parsing the human-readable label. attrPath = target.__readTree ++ lib.optionals (target ? __subtarget) [ target.__subtarget ]; }; }) drvTargets))); }