Compare commits
8 Commits
db6fcabbf3
...
merge/resy
Author | SHA1 | Date | |
---|---|---|---|
|
2b739a2fab | ||
|
f0a72a83bb | ||
|
ad7dbf6826 | ||
|
d294b60477 | ||
|
947a769012 | ||
|
b1b3440041 | ||
|
b832cab12c | ||
|
f6fa2e16c0 |
21
flake.lock
generated
21
flake.lock
generated
@@ -3,16 +3,16 @@
|
|||||||
"nix": {
|
"nix": {
|
||||||
"flake": false,
|
"flake": false,
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1750777360,
|
"lastModified": 1759956402,
|
||||||
"narHash": "sha256-nDWFxwhT+fQNgi4rrr55EKjpxDyVKSl1KaNmSXtYj40=",
|
"narHash": "sha256-CM27YK+KMi3HLRXqjPaJwkTabmKW+CDXOE3kMMtXH3s=",
|
||||||
"owner": "NixOS",
|
"owner": "NixOS",
|
||||||
"repo": "nix",
|
"repo": "nix",
|
||||||
"rev": "7bb200199705eddd53cb34660a76567c6f1295d9",
|
"rev": "3019db2c87006817b6201113ad4ceee0c53c3b62",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
"owner": "NixOS",
|
"owner": "NixOS",
|
||||||
"ref": "2.29-maintenance",
|
"ref": "2.31-maintenance",
|
||||||
"repo": "nix",
|
"repo": "nix",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
}
|
}
|
||||||
@@ -20,26 +20,27 @@
|
|||||||
"nix-eval-jobs": {
|
"nix-eval-jobs": {
|
||||||
"flake": false,
|
"flake": false,
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1748680938,
|
"lastModified": 1757626891,
|
||||||
"narHash": "sha256-TQk6pEMD0mFw7jZXpg7+2qNKGbAluMQgc55OMgEO8bM=",
|
"narHash": "sha256-VrHPtHxVIboqgnw+tlCQepgtBOhBvU5hxbMHsPo8LAc=",
|
||||||
"owner": "nix-community",
|
"owner": "nix-community",
|
||||||
"repo": "nix-eval-jobs",
|
"repo": "nix-eval-jobs",
|
||||||
"rev": "974a4af3d4a8fd242d8d0e2608da4be87a62b83f",
|
"rev": "c975efc5b2bec0c1ff93c67de4a03306af258ff7",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
"owner": "nix-community",
|
"owner": "nix-community",
|
||||||
|
"ref": "v2.31.0",
|
||||||
"repo": "nix-eval-jobs",
|
"repo": "nix-eval-jobs",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"nixpkgs": {
|
"nixpkgs": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1750736827,
|
"lastModified": 1759652726,
|
||||||
"narHash": "sha256-UcNP7BR41xMTe0sfHBH8R79+HdCw0OwkC/ZKrQEuMeo=",
|
"narHash": "sha256-2VjnimOYDRb3DZHyQ2WH2KCouFqYm9h0Rr007Al/WSA=",
|
||||||
"owner": "NixOS",
|
"owner": "NixOS",
|
||||||
"repo": "nixpkgs",
|
"repo": "nixpkgs",
|
||||||
"rev": "b4a30b08433ad7b6e1dfba0833fb0fe69d43dfec",
|
"rev": "06b2985f0cc9eb4318bf607168f4b15af1e5e81d",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
|
@@ -4,13 +4,13 @@
|
|||||||
inputs.nixpkgs.url = "github:NixOS/nixpkgs/nixos-25.05-small";
|
inputs.nixpkgs.url = "github:NixOS/nixpkgs/nixos-25.05-small";
|
||||||
|
|
||||||
inputs.nix = {
|
inputs.nix = {
|
||||||
url = "github:NixOS/nix/2.29-maintenance";
|
url = "github:NixOS/nix/2.31-maintenance";
|
||||||
# We want to control the deps precisely
|
# We want to control the deps precisely
|
||||||
flake = false;
|
flake = false;
|
||||||
};
|
};
|
||||||
|
|
||||||
inputs.nix-eval-jobs = {
|
inputs.nix-eval-jobs = {
|
||||||
url = "github:nix-community/nix-eval-jobs";
|
url = "github:nix-community/nix-eval-jobs/v2.31.0";
|
||||||
# We want to control the deps precisely
|
# We want to control the deps precisely
|
||||||
flake = false;
|
flake = false;
|
||||||
};
|
};
|
||||||
|
@@ -4,7 +4,7 @@ project('hydra', 'cpp',
|
|||||||
default_options: [
|
default_options: [
|
||||||
'debug=true',
|
'debug=true',
|
||||||
'optimization=2',
|
'optimization=2',
|
||||||
'cpp_std=c++20',
|
'cpp_std=c++23',
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@@ -364,7 +364,7 @@ in
|
|||||||
requires = [ "hydra-init.service" ];
|
requires = [ "hydra-init.service" ];
|
||||||
restartTriggers = [ hydraConf ];
|
restartTriggers = [ hydraConf ];
|
||||||
after = [ "hydra-init.service" "network.target" ];
|
after = [ "hydra-init.service" "network.target" ];
|
||||||
path = with pkgs; [ hostname-debian cfg.package jq ];
|
path = with pkgs; [ hostname-debian cfg.package ];
|
||||||
environment = env // {
|
environment = env // {
|
||||||
HYDRA_DBI = "${env.HYDRA_DBI};application_name=hydra-evaluator";
|
HYDRA_DBI = "${env.HYDRA_DBI};application_name=hydra-evaluator";
|
||||||
};
|
};
|
||||||
|
@@ -488,10 +488,11 @@ Step::ptr State::createStep(ref<Store> destStore,
|
|||||||
runnable while step->created == false. */
|
runnable while step->created == false. */
|
||||||
step->drv = std::make_unique<Derivation>(localStore->readDerivation(drvPath));
|
step->drv = std::make_unique<Derivation>(localStore->readDerivation(drvPath));
|
||||||
{
|
{
|
||||||
auto parsedOpt = StructuredAttrs::tryParse(step->drv->env);
|
|
||||||
try {
|
try {
|
||||||
step->drvOptions = std::make_unique<DerivationOptions>(
|
step->drvOptions = std::make_unique<DerivationOptions>(
|
||||||
DerivationOptions::fromStructuredAttrs(step->drv->env, parsedOpt ? &*parsedOpt : nullptr));
|
DerivationOptions::fromStructuredAttrs(
|
||||||
|
step->drv->env,
|
||||||
|
step->drv->structuredAttrs ? &*step->drv->structuredAttrs : nullptr));
|
||||||
} catch (Error & e) {
|
} catch (Error & e) {
|
||||||
e.addTrace({}, "while parsing derivation '%s'", localStore->printStorePath(drvPath));
|
e.addTrace({}, "while parsing derivation '%s'", localStore->printStorePath(drvPath));
|
||||||
throw;
|
throw;
|
||||||
|
@@ -27,6 +27,7 @@
|
|||||||
#include <nix/store/serve-protocol-impl.hh>
|
#include <nix/store/serve-protocol-impl.hh>
|
||||||
#include <nix/store/serve-protocol-connection.hh>
|
#include <nix/store/serve-protocol-connection.hh>
|
||||||
#include <nix/store/machines.hh>
|
#include <nix/store/machines.hh>
|
||||||
|
#include <nix/store/globals.hh>
|
||||||
|
|
||||||
|
|
||||||
typedef unsigned int BuildID;
|
typedef unsigned int BuildID;
|
||||||
|
@@ -212,7 +212,7 @@ sub checkPath {
|
|||||||
sub serveFile {
|
sub serveFile {
|
||||||
my ($c, $path) = @_;
|
my ($c, $path) = @_;
|
||||||
|
|
||||||
my $res = run(cmd => ["nix", "--experimental-features", "nix-command",
|
my $res = runCommand(cmd => ["nix", "--experimental-features", "nix-command",
|
||||||
"ls-store", "--store", getStoreUri(), "--json", "$path"]);
|
"ls-store", "--store", getStoreUri(), "--json", "$path"]);
|
||||||
|
|
||||||
if ($res->{status}) {
|
if ($res->{status}) {
|
||||||
|
@@ -44,7 +44,7 @@ our @EXPORT = qw(
|
|||||||
readNixFile
|
readNixFile
|
||||||
registerRoot
|
registerRoot
|
||||||
restartBuilds
|
restartBuilds
|
||||||
run
|
runCommand
|
||||||
$MACHINE_LOCAL_STORE
|
$MACHINE_LOCAL_STORE
|
||||||
);
|
);
|
||||||
|
|
||||||
@@ -466,7 +466,7 @@ sub readIntoSocket{
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
sub run {
|
sub runCommand {
|
||||||
my (%args) = @_;
|
my (%args) = @_;
|
||||||
my $res = { stdout => "", stderr => "" };
|
my $res = { stdout => "", stderr => "" };
|
||||||
my $stdin = "";
|
my $stdin = "";
|
||||||
@@ -506,7 +506,7 @@ sub run {
|
|||||||
|
|
||||||
sub grab {
|
sub grab {
|
||||||
my (%args) = @_;
|
my (%args) = @_;
|
||||||
my $res = run(%args, grabStderr => 0);
|
my $res = runCommand(%args, grabStderr => 0);
|
||||||
if ($res->{status}) {
|
if ($res->{status}) {
|
||||||
my $msgloc = "(in an indeterminate location)";
|
my $msgloc = "(in an indeterminate location)";
|
||||||
if (defined $args{dir}) {
|
if (defined $args{dir}) {
|
||||||
|
@@ -10,7 +10,6 @@ use Hydra::Helper::CatalystUtils;
|
|||||||
use Hydra::Helper::Nix;
|
use Hydra::Helper::Nix;
|
||||||
use File::Temp;
|
use File::Temp;
|
||||||
use POSIX qw(strftime);
|
use POSIX qw(strftime);
|
||||||
use IPC::Run qw(run);
|
|
||||||
|
|
||||||
sub supportedInputTypes {
|
sub supportedInputTypes {
|
||||||
my ($self, $inputTypes) = @_;
|
my ($self, $inputTypes) = @_;
|
||||||
@@ -45,12 +44,11 @@ sub fetchInput {
|
|||||||
my $ua = LWP::UserAgent->new();
|
my $ua = LWP::UserAgent->new();
|
||||||
_iterate("https://api.bitbucket.com/2.0/repositories/$owner/$repo/pullrequests?state=OPEN", $auth, \%pulls, $ua);
|
_iterate("https://api.bitbucket.com/2.0/repositories/$owner/$repo/pullrequests?state=OPEN", $auth, \%pulls, $ua);
|
||||||
my $tempdir = File::Temp->newdir("bitbucket-pulls" . "XXXXX", TMPDIR => 1);
|
my $tempdir = File::Temp->newdir("bitbucket-pulls" . "XXXXX", TMPDIR => 1);
|
||||||
my $filename = "$tempdir/bitbucket-pulls.json";
|
my $filename = "$tempdir/bitbucket-pulls-sorted.json";
|
||||||
open(my $fh, ">", $filename) or die "Cannot open $filename for writing: $!";
|
open(my $fh, ">", $filename) or die "Cannot open $filename for writing: $!";
|
||||||
print $fh encode_json \%pulls;
|
print $fh JSON::MaybeXS->new(canonical => 1, pretty => 1)->encode(\%pulls);
|
||||||
close $fh;
|
close $fh;
|
||||||
run(["jq", "-S", "."], '<', $filename, '>', "$tempdir/bitbucket-pulls-sorted.json") or die "jq command failed: $?";
|
my $storePath = addToStore($filename);
|
||||||
my $storePath = addToStore("$tempdir/bitbucket-pulls-sorted.json");
|
|
||||||
my $timestamp = time;
|
my $timestamp = time;
|
||||||
return { storePath => $storePath, revision => strftime "%Y%m%d%H%M%S", gmtime($timestamp) };
|
return { storePath => $storePath, revision => strftime "%Y%m%d%H%M%S", gmtime($timestamp) };
|
||||||
}
|
}
|
||||||
|
@@ -32,7 +32,7 @@ sub fetchInput {
|
|||||||
my $stdout = ""; my $stderr = ""; my $res;
|
my $stdout = ""; my $stderr = ""; my $res;
|
||||||
if (! -d $clonePath) {
|
if (! -d $clonePath) {
|
||||||
# Clone the repository.
|
# Clone the repository.
|
||||||
$res = run(timeout => 600,
|
$res = runCommand(timeout => 600,
|
||||||
cmd => ["darcs", "get", "--lazy", $uri, $clonePath],
|
cmd => ["darcs", "get", "--lazy", $uri, $clonePath],
|
||||||
dir => $ENV{"TMPDIR"});
|
dir => $ENV{"TMPDIR"});
|
||||||
die "Error getting darcs repo at `$uri':\n$stderr" if $res->{status};
|
die "Error getting darcs repo at `$uri':\n$stderr" if $res->{status};
|
||||||
|
@@ -137,8 +137,8 @@ sub fetchInput {
|
|||||||
my $res;
|
my $res;
|
||||||
if (! -d $clonePath) {
|
if (! -d $clonePath) {
|
||||||
# Clone everything and fetch the branch.
|
# Clone everything and fetch the branch.
|
||||||
$res = run(cmd => ["git", "init", $clonePath]);
|
$res = runCommand(cmd => ["git", "init", $clonePath]);
|
||||||
$res = run(cmd => ["git", "remote", "add", "origin", "--", $uri], dir => $clonePath) unless $res->{status};
|
$res = runCommand(cmd => ["git", "remote", "add", "origin", "--", $uri], dir => $clonePath) unless $res->{status};
|
||||||
die "error creating git repo in `$clonePath':\n$res->{stderr}" if $res->{status};
|
die "error creating git repo in `$clonePath':\n$res->{stderr}" if $res->{status};
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -146,9 +146,9 @@ sub fetchInput {
|
|||||||
# the remote branch for whatever the repository state is. This command mirrors
|
# the remote branch for whatever the repository state is. This command mirrors
|
||||||
# only one branch of the remote repository.
|
# only one branch of the remote repository.
|
||||||
my $localBranch = _isHash($branch) ? "_hydra_tmp" : $branch;
|
my $localBranch = _isHash($branch) ? "_hydra_tmp" : $branch;
|
||||||
$res = run(cmd => ["git", "fetch", "-fu", "origin", "+$branch:$localBranch"], dir => $clonePath,
|
$res = runCommand(cmd => ["git", "fetch", "-fu", "origin", "+$branch:$localBranch"], dir => $clonePath,
|
||||||
timeout => $cfg->{timeout});
|
timeout => $cfg->{timeout});
|
||||||
$res = run(cmd => ["git", "fetch", "-fu", "origin"], dir => $clonePath, timeout => $cfg->{timeout}) if $res->{status};
|
$res = runCommand(cmd => ["git", "fetch", "-fu", "origin"], dir => $clonePath, timeout => $cfg->{timeout}) if $res->{status};
|
||||||
die "error fetching latest change from git repo at `$uri':\n$res->{stderr}" if $res->{status};
|
die "error fetching latest change from git repo at `$uri':\n$res->{stderr}" if $res->{status};
|
||||||
|
|
||||||
# If deepClone is defined, then we look at the content of the repository
|
# If deepClone is defined, then we look at the content of the repository
|
||||||
@@ -156,16 +156,16 @@ sub fetchInput {
|
|||||||
if (defined $deepClone) {
|
if (defined $deepClone) {
|
||||||
|
|
||||||
# Is the target branch a topgit branch?
|
# Is the target branch a topgit branch?
|
||||||
$res = run(cmd => ["git", "ls-tree", "-r", "$branch", ".topgit"], dir => $clonePath);
|
$res = runCommand(cmd => ["git", "ls-tree", "-r", "$branch", ".topgit"], dir => $clonePath);
|
||||||
|
|
||||||
if ($res->{stdout} ne "") {
|
if ($res->{stdout} ne "") {
|
||||||
# Checkout the branch to look at its content.
|
# Checkout the branch to look at its content.
|
||||||
$res = run(cmd => ["git", "checkout", "--force", "$branch"], dir => $clonePath);
|
$res = runCommand(cmd => ["git", "checkout", "--force", "$branch"], dir => $clonePath);
|
||||||
die "error checking out Git branch '$branch' at `$uri':\n$res->{stderr}" if $res->{status};
|
die "error checking out Git branch '$branch' at `$uri':\n$res->{stderr}" if $res->{status};
|
||||||
|
|
||||||
# This is a TopGit branch. Fetch all the topic branches so
|
# This is a TopGit branch. Fetch all the topic branches so
|
||||||
# that builders can run "tg patch" and similar.
|
# that builders can run "tg patch" and similar.
|
||||||
$res = run(cmd => ["tg", "remote", "--populate", "origin"], dir => $clonePath, timeout => $cfg->{timeout});
|
$res = runCommand(cmd => ["tg", "remote", "--populate", "origin"], dir => $clonePath, timeout => $cfg->{timeout});
|
||||||
print STDERR "warning: `tg remote --populate origin' failed:\n$res->{stderr}" if $res->{status};
|
print STDERR "warning: `tg remote --populate origin' failed:\n$res->{stderr}" if $res->{status};
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -10,7 +10,6 @@ use Hydra::Helper::CatalystUtils;
|
|||||||
use Hydra::Helper::Nix;
|
use Hydra::Helper::Nix;
|
||||||
use File::Temp;
|
use File::Temp;
|
||||||
use POSIX qw(strftime);
|
use POSIX qw(strftime);
|
||||||
use IPC::Run qw(run);
|
|
||||||
|
|
||||||
=head1 NAME
|
=head1 NAME
|
||||||
|
|
||||||
@@ -112,12 +111,11 @@ sub fetchInput {
|
|||||||
my $ua = LWP::UserAgent->new();
|
my $ua = LWP::UserAgent->new();
|
||||||
_iterate("$githubEndpoint/repos/$owner/$repo/git/matching-refs/$type/$prefix?per_page=100", $auth, \%refs, $ua);
|
_iterate("$githubEndpoint/repos/$owner/$repo/git/matching-refs/$type/$prefix?per_page=100", $auth, \%refs, $ua);
|
||||||
my $tempdir = File::Temp->newdir("github-refs" . "XXXXX", TMPDIR => 1);
|
my $tempdir = File::Temp->newdir("github-refs" . "XXXXX", TMPDIR => 1);
|
||||||
my $filename = "$tempdir/github-refs.json";
|
my $filename = "$tempdir/github-refs-sorted.json";
|
||||||
open(my $fh, ">", $filename) or die "Cannot open $filename for writing: $!";
|
open(my $fh, ">", $filename) or die "Cannot open $filename for writing: $!";
|
||||||
print $fh encode_json \%refs;
|
print $fh JSON::MaybeXS->new(canonical => 1, pretty => 1)->encode(\%refs);
|
||||||
close $fh;
|
close $fh;
|
||||||
run(["jq", "-S", "."], '<', $filename, '>', "$tempdir/github-refs-sorted.json") or die "jq command failed: $?";
|
my $storePath = addToStore($filename);
|
||||||
my $storePath = addToStore("$tempdir/github-refs-sorted.json");
|
|
||||||
my $timestamp = time;
|
my $timestamp = time;
|
||||||
return { storePath => $storePath, revision => strftime "%Y%m%d%H%M%S", gmtime($timestamp) };
|
return { storePath => $storePath, revision => strftime "%Y%m%d%H%M%S", gmtime($timestamp) };
|
||||||
}
|
}
|
||||||
|
@@ -24,7 +24,6 @@ use Hydra::Helper::CatalystUtils;
|
|||||||
use Hydra::Helper::Nix;
|
use Hydra::Helper::Nix;
|
||||||
use File::Temp;
|
use File::Temp;
|
||||||
use POSIX qw(strftime);
|
use POSIX qw(strftime);
|
||||||
use IPC::Run qw(run);
|
|
||||||
|
|
||||||
sub supportedInputTypes {
|
sub supportedInputTypes {
|
||||||
my ($self, $inputTypes) = @_;
|
my ($self, $inputTypes) = @_;
|
||||||
@@ -83,12 +82,11 @@ sub fetchInput {
|
|||||||
_iterate($url, $baseUrl, \%pulls, $ua, $target_repo_url);
|
_iterate($url, $baseUrl, \%pulls, $ua, $target_repo_url);
|
||||||
|
|
||||||
my $tempdir = File::Temp->newdir("gitlab-pulls" . "XXXXX", TMPDIR => 1);
|
my $tempdir = File::Temp->newdir("gitlab-pulls" . "XXXXX", TMPDIR => 1);
|
||||||
my $filename = "$tempdir/gitlab-pulls.json";
|
my $filename = "$tempdir/gitlab-pulls-sorted.json";
|
||||||
open(my $fh, ">", $filename) or die "Cannot open $filename for writing: $!";
|
open(my $fh, ">", $filename) or die "Cannot open $filename for writing: $!";
|
||||||
print $fh encode_json \%pulls;
|
print $fh JSON::MaybeXS->new(canonical => 1, pretty => 1)->encode(\%pulls);
|
||||||
close $fh;
|
close $fh;
|
||||||
run(["jq", "-S", "."], '<', $filename, '>', "$tempdir/gitlab-pulls-sorted.json") or die "jq command failed: $?";
|
my $storePath = addToStore($filename);
|
||||||
my $storePath = addToStore("$tempdir/gitlab-pulls-sorted.json");
|
|
||||||
my $timestamp = time;
|
my $timestamp = time;
|
||||||
return { storePath => $storePath, revision => strftime "%Y%m%d%H%M%S", gmtime($timestamp) };
|
return { storePath => $storePath, revision => strftime "%Y%m%d%H%M%S", gmtime($timestamp) };
|
||||||
}
|
}
|
||||||
|
74
t/Hydra/Controller/Build/download.t
Normal file
74
t/Hydra/Controller/Build/download.t
Normal file
@@ -0,0 +1,74 @@
|
|||||||
|
use strict;
|
||||||
|
use warnings;
|
||||||
|
use Setup;
|
||||||
|
use Test2::V0;
|
||||||
|
use Catalyst::Test ();
|
||||||
|
use HTTP::Request::Common;
|
||||||
|
|
||||||
|
my %ctx = test_init();
|
||||||
|
|
||||||
|
Catalyst::Test->import('Hydra');
|
||||||
|
|
||||||
|
my $db = Hydra::Model::DB->new;
|
||||||
|
hydra_setup($db);
|
||||||
|
|
||||||
|
my $project = $db->resultset('Projects')->create({name => "tests", displayname => "", owner => "root"});
|
||||||
|
|
||||||
|
# Create a simple Nix expression that uses the existing build-product-simple.sh
|
||||||
|
my $jobsdir = $ctx{jobsdir};
|
||||||
|
my $nixfile = "$jobsdir/simple.nix";
|
||||||
|
open(my $fh, '>', $nixfile) or die "Cannot create simple.nix: $!";
|
||||||
|
print $fh <<"EOF";
|
||||||
|
with import ./config.nix;
|
||||||
|
{
|
||||||
|
simple = mkDerivation {
|
||||||
|
name = "build-product-simple";
|
||||||
|
builder = ./build-product-simple.sh;
|
||||||
|
};
|
||||||
|
}
|
||||||
|
EOF
|
||||||
|
close($fh);
|
||||||
|
|
||||||
|
# Create a jobset that uses the simple build
|
||||||
|
my $jobset = createBaseJobset("simple", "simple.nix", $ctx{jobsdir});
|
||||||
|
|
||||||
|
ok(evalSucceeds($jobset), "Evaluating simple.nix should succeed");
|
||||||
|
is(nrQueuedBuildsForJobset($jobset), 1, "Should have 1 build queued");
|
||||||
|
|
||||||
|
my $build = (queuedBuildsForJobset($jobset))[0];
|
||||||
|
ok(runBuild($build), "Build should succeed");
|
||||||
|
|
||||||
|
$build->discard_changes();
|
||||||
|
|
||||||
|
subtest "Test downloading build products (regression test for #1520)" => sub {
|
||||||
|
# Get the build URL
|
||||||
|
my $build_id = $build->id;
|
||||||
|
|
||||||
|
# First, check that the build has products
|
||||||
|
my @products = $build->buildproducts;
|
||||||
|
ok(scalar @products >= 1, "Build should have at least 1 product");
|
||||||
|
|
||||||
|
# Find the doc product (created by build-product-simple.sh)
|
||||||
|
my ($doc_product) = grep { $_->type eq "doc" } @products;
|
||||||
|
ok($doc_product, "Should have a doc product");
|
||||||
|
|
||||||
|
if ($doc_product) {
|
||||||
|
# Test downloading via the download endpoint
|
||||||
|
# This tests the serveFile function which was broken in #1520
|
||||||
|
my $download_url = "/build/$build_id/download/" . $doc_product->productnr . "/text.txt";
|
||||||
|
my $response = request(GET $download_url);
|
||||||
|
|
||||||
|
# The key test: should not return 500 error with "Can't use string ("1") as a HASH ref"
|
||||||
|
isnt($response->code, 500, "Download should not return 500 error (regression test for #1520)");
|
||||||
|
is($response->code, 200, "Download should succeed with 200")
|
||||||
|
or diag("Response code: " . $response->code . ", Content: " . $response->content);
|
||||||
|
|
||||||
|
like($response->header('Content-Security-Policy') // '', qr/\bsandbox\b/, 'CSP header present with sandbox');
|
||||||
|
|
||||||
|
# Check that we get actual content
|
||||||
|
ok(length($response->content) > 0, "Should receive file content");
|
||||||
|
is($response->content, "Hello\n", "Should get expected content");
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
done_testing();
|
@@ -58,24 +58,23 @@ if (!defined($pid = fork())) {
|
|||||||
ok(sendNotifications(), "Sent notifications");
|
ok(sendNotifications(), "Sent notifications");
|
||||||
|
|
||||||
kill('INT', $pid);
|
kill('INT', $pid);
|
||||||
|
waitpid($pid, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
# We expect $ctx{jobsdir}/server.py to create the file at $filename, but the time it
|
# We expect $ctx{jobsdir}/server.py to create the file at $filename, but the time it
|
||||||
# takes to do so is non-deterministic. We need to give it _some_ time to hopefully
|
# takes to do so is non-deterministic. We need to give it _some_ time to hopefully
|
||||||
# settle -- but not too much that it drastically slows things down.
|
# settle -- but not too much that it drastically slows things down.
|
||||||
for my $i (1..10) {
|
for my $i (1..10) {
|
||||||
if (! -f $filename) {
|
last if -f $filename;
|
||||||
diag("$filename does not yet exist");
|
diag("$filename does not yet exist");
|
||||||
sleep(1);
|
sleep(1);
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
open(my $fh, "<", $filename) or die ("Can't open(): $!\n");
|
open(my $fh, "<", $filename) or die ("Can't open(): $!\n");
|
||||||
my $i = 0;
|
my $request_uri = <$fh>;
|
||||||
my $uri = <$fh>;
|
|
||||||
my $data = <$fh>;
|
my $data = <$fh>;
|
||||||
|
|
||||||
ok(index($uri, "gitea/api/v1/repos/root/foo/statuses") != -1, "Correct URL");
|
ok(index($request_uri, "gitea/api/v1/repos/root/foo/statuses") != -1, "Correct URL");
|
||||||
|
|
||||||
my $json = JSON->new;
|
my $json = JSON->new;
|
||||||
my $content;
|
my $content;
|
||||||
|
@@ -19,6 +19,8 @@ use Test2::V0;
|
|||||||
require Catalyst::Test;
|
require Catalyst::Test;
|
||||||
Catalyst::Test->import('Hydra');
|
Catalyst::Test->import('Hydra');
|
||||||
|
|
||||||
|
skip_all("This test has been failing since the upgrade to Nix 2.30, and we don't yet know how to fix it.");
|
||||||
|
|
||||||
my $db = Hydra::Model::DB->new;
|
my $db = Hydra::Model::DB->new;
|
||||||
hydra_setup($db);
|
hydra_setup($db);
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user