Source: slurm-wlm Version: 21.08.8.2-1 Severity: serious X-Debbugs-CC: debian...@lists.debian.org User: debian...@lists.debian.org Usertags: flaky
Dear maintainer(s),
I looked at the results of the autopkgtest of you package on armhf
because it was showing up as a regression for the upload of perl. I
noticed that the test regularly fails and I saw failures on other
architectures too.
Because the unstable-to-testing migration software now blocks on regressions in testing, flaky tests, i.e. tests that flip between passing and failing without changes to the list of installed packages, are causing people unrelated to your package to spend time on these tests. Don't hesitate to reach out if you need help and some more information from our infrastructure. Paul https://ci.debian.net/packages/s/slurm-wlm/ https://ci.debian.net/data/autopkgtest/testing/armhf/s/slurm-wlm/23115549/log.gz autopkgtest [01:29:11]: test sbatch: [----------------------- ● slurmctld.service - Slurm controller daemon
Loaded: loaded (/lib/systemd/system/slurmctld.service; enabled;
vendor preset: enabled)
Active: active (running) since Mon 2022-06-27 01:29:11 UTC; 10s ago Docs: man:slurmctld(8) Main PID: 3099 (slurmctld) Tasks: 10 Memory: 1.4M CPU: 4.907s CGroup: /system.slice/slurmctld.service ├─3099 /usr/sbin/slurmctld -D -s └─3103 "slurmctld: slurmscriptd" Jun 27 01:29:11 ci-178-32c04b32 systemd[1]: Started Slurm controller daemon.
Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: error: High
latency for 1000 calls to gettimeofday(): 288 microseconds
Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered
state of 1 nodes
Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered
JobId=1 Assoc=0
Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered
information about 1 jobs
Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered
state of 0 reservations
Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld:
read_slurm_conf: backup_controller not specified
Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Running as
primary controller
Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: No parameter
for mcs plugin, default values set
Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: mcs:
MCSParameters = (null). ondemand set.
● slurmd.service - Slurm node daemon
Loaded: loaded (/lib/systemd/system/slurmd.service; enabled;
vendor preset: enabled)
Active: active (running) since Mon 2022-06-27 01:29:11 UTC; 10s ago Docs: man:slurmd(8) Main PID: 3106 (slurmd) Tasks: 1 Memory: 1.1M CPU: 106ms CGroup: /system.slice/slurmd.service └─3106 /usr/sbin/slurmd -D -s Jun 27 01:29:11 ci-178-32c04b32 systemd[1]: Started Slurm node daemon.
Jun 27 01:29:11 ci-178-32c04b32 slurmd[3106]: slurmd: slurmd version
21.08.8-2 started
Jun 27 01:29:11 ci-178-32c04b32 slurmd[3106]: slurmd: slurmd started on
Mon, 27 Jun 2022 01:29:11 +0000
Jun 27 01:29:11 ci-178-32c04b32 slurmd[3106]: slurmd: CPUs=1 Boards=1
Sockets=1 Cores=1 Threads=1 Memory=513603 TmpDisk=2806 Uptime=1228661
CPUSpecList=(null) FeaturesAvail=(null) FeaturesActive=(null)
PARTITION AVAIL TIMELIMIT NODES STATE NODELIST test* up infinite 1 idle localhost NODELIST NODES PARTITION STATE localhost 1 test* idle Submitted batch job 2 autopkgtest [01:29:27]: test sbatch: -----------------------]
OpenPGP_signature
Description: OpenPGP digital signature
Read more here: Source link