- 22 Sep, 2015 1 commit
-
-
Danny Auble authored
Correct counting for job array limits, job count limit underflow possible when master cancellation of master job record. bug 1952
-
- 21 Sep, 2015 2 commits
-
-
Danny Auble authored
Also a very minor sanity check in job_mgr.c to make sure we at least have a task count. This shouldn't matter, but just to be as robust as possible.
-
Nathan Yee authored
only 1 job was accounted (against MaxSubmitJob) for when an array was submitted.
-
- 17 Sep, 2015 2 commits
-
-
David Bigagli authored
-
Tommi Tervo authored
-
- 11 Sep, 2015 3 commits
-
-
Morris Jette authored
-
Morris Jette authored
This prevents a step from being launched if the job is killed while the prolog is running. Reproducing the original failure requires use of srun to trigger the prolog and using scancel while that prolog is running. bug 1755
-
Brian Christiansen authored
And add missing documenation. Bug 1921
-
- 10 Sep, 2015 5 commits
-
-
Morris Jette authored
GRES were not being properly tracks for multiple simultaneous steps. A step which could have run later could be rejected as never being able to run. Replacement for commit dd842d79, which was reverted in commit 6f73812875c bug 1925
-
Morris Jette authored
That commit would address a limited subset of problems and introduce other bugs rather than fixing the root of the problem.
-
David Bigagli authored
-
David Bigagli authored
-
Danny Auble authored
and you use all the GRES up instead of reporting the configuration isn't available you hold the requesting step until the GRES is available.
-
- 09 Sep, 2015 2 commits
-
-
Morris Jette authored
-
Morris Jette authored
Don't trucate task ID information in "squeue --array/-r" output. Task ID info in sview also expanded to 64 characters (from ~16 chars).
-
- 08 Sep, 2015 7 commits
-
-
Morris Jette authored
-
Morris Jette authored
At the start of a scheduling cycle, the job's "reason" field can be cleared. If the scheduler fails to reach that job and set its value to a new reason, the original reason was lost and the state reports would report NoReason. This change saves the last reason for a job being in a pending state and reports that value to the user until we have a new valid reason for it still being in a PENDING state. bug 1919
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
bug 1920
-
David Bigagli authored
-
David Bigagli authored
-
- 02 Sep, 2015 4 commits
-
-
Nicolas Joly authored
Do not use full path for true utility in testsuite. Its location differ across systems (/bin/true on Linux and /usr/bin/true on BSD).
-
David Bigagli authored
-
Morris Jette authored
Previous logic would set the avail_node_bitmap when a node was powered down, even if the initial state was DOWN or DRAINED. This made the node available for allocation to a job, which we don't want until the DOWN or DRAIN state is cleared. bug 1893
-
Morris Jette authored
This reverts commits 7660da9e 5c386455 and f6c5302b
-
- 01 Sep, 2015 6 commits
-
-
Brian Christiansen authored
Bug 1741
-
David Bigagli authored
-
Danny Auble authored
-
David Bigagli authored
-
David Bigagli authored
-
Danny Auble authored
-
- 28 Aug, 2015 4 commits
-
-
Morris Jette authored
This problem is reproducible by launching a job then killing the slurmstepd process. Under those conditions, requeue the job if possible (i.e. batch job with requeue option/configuration). This patch also improves the slurmctld logging when this happens. bug 1889
-
David Bigagli authored
-
David Bigagli authored
-
Morris Jette authored
This is a change in logic from commit 00099596 The original commit corrected the logic from CPU to core count, but used the cores_per_socket count rather than computing the total core count (cores_per_socket * sockets). bug 1830
-
- 27 Aug, 2015 3 commits
-
-
Morris Jette authored
Correct RebootProgram logic when executed outside of a maintenance reservation. Previous logic would mark the node up upon response to the reboot RPC (from slurmctld to slurmc) and when the node actually rebooted, flag that as an unexpected reboot. This new logic checks the node's up time to not mark the compute node as being usable until the reboot actually takes place. but 1866
-
Morris Jette authored
For testing purposes, add a slurmd flag so that it appears the compute node has rebooted whenever slurmd restarts.
-
Danny Auble authored
association manager.
-
- 26 Aug, 2015 1 commit
-
-
Morris Jette authored
Prevent job array task ID from being reported as NO_VAL if last task in the array gets requeued. The problem is that when that task starts, the task bitmap entry for it stays set, but the task counter gets decremented. If that job then gets requeued, under some conditions a failure to schedule it results in the array_task_id in the job record getting set to NO_VAL. Then when building the job info to report for squeue/scontrol, the string showing the pending task ID's is not rebuilt due to that counter being zero. All indications are that the job runs fine, only the information reported to squeue/scontrol is wrong. bug 1790
-