- 02 Oct, 2015 3 commits
-
-
Morris Jette authored
-
Morris Jette authored
This will only happen if a PING RPC for the node is already queued when the decision is made to power it down, then fails to get a response for the ping (since the node is already down). bug 1995
-
Morris Jette authored
If a job's CPUs/task ratio is increased due to configured MaxMemPerCPU, then increase it's allocated CPU count in order to enforce CPU limits. Previous logic would increase/set the cpus_per_task as needed if a job's --mem-per-cpu was above the configured MaxMemPerCPU, but NOT increase the min_cpus or max_cpus varilable. This resulted in allocating the wrong CPU count.
-
- 01 Oct, 2015 2 commits
-
-
Danny Auble authored
values.
-
Morris Jette authored
This required a fairly major re-write of the select plugin logic bug 1975
-
- 30 Sep, 2015 2 commits
-
-
Morris Jette authored
Correct some cgroup paths ("step_batch" vs. "step_4294967294", "step_exter" vs. "step_extern", and "step_extern" vs. "step_4294967295").
-
Morris Jette authored
Requeue/hold batch job launch request if job already running. This is possible if node went to DOWN state, but jobs remained active. In addition, if a prolog/epilog failed DRAIN the node rather than setting it down, which could kill jobs that could continue to run. bug 1985
-
- 29 Sep, 2015 2 commits
-
-
Brian Christiansen authored
Bug 1938
-
Brian Christiansen authored
Bug 1984
-
- 28 Sep, 2015 2 commits
-
-
Morris Jette authored
When nodes have been allocated to a job and then released by the job while resizing, this patch prevents the nodes from continuing to appear allocated and unavailable to other jobs. Requires exclusive node allocation to trigger. This prevents the previously reported failure, but a proper fix will be quite complex and delayed to the next major release of Slurm (v 16.05). bug 1851
-
Morris Jette authored
When nodes have been allocated to a job and then released by the job while resizing, this patch prevents the nodes from continuing to appear allocated and unavailable to other jobs. Requires exclusive node allocation to trigger. This prevents the previously reported failure, but a proper fix will be quite complex and delayed to the next major release of Slurm (v 16.05). bug 1851
-
- 25 Sep, 2015 2 commits
-
-
Morris Jette authored
-
Morris Jette authored
Add ability to change a job array's maximum running task count: "scontrol update jobid=# arraytaskthrottle=#" bug 1863
-
- 24 Sep, 2015 2 commits
-
-
Danny Auble authored
-
Danny Auble authored
option.
-
- 23 Sep, 2015 8 commits
-
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
The 2 came from the nodelist being "None assigned", which would be treated as 2 hosts when sent into hostlist.
-
Danny Auble authored
the default qos for the association.
-
Danny Auble authored
jobs. Bug 1969
-
Morris Jette authored
Pending job array records will be combined into single line by default, even if started and requeued or modified. bug 1759
-
Danny Auble authored
diversion.
-
- 22 Sep, 2015 4 commits
-
-
Danny Auble authored
-
Nathan Yee authored
only 1 job was accounted (against MaxSubmitJob) for when an array was submitted.
-
David Bigagli authored
-
Danny Auble authored
Correct counting for job array limits, job count limit underflow possible when master cancellation of master job record. bug 1952
-
- 21 Sep, 2015 4 commits
-
-
Brian Christiansen authored
-
Morris Jette authored
-
Axel Huebl authored
Implement an option NONE for not sending mails at all. Closes http://bugs.schedmd.com/show_bug.cgi?id=1962
-
Nathan Yee authored
only 1 job was accounted (against MaxSubmitJob) for when an array was submitted.
-
- 17 Sep, 2015 1 commit
-
-
David Bigagli authored
-
- 16 Sep, 2015 1 commit
-
-
Morris Jette authored
Fix teardown race condition that can result in infinite loop. bug 1947
-
- 15 Sep, 2015 1 commit
-
-
David Bigagli authored
-
- 13 Sep, 2015 1 commit
-
-
Danny Auble authored
-
- 11 Sep, 2015 5 commits
-
-
Morris Jette authored
This prevents a step from being launched if the job is killed while the prolog is running. Reproducing the original failure requires use of srun to trigger the prolog and using scancel while that prolog is running. bug 1755
-
Danny Auble authored
-
Danny Auble authored
anomaly when only asking for 1 (task_id was never set to INFINITE).
-
Danny Auble authored
-
Morris Jette authored
This prevents a step from being launched if the job is killed while the prolog is running. Reproducing the original failure requires use of srun to trigger the prolog and using scancel while that prolog is running. bug 1755
-