- 04 Nov, 2014 3 commits
-
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
This was an unrealized regression from commit 0da01963. The problem is we were clearing the job_ptr->job_resrcs too early. This patch fixes it to wait until the job is actually being requeued so it does the right thing.
-
- 31 Oct, 2014 5 commits
-
-
Danny Auble authored
pack it this way so we will not change it in 15.08
-
Danny Auble authored
-
Danny Auble authored
This isn't that big of an issue for 14.03, but 14.11 added more to this string which could overflow the buffer since sprintf is used instead of snprintf. Using xstrfmtcat fixes the issue and is easier to read code.
-
Danny Auble authored
-
Danny Auble authored
amount of tasks / number of node.
-
- 30 Oct, 2014 2 commits
-
-
David Bigagli authored
-
David Bigagli authored
to dereference it.
-
- 27 Oct, 2014 2 commits
-
-
Danny Auble authored
are specified. This is a fix to commit b9cc5b31 which just didn't know mc_ptr->ntasks_per_core is initialized to INFINITE. Without it the node_cnt packed would be set to 1 on the user tools. This fixes bug 1148.
-
Morris Jette authored
bug 1207
-
- 24 Oct, 2014 3 commits
-
-
Morris Jette authored
Use default_partition by default for the test
-
Morris Jette authored
Returned test passed, when it should have failed because of bad variable name.
-
David Singleton authored
We've seen slurmctld crashes due to negative job array indices.
-
- 23 Oct, 2014 4 commits
-
-
Morris Jette authored
The previous patch should work in most cases, but this should work more reliably and the comment is more clear bug 1196
-
David Bigagli authored
This reverts commit 7e65f924.
-
David Bigagli authored
-
Morris Jette authored
BGQ: Fix race condition when job fails due to hardware failure and is requeued. Previous code could result in slurmctld abort with NULL pointer. bug 1096
-
- 22 Oct, 2014 1 commit
-
-
Gennaro Oliva authored
-
- 21 Oct, 2014 1 commit
-
-
Morris Jette authored
Fix bug that prevented preservation of a job's GRES bitmap on slurmctld restart or reconfigure (bug was introduced in 14.03.5 "Clear record of a job's gres when requeued" and only applies when GRES mapped to specific files). bug 1192
-
- 20 Oct, 2014 4 commits
-
-
Danny Auble authored
-
David Bigagli authored
permission for the batch step.
-
David Bigagli authored
-
jette authored
Otherwise there will be no log file to write to, resulting in an abort bug 1185
-
- 18 Oct, 2014 1 commit
-
-
Nicolas Joly authored
-
- 17 Oct, 2014 6 commits
-
-
Morris Jette authored
-
Morris Jette authored
-
David Bigagli authored
-
David Bigagli authored
-
Morris Jette authored
Correct tracking of licenses for suspended jobs on slurmctld reconfigure or restart. Previously licenses for suspended jobs were not counted, so the license count could be exceeded with those jobs get resumed.
-
Danny Auble authored
commit a9dc50d4.
-
- 16 Oct, 2014 5 commits
-
-
Brian Christiansen authored
-
Danny Auble authored
to commit a9dc50d4.
-
Morris Jette authored
-
Morris Jette authored
Refine commit 5f89223f based upon feedback from David Gloe: * It's not only MPI jobs, but anything that uses PMI. That includes MPI, shmem, etc, so you may want to reword the error message. * I added the terminated flag because if multiple tasks on a node exit, you would get an error message from each of them. That reduces it to one error message per node. Cray bug 810310 prompted that change. * Since we're now relying on --kill-on-bad-exit, I think we should update the Cray slurm.conf template to default to 1 (set KillOnBadExit=1 in contribs/cray/slurm.conf.template). bug 1171
-
Morris Jette authored
Treat Cray MPI job calling exit() without mpi_fini() as fatal error for that specific task and let srun handle all timeout logic. Previous logic would cancel the entire job step and srun options for wait time and kill on exit were ignored. The new logic provides users with the following type of response: $ srun -n3 -K0 -N3 --wait=60 ./tmp Task:0 Cycle:1 Task:2 Cycle:1 Task:1 Cycle:1 Task:0 Cycle:2 Task:2 Cycle:2 slurmstepd: step 14927.0 task 1 exited without calling mpi_fini() srun: error: tux2: task 1: Killed Task:0 Cycle:3 Task:2 Cycle:3 Task:0 Cycle:4 ... bug 1171
-
- 15 Oct, 2014 3 commits
-
-
Morris Jette authored
This fixes a race condition if the slurmctld needed to power up a node shortly after startup. Previously it would execute the ResumeProgram twice for effected nodes.
-
Morris Jette authored
Without this change, a node in the cloud that failed to power up, would not have its NoResponding flag cleared, which would prevent its later use. The NoResponding flag is now cleared when manuallly when the node is modified to PowerDown.
-
Morris Jette authored
If a batch job launch to the cloud fails, permit an unlimited number of job requeues. Previously the job would abort on the second launch failure.
-