- 04 Sep, 2013 5 commits
-
-
Morris Jette authored
For a system with hyperthreads and GRES bound to specific CPUs, there was an error in assigning CPUs
-
Morris Jette authored
Previous logic would allocate job resources on such a node rather than reject the job allocation on that node.
-
Morris Jette authored
Previous logic would pick CPUs then reject jobs that can not match GRES to the allocated CPUs. New logic first filters out CPUs that can not use the GRES, next picks CPUs for the job, and finally picks the GRES that best match those CPUs. bug 410
-
Morris Jette authored
-
Morris Jette authored
No change in logic, just rename a variable for greater clarity
-
- 03 Sep, 2013 3 commits
-
-
Morris Jette authored
Eliminate extra variable and avoid changing the contents of a string Minor refactoring of commit fa0af103
-
Morris Jette authored
None of these have ever occurred, but these changes will harden Slurm bug 403
-
Morris Jette authored
-
- 30 Aug, 2013 2 commits
-
-
Morris Jette authored
Report anything that is world writable.
-
Morris Jette authored
Directories created for executables, libraries, etc. have permissions based upon the umask, which may not be desired.
-
- 29 Aug, 2013 5 commits
-
-
Danny Auble authored
/* Current code (<= 2.1) has it so we start the new * job with the next step id. This could be used * when restarting to figure out which step the * previous run of this job stopped on. */
-
Danny Auble authored
-
Morris Jette authored
-
-
Morris Jette authored
-
- 28 Aug, 2013 11 commits
-
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
-
Morris Jette authored
due to multiple free calls caused by job arrays submitted to multiple partitions. The root cause is the job priority array of the original job being re-used by the subsequent job array entries. A similar problem that could be induced by the user specifying a job accounting frequency when submitting a job array is also fixed. bug 401
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
-
Danny Auble authored
sacctmgr.
-
Danny Auble authored
-
Morris Jette authored
Some uninitialized variables, possible NULL pointers, etc. None of these have been seen in practice that we know of, but these changes will bulletproof the code
-
Morris Jette authored
Never observed, but "clang" tool reports these as possible failures
-
- 27 Aug, 2013 10 commits
-
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
If reservation create request included a CoreCnt value and more nodes are required than configured, the logic in select/cons_res could go off the end of the core_cnt array. This patch adds a check for a zero value in the core_cnt array, which terminates the user-specified array. Back-port from master of commit 211c224b
-
Morris Jette authored
-
- 26 Aug, 2013 1 commit
-
-
Danny Auble authored
-
- 24 Aug, 2013 1 commit
-
-
Danny Auble authored
-
- 23 Aug, 2013 2 commits
-
-
Morris Jette authored
-
Morris Jette authored
This is a correction of a bug introduced in commit https://github.com/SchedMD/slurm/commit/ac44db862c8d1f460e55ad09017d058942ff6499 That commit eliminated the need of reading the node state information from squeue for performance reasons (mostly for large parallel systems in which the Prolog ran squeue, which generates a lot of simultaneous RPCs, slowing down the job launch process). It also assumed 1 CPU per node. If a pending job specified a node count of 1 and a task count larger than one, squeue was reporting the node count of the job as the same as the task count. This patch moves that same calculation of a pending job's minimum node count into slurmctld, so the squeue still does not need to read the node information, but can report the correct node count for pending jobs with minimal overhead.
-