- 27 Sep, 2013 3 commits
-
-
Danny Auble authored
-
Morris Jette authored
-
Morris Jette authored
This specifies the maximum depth the backfill scheduler should go in any single partition.
-
- 25 Sep, 2013 1 commit
-
-
Morris Jette authored
-
- 24 Sep, 2013 4 commits
-
-
Danny Auble authored
srun is pending. The step creation would use the old values and possibly hang srun since the step wouldn't be able to be created in the modified allocation.
-
Morris Jette authored
-
Danny Auble authored
into the future.
-
Morris Jette authored
-
- 23 Sep, 2013 2 commits
-
-
Danny Auble authored
-
Morris Jette authored
bug 428
-
- 17 Sep, 2013 2 commits
-
-
Morris Jette authored
for setdebugflags command, avoid parsing "-flagname" as an scontrol command line option.
-
Armin Größlinger authored
for CR_Socket.
-
- 14 Sep, 2013 2 commits
-
-
Morris Jette authored
Add support for "on" and "before*" options
-
David Bigagli authored
bugfixes and the new item extraction feature.
-
- 13 Sep, 2013 4 commits
-
-
Morris Jette authored
-
Morris Jette authored
Set PBS_ACCOUNT, PBS_ENVIRONMENT, and PBS_QUEUE only for batch jobs and only if the user submission sets the account and partition.
-
Morris Jette authored
-
Morris Jette authored
No support for PBS "before" options
-
- 12 Sep, 2013 1 commit
-
-
Morris Jette authored
-l accelerator=true|false (GPU use) -l mpiprocs=# (processors per node) -l naccelerators=# (GPU count) -l select=# (node count) -l ncpus=# (task count) -v key=value (environment variable) -W umask=# (set job's umask) Note: the -v option does NOT support quoted commas.
-
- 11 Sep, 2013 2 commits
-
-
Morris Jette authored
-l accelerator=true|false (GPU use) -l mpiprocs=# (processors per node) -l naccelerators=# (GPU count) -l select=# (node count) -l ncpus=# (task count) -v key=value (environment variable) -W umask=# (set job's umask)
-
Morris Jette authored
-
- 10 Sep, 2013 3 commits
-
-
Morris Jette authored
-
David Bigagli authored
that ended in the OverTimeLimit interval.
-
David Bigagli authored
-
- 09 Sep, 2013 2 commits
-
-
Danny Auble authored
-
Danny Auble authored
-
- 06 Sep, 2013 1 commit
-
-
Morris Jette authored
Caused by allocating single adapter per node of specific adapter type.
-
- 04 Sep, 2013 1 commit
-
-
Morris Jette authored
Previous logic would pick CPUs then reject jobs that can not match GRES to the allocated CPUs. New logic first filters out CPUs that can not use the GRES, next picks CPUs for the job, and finally picks the GRES that best match those CPUs. bug 410
-
- 30 Aug, 2013 1 commit
-
-
Morris Jette authored
Report anything that is world writable.
-
- 29 Aug, 2013 3 commits
-
-
Danny Auble authored
/* Current code (<= 2.1) has it so we start the new * job with the next step id. This could be used * when restarting to figure out which step the * previous run of this job stopped on. */
-
Danny Auble authored
-
-
- 28 Aug, 2013 2 commits
-
-
Morris Jette authored
due to multiple free calls caused by job arrays submitted to multiple partitions. The root cause is the job priority array of the original job being re-used by the subsequent job array entries. A similar problem that could be induced by the user specifying a job accounting frequency when submitting a job array is also fixed. bug 401
-
Danny Auble authored
sacctmgr.
-
- 27 Aug, 2013 1 commit
-
-
Morris Jette authored
If reservation create request included a CoreCnt value and more nodes are required than configured, the logic in select/cons_res could go off the end of the core_cnt array. This patch adds a check for a zero value in the core_cnt array, which terminates the user-specified array. Back-port from master of commit 211c224b
-
- 24 Aug, 2013 1 commit
-
-
Danny Auble authored
-
- 23 Aug, 2013 1 commit
-
-
Morris Jette authored
This is a correction of a bug introduced in commit https://github.com/SchedMD/slurm/commit/ac44db862c8d1f460e55ad09017d058942ff6499 That commit eliminated the need of reading the node state information from squeue for performance reasons (mostly for large parallel systems in which the Prolog ran squeue, which generates a lot of simultaneous RPCs, slowing down the job launch process). It also assumed 1 CPU per node. If a pending job specified a node count of 1 and a task count larger than one, squeue was reporting the node count of the job as the same as the task count. This patch moves that same calculation of a pending job's minimum node count into slurmctld, so the squeue still does not need to read the node information, but can report the correct node count for pending jobs with minimal overhead.
-
- 22 Aug, 2013 2 commits
-
-
Danny Auble authored
to avoid it thinking we don't have a cluster name.
-
Danny Auble authored
-
- 21 Aug, 2013 1 commit
-
-
Hongjia Cao authored
If there are completing jobs, a reconfigure will set wrong job/node state: all nodes of the completing job will be set allocated, and the job will not be removed even if the completing nodes are released. The state can only be restored by restarting slurmctld after the completing nodes released.
-