- 04 Feb, 2015 2 commits
-
-
David Bigagli authored
-
Morris Jette authored
Enable CUDA v7.0+ use with a Slurm configuration of TaskPlugin=task/cgroup ConstrainDevices=yes (in cgroup.conf). With that configuration CUDA_VISIBLE_DEVICES will start at 0 rather than the device number. bug 1421
-
- 03 Feb, 2015 6 commits
-
-
David Bigagli authored
debug2 instead of info.
-
David Bigagli authored
SLURM_JOB_PARTITION to be the one in which the job started.
-
Morris Jette authored
If using proctrack/cgroup and gres/gpu, always start CUDA_VISIBLE_DEVICES environment variable numbering at 0. bug 1421
-
Danny Auble authored
reservations.
-
Danny Auble authored
This is an add on to commit 2e5142ef. Servicing Bug 1418.
-
David Bigagli authored
the partition in which the job runs.
-
- 02 Feb, 2015 3 commits
-
-
David Bigagli authored
-
Danny Auble authored
Front End systems.
-
David Bigagli authored
-
- 31 Jan, 2015 2 commits
-
-
Danny Auble authored
-
Danny Auble authored
initially been added.
-
- 30 Jan, 2015 3 commits
-
-
David Bigagli authored
-
David Bigagli authored
-
Danny Auble authored
-
- 29 Jan, 2015 1 commit
-
-
Brian Christiansen authored
-
- 28 Jan, 2015 5 commits
-
-
Morris Jette authored
Project seems to be defunct. Last update March 2009. http://xcpu.sourceforge.net/
-
David Bigagli authored
-
Brian Christiansen authored
-
David Bigagli authored
-
Brian Christiansen authored
Bug 1397
-
- 27 Jan, 2015 2 commits
-
-
Brian Christiansen authored
Bug 1384
-
David Bigagli authored
-
- 26 Jan, 2015 1 commit
-
-
Aaron Knister authored
-
- 23 Jan, 2015 2 commits
-
-
Dorian Krause authored
-
Morris Jette authored
-
- 22 Jan, 2015 2 commits
-
-
David Bigagli authored
-
Danny Auble authored
-
- 21 Jan, 2015 4 commits
-
-
Morris Jette authored
If some tasks of a job array are runnable and the meta-job array record is not runable (e.g. held), the old logic could start a runable task then try to start the non-runable meta-job, discover it can not run, and set its reason to "BadConstraints". Test case: Make it so no jobs can start (partition stopped, slurmd down, etc.) submit a job array hold the job array release the first two tasks of the job array Make it so jobs can start
-
Morris Jette authored
Squeue modified to not merge tasks of a job array if their wait reasons differ. bug 1388
-
Morris Jette authored
No functions currently exist, only the plugin wrapper and stubbed functions.
-
Morris Jette authored
-
- 20 Jan, 2015 3 commits
-
-
David Bigagli authored
--export sbatch/srun command line option, propagate the users' environ to the execution side. #1367
-
Danny Auble authored
-
Morris Jette authored
Interpret a partition configuration of "Nodes=ALL" in slurm.conf as including all nodes defined in the cluster. but 1382
-
- 19 Jan, 2015 1 commit
-
-
jette authored
bug 1379
-
- 17 Jan, 2015 1 commit
-
-
jette authored
bug 1375
-
- 15 Jan, 2015 2 commits
-
-
Danny Auble authored
What this does is use the core level binding after each task is laid out to skip all the extra threads in the core so it doesn't give them to another task. It probably isn't perfect, but does solve all the scenarios I found.
-
Morris Jette authored
Fix for GRES scheduling in which there is CPU topology defined or GRES types defined and there is more than 1 GPU per topology record in slurmctld. Without this fix, only one GRES could be allocated from each defined topology. bug 1369
-