- 24 Sep, 2015 1 commit
-
-
Morris Jette authored
Previously, scontrol would generate an error if passed a comma delimited list of job IDs. A space delimited list would be accepted. This increases compatability with some other Slurm commands. bug 1929
-
- 23 Sep, 2015 12 commits
-
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
Conflicts: src/sacct/print.c
-
Danny Auble authored
The 2 came from the nodelist being "None assigned", which would be treated as 2 hosts when sent into hostlist.
-
Danny Auble authored
the default qos for the association.
-
Danny Auble authored
jobs. Bug 1969
-
Morris Jette authored
Pending job array records will be combined into single line by default, even if started and requeued or modified. bug 1759
-
Danny Auble authored
diversion.
-
Nathan Yee authored
bug 1874
-
Morris Jette authored
-
- 22 Sep, 2015 13 commits
-
-
Brian Gilmer authored
If user belongs to a group which has split entries in /etc/group search for its username in all groups. Ammendment to commit 93ead71a bug 1738
-
Morris Jette authored
-
Morris Jette authored
The file is not installed, but this should eliminate any possible confusion in its use.
-
Danny Auble authored
-
Morris Jette authored
If GRES are associated with specific CPUs and a job allocation includes GRES, which are not associated with the specific CPUs allocated to the job, then when the job is deallocated, an underflow error results. To reproduce: gres.conf: Name=gpu File=/dev/tty0 CPUs=0-5 Name=gpu File=/dev/tty1 CPUs=6-11 Name=gpu File=/dev/tty2 CPUs=12-17 Name=gpu File=/dev/tty3 CPUs=18-23 Then $ srun --gres=gpu:2 -N1 --ntasks-per-node=2 hostname In slurmctld log file: error: gres/gpu: job 695 dealloc node smd1 topo gres count underflow Logic modified to increment the count based upon the specific GRES actually allocated, ignoring the associated CPUs (too late to consider that after the GRES as picked).
-
Danny Auble authored
Conflicts: NEWS src/slurmctld/acct_policy.c
-
Danny Auble authored
-
Danny Auble authored
Also a very minor sanity check in job_mgr.c to make sure we at least have a task count. This shouldn't matter, but just to be as robust as possible.
-
Nathan Yee authored
only 1 job was accounted (against MaxSubmitJob) for when an array was submitted.
-
David Bigagli authored
-
Tommi Tervo authored
-
Morris Jette authored
-
Danny Auble authored
Correct counting for job array limits, job count limit underflow possible when master cancellation of master job record. bug 1952
-
- 21 Sep, 2015 10 commits
-
-
Brian Christiansen authored
-
Brian Christiansen authored
-
Morris Jette authored
-
Morris Jette authored
-
Axel Huebl authored
Implement an option NONE for not sending mails at all. Closes http://bugs.schedmd.com/show_bug.cgi?id=1962
-
Morris Jette authored
-
Danny Auble authored
Also a very minor sanity check in job_mgr.c to make sure we at least have a task count. This shouldn't matter, but just to be as robust as possible.
-
Tim Wickberg authored
-
Nathan Yee authored
only 1 job was accounted (against MaxSubmitJob) for when an array was submitted.
-
Manuel Rodriguez-Pascual authored
I've noticed that parameter JobCheckpointDir has a (from my point of view) inconsistent behavior. * in sbatch executions, it is exported as CWD * in srun it is also exported as CWD, * except when it is manually set with "--checkpoint-dir=dir". If so, that value is exported. * value defined in slurm.conf is, as far as I know, never read. I have created this small patch to correct that behaviour. Now it is exported with the value configured on slurm.conf. If nothing is set, returned value is the defined on common/read_config.h,
-
- 18 Sep, 2015 4 commits
-
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
The "scontrol hold/release" commands accept either "name=" or "jobname=". I've modified the documentation to only show "jobname" for consistentcy with the "scontrol update" command. I have also modified the "scancel" command to accept "--jobname=" in addition to the existing "--name=" and "-n".
-
Morris Jette authored
If a sleep was interrupted or ran a bit long, the backfill scheduler run times could be significantly wrong as the sleep time was based upon the calculation of sleep_count x desired_sleep_time. This new logic captures and uses the actual sleep time for good accuracy. bug 1939
-