- 21 May, 2014 6 commits
-
-
Danny Auble authored
-
Danny Auble authored
wait for.
-
Danny Auble authored
based on the mask given.
-
Danny Auble authored
task/affinity.
-
Danny Auble authored
thread in a core.
-
Danny Auble authored
it can bind cyclically across sockets.
-
- 20 May, 2014 7 commits
-
-
Morris Jette authored
Previous logic assumed cpus_per_task=1, so ntasks_per_core option could spread the job across more cores than desired
-
Morris Jette authored
cpus-per-task support: Try to pack all CPUs of each tasks onto one socket. Previous logic could spread the tasks CPUs across multiple sockets.
-
Morris Jette authored
Previous logic was counting CPUs, but assuming each task would only use one CPU.
-
Dan Weeks authored
-
Danny Auble authored
This reverts commit b22268d8.
-
Danny Auble authored
-
Morris Jette authored
-
- 19 May, 2014 7 commits
-
-
Morris Jette authored
-
Nathan Yee authored
-
Morris Jette authored
-
Morris Jette authored
Conflicts: src/slurmctld/job_mgr.c
-
Morris Jette authored
Properly enforce job --requeue and --norequeue options. Previous logic was in three places not doing so (either ignoring the value, ANDing it with the JobRequeue configuration option or using the JobRequeue configuration option by itself). bug 821
-
Morris Jette authored
-
Morris Jette authored
There should be no change in behavior with the production code, but this will improve the robustness of the code if someone makes changes to the logic.
-
- 15 May, 2014 2 commits
-
-
Morris Jette authored
Add SelectTypeParameters option of CR_PACK_NODES to pack a job's tasks tightly on its allocated nodes rather than distributing them evenly across the allocated nodes. bug 819
-
Danny Auble authored
something you also get a signal which would produce deadlock. Fix Bug 601.
-
- 14 May, 2014 4 commits
-
-
Morris Jette authored
-
Morris Jette authored
Run EpilogSlurmctld for a job is killed during slurmctld reconfiguration. bug 806
-
Morris Jette authored
-
Morris Jette authored
Only if ALL of their partitions are hidden will a job be hidden by default. bug 812
-
- 13 May, 2014 7 commits
-
-
Morris Jette authored
If a batch job launch request can not be built (the script file is missing, a credential can not be created, or the user does not exist on the selected compute node), then cancel the job in a graceful fashion. Previously, the bad RPC would be sent to the compute node and that node DRAINED. see bug 807
-
Morris Jette authored
Correct SelectTypeParameters=CR_LLN with job selecition of specific nodes. Previous logic would in most instances allocate resources on all nodes to the job.
-
Morris Jette authored
Correct squeue's job node and CPU counts for requeued jobs. Previously, when a job was requeued, its CPU count reported was that of the previous execution. When combined with the --ntasks-per-node option, squeue would compute the expected node count. If the --exclusive option is also used, the node count reported by squeue could be off by a large margin (e.g. "sbatch --exclusive --ntasks-per-node=1 -N1 .." on requeue would use the number of CPUs on the allocated node to recompute the expected node count). bug 756
-
Danny Auble authored
jobacct_gather/cgroup.
-
Morris Jette authored
Support SLURM_CONF path which does not have "slurm.conf" as the file name. bug 803
-
Morris Jette authored
-
Morris Jette authored
-
- 12 May, 2014 7 commits
-
-
Morris Jette authored
If a job has non-responding node, retry job step create rather than returning with DOWN node error. bug 734
-
Morris Jette authored
-
Morris Jette authored
-
Puenlap Lee authored
Also correct related documentation
-
Nathan Yee authored
Add force option to all file removals ("rm ..." to "rm -f ..."). bug 673
-
Morris Jette authored
-
Hongjia Cao authored
Completing nodes is removed when calling _try_sched() for a job, which is the case in select_nodes(). If _try_sched() thinks the job can run now but select_nodes() returns ESLURM_NODES_BUSY, the backfill loop will be ended.
-