- 10 Jan, 2013 5 commits
-
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
-
Danny Auble authored
-
Danny Auble authored
-
- 09 Jan, 2013 14 commits
-
-
Danny Auble authored
-
Danny Auble authored
-
Nathan Yee authored
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
-
David Bigagli authored
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
-
Danny Auble authored
allow the use of accounting features like associations, qos and limits but not keep track of jobs or steps in accounting.
-
Danny Auble authored
-
Morris Jette authored
-
- 08 Jan, 2013 12 commits
-
-
Danny Auble authored
-
Morris Jette authored
Conflicts: testsuite/expect/test15.21
-
jette authored
-
jette authored
-
Morris Jette authored
-
Nathan Yee authored
-
Morris Jette authored
-
Rod Schultz authored
One of our testers has observed that when a long running job continues to run after a maintenance reservation comes into effect sinfo reports the node as being in the allocated state while scontrol shows it to be in the maintenance state. This can happen when a node is not completely allocated. (select cons_res, a partition which is not Shared=EXCLUSIVE, jobs allocated without –exclusive, or jobs that are allocated only some of the cpus on a node.) Execution paths leading up to calls to node_state_string (slurm_protocol_defs.c) or node_state_string_compact, in scontrol, test for allocated_cpus less that total_cpus on the node and set the node state to MIXED rather than ALLOCATED, while similar paths in sinfo do not. I think this is probably a bug, since the mixed state is defined and think it is desirable that both command return the same result. The problem can be fixed with two logic changes (in multiple places) 1) node_state_string and node_state_string_compact have to check for mixed as well as allocated before returning the MAINT state. This means that the reported state for the node with the allocated job will be MIXED. 2) Sinfo must also check allocated_cpus less than total_cpus and set the state to MIXED before calling either node_state_string or node_state_string_compact. The attached patch (against 2.5.1) makes these changes. The attached script is a test case.
-
Morris Jette authored
-
Morris Jette authored
Phase 1 of effort. See "man sbatch" option -a/--array option for details. Creates job records using sbatch. Reports job arrays using scontrol or squeue. More work coming soon...
-
Danny Auble authored
today) so bitmaps are always 32bits. If one would like to use 64bit bitmaps just #define USE_64BIT_BITSTR in config.h.
-
Danny Auble authored
instead of a large array. This appears to speed up the process a big deal before we were seeing times of over 6000 usecs just to memset the array for a 5D system. With this patch on average the whole process takes around 1000 usecs with many being way under that.
-
- 07 Jan, 2013 1 commit
-
-
Danny Auble authored
-
- 04 Jan, 2013 5 commits
-
-
jette authored
Make sure out of memory gets logged properly for slurmctld in foreground Fix slurmd and slurmdbd to log out of memory to stdout in foreground
-
jette authored
-
Mark A. Grondona authored
The MPIRUN_PROCESSES variable set by the mpi/mvapich plugin probably is not needed for most if not all recent versions of mvapich. This environment variable also negatively affects job scalability since its length is proportional to the number of tasks in a job. In fact, for very large jobs, the increased environment size can lead to failures in execve(2). Since MPIRUN_PROCESSES *might* be required in some older versions of mvapich, this patch disables the setting of that variable completely only if SLURM_NEED_MVAPICH_MPIRUN_PROCESSES is not set in the job's environment. (Thus, by default MPIRUN_PROCESSES is disabled, but the old behavior may be restored by setting the environment variable above)
-
https://github.com/SchedMD/slurmjette authored
-
jette authored
-
- 03 Jan, 2013 3 commits
-
-
Morris Jette authored
Conflicts: META NEWS
-
Morris Jette authored
-
Morris Jette authored
-