1. 28 Jan, 2015 1 commit
  2. 27 Jan, 2015 1 commit
  3. 26 Jan, 2015 1 commit
  4. 23 Jan, 2015 1 commit
  5. 22 Jan, 2015 1 commit
  6. 21 Jan, 2015 2 commits
    • Morris Jette's avatar
      fix job array scheduling anomaly · 3787c01f
      Morris Jette authored
      If some tasks of a job array are runnable and the meta-job array
      record is not runable (e.g. held), the old logic could start a
      runable task then try to start the non-runable meta-job, discover
      it can not run, and set its reason to "BadConstraints".
      
      Test case:
      Make it so no jobs can start (partition stopped, slurmd down, etc.)
      submit a job array
      hold the job array
      release the first two tasks of the job array
      Make it so jobs can start
      3787c01f
    • Morris Jette's avatar
      fix squeue merging of job arrays · 261580be
      Morris Jette authored
      Squeue modified to not merge tasks of a job array if their wait reasons
      differ.
      bug 1388
      261580be
  7. 20 Jan, 2015 2 commits
  8. 19 Jan, 2015 1 commit
  9. 17 Jan, 2015 1 commit
  10. 15 Jan, 2015 3 commits
    • Danny Auble's avatar
      Make CR_ONE_TASK_PER_CORE work correctly with task/affinity. · db926ab7
      Danny Auble authored
      What this does is use the core level binding after each task is laid out
      to skip all the extra threads in the core so it doesn't give them to
      another task.
      
      It probably isn't perfect, but does solve all the scenarios I found.
      db926ab7
    • Morris Jette's avatar
      GRES scheduling fix · 72cefd54
      Morris Jette authored
      Fix for GRES scheduling in which there is CPU topology defined or
      GRES types defined and there is more than 1 GPU per topology record
      in slurmctld. Without this fix, only one GRES could be allocated
      from each defined topology.
      bug 1369
      72cefd54
    • Morris Jette's avatar
      Fix for slurmctld abort on gres error · ce1d99f5
      Morris Jette authored
      The slurmctld could abort with a gres configuration having
      Type= configured, but no CPU binding configured.
      ce1d99f5
  11. 14 Jan, 2015 2 commits
  12. 13 Jan, 2015 2 commits
  13. 09 Jan, 2015 2 commits
  14. 07 Jan, 2015 4 commits
    • Danny Auble's avatar
      update NEWS from Merge · 8e7a62d1
      Danny Auble authored
      8e7a62d1
    • Aaron Knister's avatar
      Add pbs parser fix to NEWS · 729a58ac
      Aaron Knister authored
      729a58ac
    • Rémi Palancher's avatar
      avoid delay on commit for PMI task at rank 0 · bb6656dc
      Rémi Palancher authored
      Intel MPI, on MPI jobs initialisation through PMI, uses to call PMI_KVS_Put()
      many many times from task at rank 0, and each on these call is followed by
      PMI_KVS_Commit(). Slurm implementation of PMI_KVS_Commit() imposes a delay
      to avoid DDOS on original srun. This delay is proportional to the total number.
      It could be up to 3 secs for large jobs for ex. with 7168 tasks. Therefore,
      when Intel MPI calls PMI_KVS_Commit() 475 times (mesured on a test case) from
      task at rank 0, 28 minutes are spent in delay function.
      All other tasks in the job are waiting for a PMI_Barrier. Therefore, there is
      no risk for a DDOS from this single task 0. The patch alters the delaying time
      calculation to make sure task at rank 0 will does not be delayed. All other
      tasks are globally spreaded in the same time range as before.
      bb6656dc
    • Aaron Knister's avatar
      Add PMI2 fix to NEWS · 84d61f94
      Aaron Knister authored
      84d61f94
  15. 06 Jan, 2015 4 commits
  16. 05 Jan, 2015 1 commit
  17. 02 Jan, 2015 2 commits
  18. 01 Jan, 2015 1 commit
  19. 30 Dec, 2014 2 commits
  20. 29 Dec, 2014 1 commit
  21. 26 Dec, 2014 1 commit
  22. 23 Dec, 2014 3 commits
    • Morris Jette's avatar
      Fix bad job array task ID value · 48016f86
      Morris Jette authored
      Prevent invalid job array task ID value if a task is started using gang
      scheduling (i.e. the task starts in a SUSPENDED state). The task ID gets
      set to NO_VAL and the task string is also cleared.
      48016f86
    • Morris Jette's avatar
      Prevent gang resume of suspended job · 161d0336
      Morris Jette authored
      Prevent a job manually suspended from being resumed by gang scheduler once
      free resources are available.
      bug 1335
      161d0336
    • Dorian Krause's avatar
      set node state RESERVED on maint reservation delete · cf846644
      Dorian Krause authored
      we have hit the following problem that seems to be present in Slurm
      slurm-14-11-2-1 and previous versions. When a node is reserved and an
      overlapping maint reservation is created and later deleted the scontrol
      output will report the node as IDLE rather than RESERVED:
      
      + scontrol show node node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=node1 ReservationName=X
      Reservation created: X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=RESERVED ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=ALL flags=maint,ignore_jobs ReservationName=Y
      Reservation created: Y
      + sleep 10
      + grep State
      + scontrol show nodes node1
         State=MAINT ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol delete ReservationName=Y
      + sleep 10
      + scontrol show nodes node1
      + grep State
      *   State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1*
      + scontrol delete ReservationName=X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      
      Note that the after the deletion of reservation "X" the State=IDLE
      instead of State=RESERVED. I think that the delete_resv() function in
      slurmctld/reservation.c should call set_node_maint_mode(true) like
      update_resv() does. With the patch pasted at the end of this e-mail I
      get the following output which matches my expectation:
      
      + scontrol show node node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=node1 ReservationName=X
      Reservation created: X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=RESERVED ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=ALL flags=maint,ignore_jobs ReservationName=Y
      Reservation created: Y
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=MAINT ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol delete ReservationName=Y
      + sleep 10
      + scontrol show nodes node1
      + grep State
      *   State=RESERVED ThreadsPerCore=1 TmpDisk=0 Weight=1*
      + scontrol delete ReservationName=X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      
      Thanks,
      Dorian
      cf846644
  23. 22 Dec, 2014 1 commit