1. 20 Jan, 2015 1 commit
  2. 19 Jan, 2015 1 commit
  3. 17 Jan, 2015 1 commit
  4. 15 Jan, 2015 3 commits
    • Danny Auble's avatar
      Make CR_ONE_TASK_PER_CORE work correctly with task/affinity. · db926ab7
      Danny Auble authored
      What this does is use the core level binding after each task is laid out
      to skip all the extra threads in the core so it doesn't give them to
      another task.
      
      It probably isn't perfect, but does solve all the scenarios I found.
      db926ab7
    • Morris Jette's avatar
      GRES scheduling fix · 72cefd54
      Morris Jette authored
      Fix for GRES scheduling in which there is CPU topology defined or
      GRES types defined and there is more than 1 GPU per topology record
      in slurmctld. Without this fix, only one GRES could be allocated
      from each defined topology.
      bug 1369
      72cefd54
    • Morris Jette's avatar
      Fix for slurmctld abort on gres error · ce1d99f5
      Morris Jette authored
      The slurmctld could abort with a gres configuration having
      Type= configured, but no CPU binding configured.
      ce1d99f5
  5. 14 Jan, 2015 2 commits
  6. 13 Jan, 2015 2 commits
  7. 09 Jan, 2015 2 commits
  8. 07 Jan, 2015 4 commits
    • Danny Auble's avatar
      update NEWS from Merge · 8e7a62d1
      Danny Auble authored
      8e7a62d1
    • Aaron Knister's avatar
      Add pbs parser fix to NEWS · 729a58ac
      Aaron Knister authored
      729a58ac
    • Rémi Palancher's avatar
      avoid delay on commit for PMI task at rank 0 · bb6656dc
      Rémi Palancher authored
      Intel MPI, on MPI jobs initialisation through PMI, uses to call PMI_KVS_Put()
      many many times from task at rank 0, and each on these call is followed by
      PMI_KVS_Commit(). Slurm implementation of PMI_KVS_Commit() imposes a delay
      to avoid DDOS on original srun. This delay is proportional to the total number.
      It could be up to 3 secs for large jobs for ex. with 7168 tasks. Therefore,
      when Intel MPI calls PMI_KVS_Commit() 475 times (mesured on a test case) from
      task at rank 0, 28 minutes are spent in delay function.
      All other tasks in the job are waiting for a PMI_Barrier. Therefore, there is
      no risk for a DDOS from this single task 0. The patch alters the delaying time
      calculation to make sure task at rank 0 will does not be delayed. All other
      tasks are globally spreaded in the same time range as before.
      bb6656dc
    • Aaron Knister's avatar
      Add PMI2 fix to NEWS · 84d61f94
      Aaron Knister authored
      84d61f94
  9. 06 Jan, 2015 4 commits
  10. 05 Jan, 2015 1 commit
  11. 02 Jan, 2015 2 commits
  12. 01 Jan, 2015 1 commit
  13. 30 Dec, 2014 2 commits
  14. 29 Dec, 2014 1 commit
  15. 26 Dec, 2014 1 commit
  16. 23 Dec, 2014 3 commits
    • Morris Jette's avatar
      Fix bad job array task ID value · 48016f86
      Morris Jette authored
      Prevent invalid job array task ID value if a task is started using gang
      scheduling (i.e. the task starts in a SUSPENDED state). The task ID gets
      set to NO_VAL and the task string is also cleared.
      48016f86
    • Morris Jette's avatar
      Prevent gang resume of suspended job · 161d0336
      Morris Jette authored
      Prevent a job manually suspended from being resumed by gang scheduler once
      free resources are available.
      bug 1335
      161d0336
    • Dorian Krause's avatar
      set node state RESERVED on maint reservation delete · cf846644
      Dorian Krause authored
      we have hit the following problem that seems to be present in Slurm
      slurm-14-11-2-1 and previous versions. When a node is reserved and an
      overlapping maint reservation is created and later deleted the scontrol
      output will report the node as IDLE rather than RESERVED:
      
      + scontrol show node node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=node1 ReservationName=X
      Reservation created: X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=RESERVED ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=ALL flags=maint,ignore_jobs ReservationName=Y
      Reservation created: Y
      + sleep 10
      + grep State
      + scontrol show nodes node1
         State=MAINT ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol delete ReservationName=Y
      + sleep 10
      + scontrol show nodes node1
      + grep State
      *   State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1*
      + scontrol delete ReservationName=X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      
      Note that the after the deletion of reservation "X" the State=IDLE
      instead of State=RESERVED. I think that the delete_resv() function in
      slurmctld/reservation.c should call set_node_maint_mode(true) like
      update_resv() does. With the patch pasted at the end of this e-mail I
      get the following output which matches my expectation:
      
      + scontrol show node node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=node1 ReservationName=X
      Reservation created: X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=RESERVED ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol create reservation starttime=now duration=120 user=usr01000
      nodes=ALL flags=maint,ignore_jobs ReservationName=Y
      Reservation created: Y
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=MAINT ThreadsPerCore=1 TmpDisk=0 Weight=1
      + scontrol delete ReservationName=Y
      + sleep 10
      + scontrol show nodes node1
      + grep State
      *   State=RESERVED ThreadsPerCore=1 TmpDisk=0 Weight=1*
      + scontrol delete ReservationName=X
      + sleep 10
      + scontrol show nodes node1
      + grep State
         State=IDLE ThreadsPerCore=1 TmpDisk=0 Weight=1
      
      Thanks,
      Dorian
      cf846644
  17. 22 Dec, 2014 2 commits
    • Daniel Ahlin's avatar
      Auth/munge - Correct AccountingStoragePass parsing · 2edef50d
      Daniel Ahlin authored
      Correct parsing of AccountingStoragePass when specified in old format
      (just a path name)
      2edef50d
    • Rémi Palancher's avatar
      avoid delay on commit for PMI task at rank 0 · fcc11e22
      Rémi Palancher authored
      Intel MPI, on MPI jobs initialisation through PMI, uses to call PMI_KVS_Put()
      many many times from task at rank 0, and each on these call is followed by
      PMI_KVS_Commit(). Slurm implementation of PMI_KVS_Commit() imposes a delay
      to avoid DDOS on original srun. This delay is proportional to the total number.
      It could be up to 3 secs for large jobs for ex. with 7168 tasks. Therefore,
      when Intel MPI calls PMI_KVS_Commit() 475 times (mesured on a test case) from
      task at rank 0, 28 minutes are spent in delay function.
      All other tasks in the job are waiting for a PMI_Barrier. Therefore, there is
      no risk for a DDOS from this single task 0. The patch alters the delaying time
      calculation to make sure task at rank 0 will does not be delayed. All other
      tasks are globally spreaded in the same time range as before.
      fcc11e22
  18. 20 Dec, 2014 1 commit
  19. 19 Dec, 2014 3 commits
  20. 17 Dec, 2014 2 commits
  21. 16 Dec, 2014 1 commit
    • Morris Jette's avatar
      Fix job hash table bug · f293ce7c
      Morris Jette authored
      Fix job array hash table bug, could result in slurmctld infinite loop or
      invalid memory reference.
      bug 1309
      f293ce7c