1. 12 Nov, 2014 2 commits
  2. 10 Nov, 2014 1 commit
    • Danny Auble's avatar
      Fix issue where exclusive allocations wouldn't lay tasks out correctly · 7461c119
      Danny Auble authored
      with CR_PACK_NODES.
      
      Really do commit d388dd67 a different way to get the same info and
      be able to lay out tasks correctly when --hint=nomultithread.
      
      tests on a 4 core 8 thread system are
      srun -n6 --hint=nomultithread --exclusive whereami | sort -h
      srun: cpu count 6
         0 snowflake0 - MASK:0x1
         1 snowflake0 - MASK:0x2
         2 snowflake0 - MASK:0x4
         3 snowflake0 - MASK:0x8
         4 snowflake1 - MASK:0x1
         5 snowflake1 - MASK:0x2
      
      and
      
      srun -n10 -N5 --hint=nomultithread --exclusive whereami | sort -h
      srun: cpu count 10
         0 snowflake0 - MASK:0x1
         1 snowflake0 - MASK:0x2
         2 snowflake0 - MASK:0x4
         3 snowflake0 - MASK:0x8
         4 snowflake1 - MASK:0x1
         5 snowflake1 - MASK:0x2
         6 snowflake1 - MASK:0x4
         7 snowflake2 - MASK:0x1
         8 snowflake3 - MASK:0x1
         9 snowflake4 - MASK:0x1
      7461c119
  3. 07 Nov, 2014 2 commits
  4. 06 Nov, 2014 4 commits
  5. 05 Nov, 2014 1 commit
  6. 04 Nov, 2014 2 commits
  7. 31 Oct, 2014 4 commits
  8. 30 Oct, 2014 1 commit
  9. 27 Oct, 2014 1 commit
  10. 24 Oct, 2014 1 commit
  11. 23 Oct, 2014 1 commit
  12. 21 Oct, 2014 1 commit
    • Morris Jette's avatar
      Fix job gres info clear on slurmctld restart · 1209a664
      Morris Jette authored
      Fix bug that prevented preservation of a job's GRES bitmap on slurmctld
      restart or reconfigure (bug was introduced in 14.03.5 "Clear record of a
      job's gres when requeued" and only applies when GRES mapped to specific
      files).
      bug 1192
      1209a664
  13. 20 Oct, 2014 4 commits
  14. 18 Oct, 2014 1 commit
  15. 17 Oct, 2014 3 commits
  16. 16 Oct, 2014 2 commits
    • Brian Christiansen's avatar
      e1c42895
    • Morris Jette's avatar
      Change Cray mpi_fini failure logic · 5f89223f
      Morris Jette authored
      Treat Cray MPI job calling exit() without mpi_fini() as fatal error for
      that specific task and let srun handle all timeout logic.
      Previous logic would cancel the entire job step and srun options
      for wait time and kill on exit were ignored. The new logic provides
      users with the following type of response:
      
      $ srun -n3 -K0 -N3 --wait=60 ./tmp
      Task:0 Cycle:1
      Task:2 Cycle:1
      Task:1 Cycle:1
      Task:0 Cycle:2
      Task:2 Cycle:2
      slurmstepd: step 14927.0 task 1 exited without calling mpi_fini()
      srun: error: tux2: task 1: Killed
      Task:0 Cycle:3
      Task:2 Cycle:3
      Task:0 Cycle:4
      ...
      
      bug 1171
      5f89223f
  17. 15 Oct, 2014 4 commits
  18. 14 Oct, 2014 2 commits
  19. 10 Oct, 2014 3 commits
    • Danny Auble's avatar
      6bf40ed9
    • Brian Christiansen's avatar
      5d6a2dc2
    • Dorian Krause's avatar
      Job step memory allocation logic fix · 0dd12469
      Dorian Krause authored
      This commit fixes a bug we observed when combining select/linear with
      gres. If an allocation was requested with a --gres argument an srun
      execution within that allocation would stall indefinitely:
      
      -bash-4.1$ salloc -N 1 --gres=gpfs:100
      salloc: Granted job allocation 384049
      bash-4.1$ srun -w j3c017 -n 1 hostname
      srun: Job step creation temporarily disabled, retrying
      
      The slurmctld log showed:
      
      debug3: StepDesc: user_id=10034 job_id=384049 node_count=1-1 cpu_count=1
      debug3:    cpu_freq=4294967294 num_tasks=1 relative=65534 task_dist=1 node_list=j3c017
      debug3:    host=j3l02 port=33608 name=hostname network=(null) exclusive=0
      debug3:    checkpoint-dir=/home/user checkpoint_int=0
      debug3:    mem_per_node=62720 resv_port_cnt=65534 immediate=0 no_kill=0
      debug3:    overcommit=0 time_limit=0 gres=(null) constraints=(null)
      debug:  Configuration for job 384049 complete
      _pick_step_nodes: some requested nodes j3c017 still have memory used by other steps
      _slurm_rpc_job_step_create for job 384049: Requested nodes are busy
      
      If srun --exclusive would have be used instead everything would work fine.
      The reason is that in exclusive mode the code properly checks whether memory
      is a reserved resource in the _pick_step_node() function.
      This commit modifies the alternate code path to do the same.
      0dd12469