1. 16 Nov, 2015 1 commit
  2. 13 Nov, 2015 2 commits
  3. 04 Nov, 2015 1 commit
  4. 22 Oct, 2015 2 commits
  5. 19 Oct, 2015 1 commit
  6. 09 Oct, 2015 1 commit
  7. 07 Oct, 2015 2 commits
  8. 06 Oct, 2015 3 commits
  9. 05 Oct, 2015 1 commit
  10. 03 Oct, 2015 1 commit
  11. 02 Oct, 2015 1 commit
    • Morris Jette's avatar
      Don't mark powered down node as not responding · 8c03a8bc
      Morris Jette authored
      This will only happen if a PING RPC for the node is already queued
        when the decision is made to power it down, then fails to get
        a response for the ping (since the node is already down).
      bug 1995
      8c03a8bc
  12. 30 Sep, 2015 3 commits
    • Morris Jette's avatar
      Reset job CPU count if CPUs/task ratio increased for mem limit · 836912bf
      Morris Jette authored
      If a job's CPUs/task ratio is increased due to configured MaxMemPerCPU,
      then increase it's allocated CPU count in order to enforce CPU limits.
      Previous logic would increase/set the cpus_per_task as needed if a
      job's --mem-per-cpu was above the configured MaxMemPerCPU, but NOT
      increase the min_cpus or max_cpus varilable. This resulted in allocating
      the wrong CPU count.
      836912bf
    • Brian Christiansen's avatar
      Enable srun -I to use pending step logic. · 0bf0e71f
      Brian Christiansen authored
      Continuation of 1252d1a1
      Bug 1938
      0bf0e71f
    • Morris Jette's avatar
      Don't start duplicate batch job · c1513956
      Morris Jette authored
      Requeue/hold batch job launch request if job already running. This is
        possible if node went to DOWN state, but jobs remained active.
      In addition, if a prolog/epilog failed DRAIN the node rather than
        setting it down, which could kill jobs that could continue to
        run.
      bug 1985
      c1513956
  13. 29 Sep, 2015 3 commits
  14. 28 Sep, 2015 1 commit
    • Morris Jette's avatar
      Fix for node state when shrinking jobs · 6c9d4540
      Morris Jette authored
      When nodes have been allocated to a job and then released by the
        job while resizing, this patch prevents the nodes from continuing
        to appear allocated and unavailable to other jobs. Requires
        exclusive node allocation to trigger. This prevents the previously
        reported failure, but a proper fix will be quite complex and
        delayed to the next major release of Slurm (v 16.05).
      bug 1851
      6c9d4540
  15. 23 Sep, 2015 1 commit
  16. 22 Sep, 2015 2 commits
  17. 21 Sep, 2015 2 commits
  18. 17 Sep, 2015 2 commits
  19. 11 Sep, 2015 3 commits
  20. 10 Sep, 2015 5 commits
  21. 09 Sep, 2015 2 commits