1. 13 Jul, 2018 1 commit
  2. 12 Jul, 2018 3 commits
  3. 09 Jul, 2018 1 commit
  4. 06 Jul, 2018 1 commit
    • Marshall Garey's avatar
      Fix leaking freezer cgroups. · 7f9c4f73
      Marshall Garey authored
      Continuation of 923c9b37.
      
      There is a delay in the cgroup system when moving a PID from one cgroup
      to another. It is usually short, but if we don't wait for the PID to
      move before removing cgroup directories the PID previously belonged to,
      we could leak cgroups. This was previously fixed in the cpuset and
      devices subsystems. This uses the same logic to fix the freezer
      subsystem.
      
      Bug 5082.
      7f9c4f73
  5. 04 Jul, 2018 1 commit
  6. 03 Jul, 2018 1 commit
  7. 26 Jun, 2018 4 commits
  8. 25 Jun, 2018 1 commit
  9. 22 Jun, 2018 1 commit
  10. 20 Jun, 2018 1 commit
    • Alejandro Sanchez's avatar
      Make job_start_data() multi partition aware on REQUEST_JOB_WILL_RUN. · 35a13703
      Alejandro Sanchez authored
      Previously the function was only testing against the first partition in
      the job_record. Now it detects if the job request is multi partition and
      if so then loops through all of them until the job will run in any or
      until the end of the list, returning the error code from the last one if
      the job won't run in any partition.
      
      Bug 5185
      35a13703
  11. 19 Jun, 2018 2 commits
  12. 18 Jun, 2018 1 commit
  13. 15 Jun, 2018 2 commits
  14. 12 Jun, 2018 3 commits
  15. 08 Jun, 2018 2 commits
  16. 06 Jun, 2018 1 commit
    • Brian Christiansen's avatar
      Don't allocate downed cloud nodes · be449407
      Brian Christiansen authored
      which were marked down due to ResumeTimeout.
      
      If a cloud node was marked down due to not responding by ResumeTimeout,
      the code inadvertently added the node back to the avail_node_bitmap --
      after being cleared by set_node_down_ptr(). The scheduler would then
      attempt to allocate the node again, which would cause a loop of hitting
      ResumeTimeout and allocating the downed node again.
      
      Bug 5264
      be449407
  17. 05 Jun, 2018 1 commit
  18. 31 May, 2018 1 commit
    • Alejandro Sanchez's avatar
      Fix incomplete RESPONSE_[RESOURCE|JOB_PACK]_ALLOCATION building path. · 17392e76
      Alejandro Sanchez authored
      There were two code paths building an allocation response by calling
      its own static _build_alloc_msg() function:
      
      1. src/slurmctld/proc_req.c
      2. src/slurmctld/srun_comm.c
      
      These two functions diverged and both had members that were not filled
      in but were filled in the other. This patch makes it so we change the
      signature of the one in proc_req.c to make it extern and then in
      srun_comm.c we call this newly common function.
      
      Also added cpu_freq_[min|max|gov] members in the common one since these
      were the only members missing in proc_req.c function (the one in
      srun_comm.c had more members missing, like all the ntasks_per*, account,
      qos or resv_name).
      
      Bug 4999.
      17392e76
  19. 30 May, 2018 7 commits
  20. 24 May, 2018 1 commit
    • Brian Christiansen's avatar
      Notify srun and ctld when unkillable stepd exits · 956a808d
      Brian Christiansen authored
      Commits f18390e8 and eed76f85 modified the stepd so that if the
      stepd encountered an unkillable step timeout that the stepd would just
      exit the stepd. If the stepd is a batch step then it would reply back
      to the controller with a non-zero exit code which will drain the node.
      But if an srun allocation/step were to get into the unkillable step
      code, the steps wouldn't let the waiting srun or controller know about
      the step going away -- leaving a hanging srun and job.
      
      This patch enables the stepd to notify the waiting sruns and the ctld of
      the stepd being done and drains the node for srun'ed alloction and/or
      steps.
      
      Bug 5164
      956a808d
  21. 21 May, 2018 1 commit
  22. 17 May, 2018 1 commit
  23. 15 May, 2018 1 commit
    • Alejandro Sanchez's avatar
      PMIx - override default paths at configure time if --with-pmix is used. · 635c0232
      Alejandro Sanchez authored
      Previously the default paths continued to be tested even when new ones
      were requested. This had as a consequence that if any of the new paths
      was the same as any of the default ones (i.e. /usr or /usr/local), the
      configure script was incorrectly erroring out specifying that a version
      of PMIx was already found in a previous path.
      
      Bug 5168.
      635c0232
  24. 10 May, 2018 1 commit
    • Alejandro Sanchez's avatar
      Fix different issues when requesting memory per cpu/node. · bf4cb0b1
      Alejandro Sanchez authored
      
      
      First issue was identified on multi partition requests. job_limits_check()
      was overriding the original memory requests, so the next partition
      Slurm validating limits against was not using the original values. The
      solution consists in adding three members to job_details struct to
      preserve the original requests. This issue is reported in bug 4895.
      
      Second issue was memory enforcement behavior being different depending on
      job the request issued against a reservation or not.
      
      Third issue had to do with the automatic adjustments Slurm did underneath
      when the memory request exceeded the limit. These adjustments included
      increasing pn_min_cpus (even incorrectly beyond the number of cpus
      available on the nodes) or different tricks increasing cpus_per_task and
      decreasing mem_per_cpu.
      
      Fourth issue was identified when requesting the special case of 0 memory,
      which was handled inside the select plugin after the partition validations
      and thus that could be used to incorrectly bypass the limits.
      
      Issues 2-4 were identified in bug 4976.
      
      Patch also includes an entire refactor on how and when job memory is
      is both set to default values (if not requested initially) and how and
      when limits are validated.
      
      Co-authored-by: default avatarDominik Bartkiewicz <bart@schedmd.com>
      bf4cb0b1