- 22 Jul, 2015 1 commit
-
-
David Bigagli authored
-
- 21 Jul, 2015 2 commits
-
-
Chandler Wilkerson authored
This patch provides a rewrite of how /proc/cpuinfo is parsed in common_jag.c, as the original code made the incorrect assumption that cpuinfo follows a sane format across architectures ;-) The motivation for this patch is that the original code was producing stack smashing on a POWER7 running RHEL6.4 Red Hat adds -fstack-protector along with a lot of other protective CFLAGS when building RPMs. The code ran okay with -fno-stack-protector, but that is not the best work-around. So, the relevant /proc/cpuinfo line on an Intel (Xeon X5675) system looks like: cpu MHz : 3066.915 Whereas the relevant line in a POWER7 system is clock : 3550.000000MHz My patch replaces the assumption that the relevant number starts 11 characters into the string with another assumption: That the relevant number starts two characters after a colon in a string that matches (M|G)Hz. All in all, the function has a few more calls, which may be a performance issue if it has to be called multiple times, but since the section I edited only gets evaluated if we don't know the cpu frequency, getting it right will actually result in fewer string operations and unnecessary opens of /proc/cpuinfo for systems likewise affected. Finally, I also read the actual value into a double and multiply it up to the size indicated by the suffix, so we end up with KHz? It was unclear what the original code intended, since it matched both MHz and GHz, replaced the decimal point with a zero, and read the result as an int. -- Chandler Wilkerson Center for Research Computing Rice University
-
Danny Auble authored
This reverts commit 2c95e2d2. Conflicts: src/plugins/select/alps/basil_interface.c This is related to bug 1822. It isn't clear why the code was taken out in this commit in the first place and based off of commit 2e2de6a4 (which is the reason for the conflict) we tried unsuccessfully to put it back. It appears the only difference here is the addition of always setting mppnppn = 1 instead of always to job_ptr->details->ntasks_per_node when no ntasks is set. This appears to only be an issue with salloc or sbatch as ntasks is always set for srun.
-
- 18 Jul, 2015 1 commit
-
-
Brian Christiansen authored
Prevent slurmctld abort on update of advanced reservation that contains no nodes. bug 1814
-
- 17 Jul, 2015 4 commits
-
-
Morris Jette authored
srun command line of either --mem or --mem-per-cpu will override both the SLURM_MEM_PER_CPU and SLURM_MEM_PER_NODE environment variables. Without this change, salloc or sbatch setting --mem-per-cpu (or a configuration of DefMemPerCPU) would over-ride the step's --mem value.
-
Danny Auble authored
change was made.
-
Danny Auble authored
when removing a limit from an association on multiple clusters at the same time.
-
Danny Auble authored
to gain the correct limit when a parent account is root and you remove a subaccount's limit which exists on the parent account.
-
- 16 Jul, 2015 1 commit
-
-
Morris Jette authored
-
- 15 Jul, 2015 3 commits
-
-
Morris Jette authored
-
Nathan Yee authored
-
Nathan Yee authored
Bug 1798
-
- 14 Jul, 2015 2 commits
-
-
Danny Auble authored
-
Morris Jette authored
Previous logic could fail to update some tasks of a job array for some fields. bug 1777
-
- 13 Jul, 2015 2 commits
-
-
Morris Jette authored
Fix to job array update logic that can result in a task ID of 4294967294. To reproduce: $ sbatch --exclusive -a 1,3,5 tmp Submitted batch job 11825 $ scontrol update jobid=11825_[3,4,5] timelimit=3 $ squeue JOBID PARTITION NAME USER ST TIME NODES NODELIST(REASON) 11825_3 debug tmp jette PD 0:00 1 (None) 11825_4 debug tmp jette PD 0:00 1 (None) 11825_5 debug tmp jette PD 0:00 1 (None) 11825 debug tmp jette PD 0:00 1 (Resources) A new job array entry was created for task ID 4 and the "master" job array record now has a task ID of 4294967294. The logic with the bug was using the wrong variable in a test. bug 1790
-
Gene Soudlenkov authored
Bug 1799
-
- 10 Jul, 2015 2 commits
-
-
Morris Jette authored
remove new capabilities added in comit ad9c2413 Leave the new logic only in version 15.08, which has related performance improvements in the slurmctld agent code, see commit 53534f49
-
Morris Jette authored
Correct "sdiag" backfill cycle time calculation if it yields locks. A microsecond value was being treated as a second value resulting in an overflow in the calcuation. bug 1788
-
- 09 Jul, 2015 1 commit
-
-
Morris Jette authored
The slurmctld logic throttles some RPCs so that only one of them can execute at a time in order to reduce contention for the job, partition and node locks (only one of the effected RPCs can execute at any time anyway and this lets other RPC types run). While an RPC is stuck in the throttle function, do not count that thread against the slurmctld thread limit. but 1794
-
- 08 Jul, 2015 1 commit
-
-
Morris Jette authored
-
- 07 Jul, 2015 3 commits
-
-
Trey Dockendorf authored
This patch moves the QOS update of an existing job to be before the partition update. This ensures a new QOS value is the value used when doing validations against things like a partition's AllowQOS and DenyQOS. Currently if a two partitions have AllowQOS that do not share any QOS, the order of updates prevents a job from being moved from one partition to another using something like the following: scontrol update job=<jobID> partition=<new part> qos=<new qos>
-
David Bigagli authored
-
Morris Jette authored
Correct task layout with CR_Pack_Node option and more than 1 CPU per task. Previous logic would place one task per CPU launch too few tasks. bug 1781
-
- 06 Jul, 2015 2 commits
-
-
Morris Jette authored
Backfill scheduler now considers OverTimeLimit and KillWait configuration parameters to estimate when running jobs will exit. Initially the job's end time is estimated based upon it's time limit. After the time limit is reached, the end time estimate is based upon the OverTimeLimit and KillWait configuration parameters. bug 1774
-
Morris Jette authored
Backfill scheduler: The configured backfill_interval value (default 30 seconds) is now interpretted as a maximum run time for the backfill scheduler. Once reached, the scheduler will build a new job queue and start over, even if not all jobs have been tested. bub 1774
-
- 30 Jun, 2015 2 commits
-
-
Thomas Cadeau authored
Bug 1745
-
Brian Christiansen authored
This reverts commit 3f91f4b2.
-
- 29 Jun, 2015 1 commit
-
-
Nathan Yee authored
Bug 1745
-
- 25 Jun, 2015 1 commit
-
-
Morris Jette authored
-
- 24 Jun, 2015 1 commit
-
-
David Bigagli authored
-
- 23 Jun, 2015 1 commit
-
-
David Bigagli authored
-
- 22 Jun, 2015 3 commits
-
-
Morris Jette authored
Updates of existing bluegene advanced reservations did not work at all. Some multi-core configurations resulting in an abort due to creating core_bitmaps for the reservation that only had one bit per node rather than one bit per core. These bugs were introduced in commit 5f258072
-
David Bigagli authored
-
David Bigagli authored
-
- 19 Jun, 2015 1 commit
-
-
David Bigagli authored
-
- 15 Jun, 2015 1 commit
-
-
Morris Jette authored
Logic was assuming the reservation had a node bitmap which was being used to check for overlapping jobs. If there is no node bitmap (e.g. a licenses only reservation), an abort would result.
-
- 12 Jun, 2015 2 commits
-
-
Brian Christiansen authored
Bug 1739
-
Brian Christiansen authored
Bug 1743
-
- 11 Jun, 2015 1 commit
-
-
Brian Christiansen authored
Bug 1733
-
- 10 Jun, 2015 1 commit
-
-
Morris Jette authored
-