- 16 Jul, 2014 1 commit
-
-
Morris Jette authored
For improved support of job arrays
-
- 15 Jul, 2014 6 commits
-
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
-
David Bigagli authored
-
David Bigagli authored
which then kill the job or the job array.
-
Morris Jette authored
Added SchedulerParameters options of bf_max_job_array_resv to control how many tasks of a job array should have resources reserved for them.
-
- 14 Jul, 2014 3 commits
-
-
Morris Jette authored
Modify to work if the bitmap sizes of the job and user specification vary.
-
Morris Jette authored
-
Morris Jette authored
-
- 12 Jul, 2014 1 commit
-
-
Morris Jette authored
This adds a new function job_str_signal() which can be used to cancel/signal a job array using a single RPC rather than a series if RPCs, one per job record, which does not work with the new job array data structure. This new logic is untested.
-
- 11 Jul, 2014 1 commit
-
-
Morris Jette authored
-
- 07 Jul, 2014 7 commits
-
-
Morris Jette authored
Partitial fix for new job array data structure
-
Morris Jette authored
Fix squeue --array/-r option to print individual job array task records even if then are all in a single job record from slurmctld.
-
Morris Jette authored
-
Morris Jette authored
Change un/pack_bit_str() to pack_bit_str_hex(), which has much better scalability and performance.
-
Morris Jette authored
-
Morris Jette authored
set exit code on a failure mode previously ignored
-
Morris Jette authored
-
- 04 Jul, 2014 2 commits
-
-
Morris Jette authored
Move the formatting of a job array's task ID string from slurmctld to the client to reduce system overhead. There is logic in place to translate the full bitmap into a string, but in the worst case this can take roughly one minute (e.g. 1-1000000:2,2) so by default only the first 1024 bytes in the string are set. There is logic in the code to translate the full string, but commented out.
-
Morris Jette authored
-
- 03 Jul, 2014 5 commits
-
-
Morris Jette authored
Fix some anomalies in the main scheduling loop for job arrays Improve logging in the main scheduling loop for job arrays
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
Now able to run job arrays based upon the new job array data structure.
-
Morris Jette authored
This adds logic to copy a job record for a job array when we actually schedule one of them. More work is still needed.
-
- 02 Jul, 2014 5 commits
-
-
Morris Jette authored
Add logic to build job array task ID string with a step function rather than using the bit_unfmt() function, which is terribly slow under those conditions for large job arrays.
-
Morris Jette authored
-
Morris Jette authored
-
Morris Jette authored
Modify the job state RPC, squeue, sview, and scontrol to display the new job array task ID string.
-
Morris Jette authored
Add new data structure for job arrays so only one job record is required for all tasks in a pending job array Initialize and free the new structure
-
- 01 Jul, 2014 9 commits
-
-
Morris Jette authored
Add squeue support for SQUEUE_FORMAT2 environment variable, which is equivalent to the -O/--Format option.
-
Morris Jette authored
-
Morris Jette authored
Report correct partition associated with a step if the job is submitted to multiple partitions. Previously all of the partitions that the job could run in would be reported for the step.
-
Nathan Yee authored
Added squeue -O/--Format option that makes all job and step fields available for printing.
-
jette authored
Add the uid cache functions thread safe add function to clear the uid cache when checking for memory leaks
-
jette authored
-
Anders Halager authored
squeue has to convert uid_t to a char* username a lot. Every line of output(with the default format) and for every comparison when sorting by username. This causes multiple system calls to open and read /etc/passwd on standard systems and on my system running YP there were 17 unnecessary entries in strace per line in the output. The cache is very simple, just a sorted array that is resorted for every new name. The number of unique usernames is likely going to be so low that any machine can handle it. No attempt is made at freeing memory for the same reason.
-
Morris Jette authored
-
Morris Jette authored
Add more detail to sched/backfill logging
-