English 中文(简体)
how to limit number of concurrently running PBS jobs
原标题:
  • 时间:2010-01-12 23:35:14
  •  标签:
  • pbs
  • qsub

I have a 64-node cluster, running PBS Pro. If I submit many hundreds of jobs, I can get 64 running at once. This is great, except when all 64 jobs happen to be nearly I/O bound, and are reading/writing to the same disk. In such cases, I d like to be able to still submit all the jobs, but have a max of (say) 10 jobs running at a given time. Is there an incantation to qsub that will allow me to do such, without having administrative access to the cluster s PBS server?

问题回答

In TORQUE you can do this by setting a slot limit on a job array, as long as you can arrange the jobs as an array:

qsub script.sh -t 0-99%10

would limit 10 of them to running at once. If PBSPro has an equivalent to this then you can use that.

you could make them dependent on each other. or schedule them to start at different timepoints.

otherwise, your admin can reduce the number of simultaneous jobs you can run at the same time.





相关问题
Making qsub block until job is done?

Currently, I have a driver program that runs several thousand instances of a "payload" program and does some post-processing of the output. The driver currently calls the payload program directly, ...

Captuing job ID when submitting Perl-generated scripts to SGE

Perl can t open command pipes that run both in and out, which causes a problem when submitting to SGE via qsub because I lose the job id of my submission. How can I submit a generated script to SGE ...

SGE - QSUB fails to submit jobs in -sync mode

I have a perl script that prepares files for input to a binary program and submits the execution of the binary program to the SGE queueing system version 6.2u2. The jobs are submitted with the -sync ...

Running a job on multiple nodes of a GridEngine cluster

I have access to a 128-core cluster on which I would like to run a parallelised job. The cluster uses Sun GridEngine and my program is written to run using Parallel Python, numpy, scipy on Python 2.5....

how to limit number of concurrently running PBS jobs

I have a 64-node cluster, running PBS Pro. If I submit many hundreds of jobs, I can get 64 running at once. This is great, except when all 64 jobs happen to be nearly I/O bound, and are reading/...

热门标签