Mailing List Archives
Authenticated access
|
|
|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [HTCondor-users] Condor submit for multiple small jobs
- Date: Thu, 16 May 2013 13:52:29 +0100
- From: Brian Candler <B.Candler@xxxxxxxxx>
- Subject: Re: [HTCondor-users] Condor submit for multiple small jobs
On Thu, May 16, 2013 at 05:59:43PM +0530, Batsayan Das wrote:
> I am trying to understand the common practice used to submit multiple
> small jobs with low overhead. I have a pool of small jobs. I do not
> want to invoke condor_submit each and every time for each job.
I have found that dagman works well for this, as described at
http://research.cs.wisc.edu/htcondor/manual/current/2_10DAGMan_Applications.html#SECTION0031013000000000000000
By default, dagman only submits 5 jobs every 5 seconds, but you can crank
this up by setting e.g.
DAGMAN_MAX_SUBMITS_PER_INTERVAL = 200
in your condor_config.local
Useful tools:
condor_submit_dag -maxidle 500 foo.dag
(to stop your queue getting too huge at one time)
condor_q -dag
(shows DAG node names)
condor_q -run -dag
(shows DAG node names and which host each job is running on)
More info:
http://research.cs.wisc.edu/htcondor/manual/current/2_10DAGMan_Applications.html
http://research.cs.wisc.edu/htcondor/manual/current/condor_submit_dag
If the jobs are almost identical, you can submit a "cluster" of jobs in a
single submission:
executable = myjob.sh
arguments = $(process)
queue 1000
This will run myjob.sh 0, myjob.sh 1 ... myjob.sh 999
More info:
http://research.cs.wisc.edu/htcondor/manual/current/condor_submit.html
Other options you could consider are the python API and the SOAP API for
submitting jobs, but I have not tried these.
HTH,
Brian.