On Friday, January 26, 2018, 7:07:02 PM GMT+3:30, Jason Patton <jpatton@xxxxxxxxxxx> wrote:
Let's take your example with a condor pool containing:
Machine1 with 1 core
Machine2 with 2 cores
If you submit a parallel universe job with "machine_count = 3" as the
only requirement, condor will try to schedule three slots in the pool
with one core each. This will work **if** your pool is configured to
have one static slot per core (the default) or if you are using
partitionable slots. However, if your pool is configured with a single
static slot on each machine (perhaps with each slot containing all of
the cores), then your job will not match because you will only have
two slots -- one on Machine1 with one core, one on Machine2 with two
cores.
It's difficult to address if specific examples will work without
knowing exactly how your pool is configured.
Based on the condor_status output you've provided in these threads, it
seems that you have one static slot per core. This means that you can
only submit jobs that request a single core per slot (or per
node/machine in the parallel universe), but you can request as many
nodes (machine_count) as you want up to the number of slots in your
pool.
If you want to submit jobs that request more than a single core (e.g.
request_cpus = 2), then you will need to reconfigure your pool to have
more than one core per slot or consider using partitionable slots.
Here's the manual for configuring slots (Section 3.5.10):
Jason