Subject: [Condor-users] Globus to Condor cluster submission
Hello All,
I have configured a condor pool.
[grid-node@grid-server ~]$ condor_status
Name OpSys Arch State Activity LoadAv Mem ActvtyTime
slot1@an-lab-node1 LINUX INTEL Unclaimed Idle 0.000 497 0+00:14:18
slot2@an-lab-node1 LINUX INTEL Unclaimed Idle 0.000 497 0+00:14:18 slot1@an-lab-node3 LINUX INTEL Unclaimed Idle 0.000 497 0+00:14:24 slot2@an-lab-node3 LINUX INTEL Unclaimed Idle 0.160 497 0+00:14:25
slot1@grid-server. LINUX INTEL Unclaimed Idle 0.380 1006 0+00:25:04 slot2@grid-server. LINUX INTEL Unclaimed Idle 0.000 1006 0+00:25:05
Total Owner Claimed Unclaimed Matched Preempting Backfill
INTEL/LINUX 6 0 0 6 0 0 0
Total 6 0 0 6 0 0 0 [grid-node@grid-server ~]$
I have a submit file like
[grid-node@grid-server ~]$ cat memory-grid.submit #################### # # Example 2: Show off some fancy features including # use of pre-defined macros and logging.
# ####################
-- Submitter: grid-server.iiitm.ac.in : <192.168.33.188:34401> : grid-server.iiitm.ac.in
ID OWNER SUBMITTED RUN_TIME ST PRI SIZE CMD 209.0 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.1 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello
209.2 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.3 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.4 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello
209.5 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.6 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.7 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello
209.8 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.9 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 210.0 grid-node 10/28 01:10 0+00:00:33 R 0 0.0 gridftp_wrapper.sh
11 jobs; 10 idle, 1 running, 0 held [grid-node@grid-server ~]$
[grid-node@grid-server ~]$ condor_q
-- Submitter: grid-server.iiitm.ac.in : <192.168.33.188:34401> : grid-server.iiitm.ac.in
ID OWNER SUBMITTED RUN_TIME ST PRI SIZE CMD 209.0 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.1 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello
209.2 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.3 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.4 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello
209.5 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.6 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.7 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello
209.8 grid-node 10/28 01:10 0+00:00:00 I 0 29.3 hello 209.9 grid-node 10/28 01:10 0+00:00:04 C 0 29.3 hello 210.0 grid-node 10/28 01:10 0+00:02:14 R 0 0.0 gridftp_wrapper.sh
212.0 grid-node 10/28 01:12 0+00:00:00 I 0 1.7 hello 213.0 grid-node 10/28 01:12 0+00:00:00 I 0 1.7 hello 214.0 grid-node 10/28 01:12 0+00:00:00 I 0 1.7 hello
215.0 grid-node 10/28 01:12 0+00:00:00 I 0 1.7 hello 216.0 grid-node 10/28 01:12 0+00:00:00 I 0 1.7 hello 217.0 grid-node 10/28 01:12 0+00:00:00 I 0 1.7 hello
218.0 grid-node 10/28 01:12 0+00:00:00 I 0 1.7 hello 219.0 grid-node 10/28 01:12 0+00:00:00 I 0 1.7 hello 220.0 grid-node 10/28 01:12 0+00:00:00 I 0 1.7 hello
19 jobs; 18 idle, 1 running, 0 held You have new mail in /var/spool/mail/grid-node [grid-node@grid-server ~]$
When i run this file, the whole job is being execute on the slots1@grid-server and slot2@grid-server, but it should execute on other two nodes of the condor pool...
When i tun the above file without grid universe it runs on other slots which are satisfying the parameters, what could be the problem
2009-10-28T01:12:09.552+05:30 INFO handler.SubmitStateHandler [pool-1-thread-2,process:172] Job 92bcb220-c330-11de-bb54-b421c00b8b13 submitted with local job ID '211.000.000' 2009-10-28T01:12:11.804+05:30 INFO handler.SubmitStateHandler [pool-1-thread-10,process:172] Job 92a42110-c330-11de-bb53-b421c00b8b13 submitted with local job ID '213.000.000'
2009-10-28T01:12:11.825+05:30 INFO handler.SubmitStateHandler [pool-1-thread-6,process:172] Job 929c31d0-c330-11de-bb53-b421c00b8b13 submitted with local job ID '212.000.000' 2009-10-28T01:12:11.848+05:30 INFO handler.SubmitStateHandler [pool-1-thread-8,process:172] Job 9293cd60-c330-11de-bb53-b421c00b8b13 submitted with local job ID '214.000.000'
2009-10-28T01:12:12.210+05:30 INFO handler.SubmitStateHandler [pool-1-thread-3,process:172] Job 936acc70-c330-11de-bb54-b421c00b8b13 submitted with local job ID '215.000.000' 2009-10-28T01:12:13.163+05:30 INFO handler.SubmitStateHandler [pool-1-thread-5,process:172] Job 926606a0-c330-11de-bb53-b421c00b8b13 submitted with local job ID '216.000.000'
2009-10-28T01:12:13.979+05:30 INFO handler.SubmitStateHandler [pool-1-thread-4,process:172] Job 93d2dd10-c330-11de-bb54-b421c00b8b13 submitted with local job ID '217.000.000' 2009-10-28T01:12:13.983+05:30 INFO handler.SubmitStateHandler [pool-1-thread-10,process:172] Job 93deeb00-c330-11de-bb54-b421c00b8b13 submitted with local job ID '218.000.000'
2009-10-28T01:12:13.997+05:30 INFO handler.SubmitStateHandler [pool-1-thread-6,process:172] Job 9419f830-c330-11de-bb54-b421c00b8b13 submitted with local job ID '219.000.000' 2009-10-28T01:12:14.045+05:30 INFO handler.SubmitStateHandler [pool-1-thread-2,process:172] Job 9412cc40-c330-11de-bb54-b421c00b8b13 submitted with local job ID '220.000.000'
2009-10-28T01:12:15.613+05:30 INFO handler.DoneStateHandler [pool-1-thread-7,process:66] Job 92bcb220-c330-11de-bb54-b421c00b8b13 finished successfully 2009-10-28T01:12:34.790+05:30 INFO handler.DoneStateHandler [pool-1-thread-4,process:66] Job 929c31d0-c330-11de-bb53-b421c00b8b13 finished successfully
2009-10-28T01:12:35.955+05:30 INFO handler.DoneStateHandler [pool-1-thread-7,process:66] Job 92a42110-c330-11de-bb53-b421c00b8b13 finished successfully 2009-10-28T01:12:37.645+05:30 INFO handler.DoneStateHandler [pool-1-thread-9,process:66] Job 936acc70-c330-11de-bb54-b421c00b8b13 finished successfully
2009-10-28T01:12:37.995+05:30 INFO handler.DoneStateHandler [pool-1-thread-1,process:66] Job 93d2dd10-c330-11de-bb54-b421c00b8b13 finished successfully 2009-10-28T01:12:38.474+05:30 INFO handler.DoneStateHandler [pool-1-thread-5,process:66] Job 9419f830-c330-11de-bb54-b421c00b8b13 finished successfully
2009-10-28T01:12:39.019+05:30 INFO handler.DoneStateHandler [pool-1-thread-6,process:66] Job 926606a0-c330-11de-bb53-b421c00b8b13 finished successfully 2009-10-28T01:12:39.177+05:30 INFO handler.DoneStateHandler [pool-1-thread-3,process:66] Job 93deeb00-c330-11de-bb54-b421c00b8b13 finished successfully
2009-10-28T01:12:39.388+05:30 INFO handler.DoneStateHandler [pool-1-thread-10,process:66] Job 9293cd60-c330-11de-bb53-b421c00b8b13 finished successfully 2009-10-28T01:12:39.502+05:30 INFO handler.DoneStateHandler [pool-1-thread-5,process:66] Job 9412cc40-c330-11de-bb54-b421c00b8b13 finished successfully