[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Condor-users] Jobs retuning blank output



The directories all exist and permissions are correct.

I have been using it without the log entry anyway and its still
the same problem so I dont think thats anything to do with it.

Chris
----- Original Message ----- From: "Alexander Lincoln Read" <a.l.read@xxxxxxxxxx>
To: "Condor-Users Mail List" <condor-users@xxxxxxxxxxx>
Sent: Tuesday, October 04, 2005 2:07 PM
Subject: Re: [Condor-users] Jobs retuning blank output



You can't have a separate log file for each job.
Don't you get a warning that the log file is invalid?
Try e.g. "log =log/hostnames.log" instead.

The subdirectories output, error and log have to
exist before the submit will work.

Cheers, Alex

Chris Miles wrote:

Thanks for the response but still no luck.

Here is the job file I am submitting.

executable = /bin/hostname
universe = vanilla

should_transfer_files = YES
when_to_transfer_output = ON_EXIT

requirements = (Arch == "X86_64") && (OpSys == "LINUX")

output  = output/$(Process).out
error = error/$(Process).err
log = log/$(Process).log

Queue 5
The output from the job is blank
The error from the job is blank
The log from the job reports

001 (039.000.000) 10/04 13:32:24 Job executing on host:
<192.168.1.101:34179>
...
005 (039.000.000) 10/04 13:32:24 Job terminated.
        (1) Normal termination (return value 0)
                Usr 0 00:00:00, Sys 0 00:00:00  -  Run Remote Usage
                Usr 0 00:00:00, Sys 0 00:00:00  -  Run Local Usage
                Usr 0 00:00:00, Sys 0 00:00:00  -  Total Remote Usage
                Usr 0 00:00:00, Sys 0 00:00:00  -  Total Local Usage
        0  -  Run Bytes Sent By Job
        9384  -  Run Bytes Received By Job
        0  -  Total Bytes Sent By Job
        9384  -  Total Bytes Received By Job

And finally condor_q reports this about one of the jobs I submitted.

MyType = "Job"
TargetType = "Machine"
ClusterId = 40
QDate = 1128429240
CompletionDate = 0
Owner = "condor"
RemoteWallClockTime = 0.000000
LocalUserCpu = 0.000000
LocalSysCpu = 0.000000
RemoteUserCpu = 0.000000
RemoteSysCpu = 0.000000
ExitStatus = 0
NumCkpts = 0
NumRestarts = 0
NumSystemHolds = 0
CommittedTime = 0
TotalSuspensions = 0
LastSuspensionTime = 0
CumulativeSuspensionTime = 0
ExitBySignal = FALSE
CondorVersion = "$CondorVersion: 6.6.10 Jun 13 2005 $"
CondorPlatform = "$CondorPlatform: I386-LINUX_RH9 $"
RootDir = "/"
Iwd = "/home/condor/jobs/test"
JobUniverse = 5
Cmd = "/bin/hostname"
MinHosts = 1
MaxHosts = 1
CurrentHosts = 0
WantRemoteSyscalls = FALSE
WantCheckpoint = FALSE
JobStatus = 1
EnteredCurrentStatus = 1128429240
JobPrio = 0
User = "condor@xxxxxxxxxxxxxxxxxxxx <mailto:condor@xxxxxxxxxxxxxxxxxxxx>"
NiceUser = FALSE
Env = ""
JobNotification = 2
CoreSize = 0
KillSig = "SIGTERM"
Rank = 0.000000
In = "/dev/null"
TransferIn = FALSE
BufferSize = 524288
BufferBlockSize = 32768
ShouldTransferFiles = "YES"
WhenToTransferOutput = "ON_EXIT"
TransferFiles = "ONEXIT"
ImageSize = 10
ExecutableSize = 10
DiskUsage = 10
Requirements = ((Arch == "X86_64") && (OpSys == "LINUX")) && (Disk >=
DiskUsage)
 && ((Memory * 1024) >= ImageSize) && (HasFileTransfer)
PeriodicHold = FALSE
PeriodicRelease = FALSE
PeriodicRemove = FALSE
OnExitHold = FALSE
OnExitRemove = TRUE
LeaveJobInQueue = FALSE
Args = ""
GlobalJobId = "thebeast.cluster.int#1128429240#40.4"
ProcId = 4
UserLog = "/home/condor/jobs/test/log/4.log"
OutOrig = "output/4.out"
Out = "_condor_stdout_40.4"
ErrOrig = "error/4.err"
Err = "_condor_stderr_40.4"
WantMatchDiagnostics = TRUE
LastMatchTime = 1128429240
NumJobMatches = 1
ServerTime = 1128429241


thanks

Chris Miles


----- Original Message ----- *From:* Jaime Frey <mailto:jfrey@xxxxxxxxxxx> *To:* Condor-Users Mail List <mailto:condor-users@xxxxxxxxxxx> *Sent:* Friday, September 30, 2005 7:42 PM *Subject:* Re: [Condor-users] Jobs retuning blank output

    On Sep 30, 2005, at 8:25 AM, Chris Miles wrote:

    My jobs are submitting successfully now i have upgraded to 6.7
    but my output files
    are retuning blank. It does this using my own console application
    or even by calling /bin/hostname
    as the executable. Is there anything that could be causing this?
    problem with stdout?


    One likely cause is Condor thinks you have a shared filesystem and
    either 1) you don't or 2) your output files are on a local disk
    instead. On Windows, Condor assumes you don't have a shared
    filesystem by default and will transfer your job files between
    machines. On unix, it assumes you do have a shared filesystem by
    default. You can tell Condor to transfer the job files by adding
    the following to your submit file:

    should_transfer_files = YES
    when_to_transfer_output = ON_EXIT


+----------------------------------+---------------------------------+

| Jaime Frey | Public Split on Whether |

    |        jfrey@xxxxxxxxxxx <mailto:jfrey@xxxxxxxxxxx>
      |  Bush Is a Divider              |

    |  http://www.cs.wisc.edu/~jfrey/
    <http://www.cs.wisc.edu/%7Ejfrey/>  |         -- CNN Scrolling
    Banner |


+----------------------------------+---------------------------------+



    ------------------------------------------------------------------------
    _______________________________________________
    Condor-users mailing list
    Condor-users@xxxxxxxxxxx
    https://lists.cs.wisc.edu/mailman/listinfo/condor-users

------------------------------------------------------------------------

_______________________________________________
Condor-users mailing list
Condor-users@xxxxxxxxxxx
https://lists.cs.wisc.edu/mailman/listinfo/condor-users


_______________________________________________ Condor-users mailing list Condor-users@xxxxxxxxxxx https://lists.cs.wisc.edu/mailman/listinfo/condor-users