Mailing List Archives
Authenticated access
|
|
|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Condor-users] condor_rm not killing subprocesses
- Date: Fri, 03 Jun 2005 19:33:28 +0300
- From: Mark Silberstein <marks@xxxxxxxxxxxxxxxxxxxxxxx>
- Subject: Re: [Condor-users] condor_rm not killing subprocesses
It seems that your condor setup doesn't give a time to a program to
finish nicely when condor is evicting it - look at KILL expression.
Usually Condor first tries to kill with SIGTERM, and then when KILL
expression is true - it will kill with -9. It seems that bash doesn't
have a chance to clean up all its processes, which it does when you kill
with Ctl-C.
You may also want to specify kill_sig=SIGQUIT, which will cause Condor
to kill it with SIGQUIT first.
On Fri, 2005-06-03 at 01:18 -0400, Jacob Joseph wrote:
> Hi. I have a number of users who have taken to wrapping their jobs
> within shell scripts. Often, they'll use a for or while loop to execute
> a single command with various permutations. When such a job is removed
> with condor_rm, the main script is killed, but subprocesses spawned from
> inside a loop will not be killed and will continue to run on the compute
> machine. This naturally interferes with jobs which are later assigned
> to that machine.
>
> Does anyone know of a way to force bash subprocesses to be killed along
> with the parent upon removal with condor_rm? (This behavior is not
> unique to condor_rm. A kill to the parent also leaves the subprocess
> running.)
>
> -Jacob
> _______________________________________________
> Condor-users mailing list
> Condor-users@xxxxxxxxxxx
> https://lists.cs.wisc.edu/mailman/listinfo/condor-users