Hello,
i'm trying to adapt a configuration we have with LSF:
jobs from certain user groups are forced to run in a sl6 singularity
container;
the LSF job starter does something like this (minor details stripped)
if [[ "$USER" =~ ^(user001|user002)$ ]]; then
job_cmd=$*
singularity exec --home $HOME --pwd $HOME -B /cvmfs:/cvmfs -B
/tmp:/tmp -B /opt/exp_software/common:/opt/exp_software/common -B
/data:/data /opt/exp_software/common/UI_SL6.img ${job_cmd}
fi
With HTCondor (8.8.4) , I used the following conf in the WN:
EXECUTE = /home/condor/execute/
SINGULARITY_JOB = RegExp("^(user001|user002)$", TARGET.Owner)
SINGULARITY_IMAGE_EXPR = "/opt/exp_software/common/UI_SL6.img"
SINGULARITY_BIND_EXPR = "/cvmfs /tmp /opt/exp_software/common /data"
However i'm not sure how to pass: --home $HOME --pwd $HOME
as these vales are set at runtime, and trying something like
SINGULARITY_EXTRA_ARGUMENTS = "--home $HOME --pwd $HOME"
fails (does not resolve $HOME).
Looking at the job env i see:
HOME=/home/user001
PWD=/home/user001
And PWD looks wrong to me, as it should better be
PWD=/home/condor/execute/dir_<number>
which is the case when running without singularity (SINGULARITY_JOB =
False) where the job ends succesfully.
Looking inside job.err there is:
2020-01-22 17:55:37 (245 KB/s) - `dirac-install.py' saved [91191/91191]
tar: DIRAC/releasehistory.pdf: Wrote only 7680 of 10240 bytes
tar:
DIRAC/RequestManagementSystem/scripts/dirac-rms-list-req-cache.py:
Cannot write: No space left on device
[... lots more of these ...]
Any suggestion on what i could be doing wrong? Thank You
Stefano
_______________________________________________
HTCondor-users mailing list
To unsubscribe, send a message to htcondor-users-request@xxxxxxxxxxx
with a
subject: Unsubscribe
You can also unsubscribe by visiting
https://lists.cs.wisc.edu/mailman/listinfo/htcondor-users
The archives can be found at:
https://lists.cs.wisc.edu/archive/htcondor-users/