Mailing List Archives
Authenticated access
|
|
|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Condor-users] out-of-memory issues in parallel universe
- Date: Sun, 16 Mar 2008 23:18:53 -0400
- From: "Robert E. Parrott" <parrott@xxxxxxxxxxxxxxxx>
- Subject: [Condor-users] out-of-memory issues in parallel universe
Hi folks,
We have some macih es with 4 Gb RAM, and some with 6 Gb RAM all in the
same cluster, and may be seeing some over-extension of memory for MPI
jobs that are asssuming that there's 6 Gb of RAM available.
Is there some way of specifying the image size, and restricting jobs
to larger memory compute nodes, for MPI jobs submitted in the parallel
universe?
Also, what's the recommended way to enforce memory constraints on
jobs, in general and in particular in the parallel universe? ulimit?
something else?
thanks,
rob
==========================
Robert E. Parrott, Ph.D. (Phys. '06)
Associate Director, Grid and
Supercomputing Platforms
Project Manager, CrimsonGrid Initiative
Harvard University Sch. of Eng. and App. Sci.
Maxwell-Dworkin 211,
33 Oxford St.
Cambridge, MA 02138
(617)-495-5045