UConn-OSG Use Policy Statement
Richard Jones, University of Connecticut
last updated May 8, 2013
User Policies:
- Your use of the Open Science Grid (OSG) is subject to your
acceptance of the
OSG
User Acceptable Use Policy (AUP).
- Access to UConn-OSG resources is provided via GUMS/VOMS authorization
software and is restricted to OSG member VOs. Users are mapped to VO group
accounts exclusively when they obtain VOMS proxy credentials for their
session. No individual accounts are permitted.
- Any planned event that would impact the regular operation of processing
at our site will by preceded by public notices on
the UConn cluster message
board at least 24 hours in advance. Unplanned outages will be
reported on the same message board as soon as the issue is detected,
together with an estimate for the time when normal operations are expected
to resume.
Processing Policies:
- UConn-OSG consists primarily of AMD x86_64 and Intel i7 processors,
together with a smaller number of older 32-bit nodes. Compute resources are
shared between local users and OSG VO group accounts, with priority given
to local users. There are 516 job slots on the UConn site currently. The
Condor scheduler configuration allocates approximately half of the processors
to OSG Gluex jobs, and opportunistic running to the other half subject to
preemption by local jobs. Compute resources are made available to non-Gluex
OSG VO users on an opportunistic basis.
- All jobs are limited to a maximum of 72 hours
of runtime. Jobs exceeding
this limit may be terminated without warning and removed from the queue of
executable jobs.
- Jobs must be able to run within a small local filesystem on the worker
nodes pointed to by the environment variable $OSG_WN_TMP, which have about
10GB of local storage capacity
and 1GB of resident memory per job.
This local per-job storage quota includes all files that are locally
staged during job startup, plus whatever output and log files are generated by
the job during execution. Jobs that overflow this limit should consider how
to make use of space in the $OSG_APP and $OSG_DATA areas (see the Storage
section below), and also the local dCache Storage Element.
- All access to processing resources by OSG users is opportunistic.
Jobs submitted by local users have priority over those submitted by
non-local VO members. Any jobs that are preempted to allow a local user's
job to run will be requeued and eventually restarted. Any previously
executed work performed by such jobs will be lost, and the job will be
restarted from the beginning.
- UConn-OSG may suspend or ban DNs, VOs, Roles and/or CAs as necessary to
assure operations. Under normal running conditions, we will attempt to
contact the affected VOs and/or users through the OSG Operations Center
and attempt to provide up to 4 hours of warning before taking these
actions. During some running conditions we may be forced to take this
action without warning.
- VOs and VO members that desire to use UConn-OSG grid resources must
initialize their credentials using:
$VDT_LOCATION/voms/bin/voms-proxy-init
Those VOs and VO members that fail to use voms-proxy-init may be blocked
from accessing UConn-OSG grid resources.
- UConn-OSG requests that VOs submit using Condor Grid Monitor. VOs unable
to submit through grid monitor may be limited in the number of active
processes allowed. UConn-OSG requires that VO's submitting multi user pilot
jobs use the glexec package to validate the grid credentials of the actual
user who has submitted the job.
Storage Policies:
The following three canonical storage areas are provided for the use of
VO's to install software builds, stage shared data that is accessed by
multiple jobs, and locally cache large data files in the local dCache
Storage Element. Remote access to any of these areas should be over
gridftp, or through SRM in the case of the SE.
- $OSG_APP is an area for installing software builds of the applications to
be run by a given VO on this site. It is mounted read/write from the head
node, accessible to remote users via the jobmanager-fork queue on the CE.
It is mounted read-only on the worker nodes. Default quotas are 30 GB for
all the users in a VO. VO's that have more than one user may request that
only a single privileged user be allowed quota to write files. By default,
the directory has permission mode 1777.
- $OSG_DATA is a NFS-mounted area which is accessible read-write from the
head node, accessible to remote users via the jobmanager-fork queue on the CE.
It is mounted read-only on the worker nodes. Guest VO's are granted a small
default quota, which is shared between all members of the VO. The quota for
a guest VO can be increased upon request.
- OSG also defines $SITE_READ and $SITE_WRITE which are URL's that point
to a local storage element accessible by SRM and gridftp for reading, and
by SRM for writing. This is volatile storage with a total available space
of 100 TB. Files are deleted, if necessary, on a least recently used basis
and there are no quotas. This space is not backed up to tape.
- VO's can request opportunistic storage on the UConn SE in order to stage
data locally on the site, and make job startup more efficient. Space
allocation up to 1 TB will be made available to VO's upon request, subject
to availability of free space on the SE.