Compute Element Configuration

From SURAgrid
Jump to: navigation, search

Contents

Reference

OSG Documentation: Configuration File Help
OSG Documentation: Configuration File Format
OSG Documentation: Compute Element Installation

Configuration

Edit (with care!) $VDT_LOCATION/osg/etc/config.ini. Make a backup copy of the original. Because we don't have any certificates at this point, we will not be enabling RSV and GIP, which will attempt (and fail) to push data up to OSG. These will be noted below. For production these will be enabled. Here are some key parts of the config.ini file. A link to the one used for the Calclab CE is at the bottom of the page.


;=================================================================== 
;IMPORTANT 
;=================================================================== 

; You can get documentation on the syntax of this file at: 
; https://twiki.grid.iu.edu/bin/view/ReleaseDocumentation/ConfigurationFileFormat
; You can get documentation on the options for each section at:
; https://twiki.grid.iu.edu/bin/view/ReleaseDocumentation/ConfigurationFileHelp

;[DEFAULT]
; Use this section to define variables that will be used in other sections 
; For example, if you define a variable called dcache_root here 
; you can use it in the gip section as %(dcache_root)s (e.g. 
; my_vo_1_dir = %(dcache_root)s/my_vo_1 
; my_vo_2_dir = %(dcache_root)s/my_vo_2

; Defaults, please don't modify these variables

unavailable = UNAVAILABLE
default = UNAVAILABLE 

; Name these variables disable and enable rather than disabled and enabled 
; to avoid infinite recursions

disable = False
enable = True 

; You can modify the following and use them

localhost = [Replace with fully qualified domain name of compute element]
admin_email = [Replace with email address of OSG system administrator] 

;=================================================================== 
;Site Information 
;===================================================================

[Site Information] 

; The group option indicates the group that the OSG site should be listed in, 
; for production sites this should be OSG, for vtb or itb testing it should be 
; OSG-ITB [If you are not ready to advertise your compute element and just want 
; to test it, choose OSG-ITB and coordinate with OSG Integration Test Bed staff.] 
;
; YOU WILL NEED TO CHANGE THIS

group = OSG 

; The host_name setting should give the host name of the CE that is being 
; configured, this setting must be a valid dns name that resolves 
;

host_name = %(localhost)s 

; The resource setting should be set to the same value (OIM_resource_name) as used in the OIM 
; registration at the goc 
;
; YOU WILL NEED TO CHANGE THIS

resource = OIM_resource_name 

; The resource_group setting should be set to the same value (OIM_resource_group_name) as used in the OIM 
; registration at the goc 
;
; YOU WILL NEED TO CHANGE THIS

resource_group = OIM_resource_group_name 

; The sponsor setting should list the sponsors for your cluster, if your cluster 
; has multiple sponsors, you can separate them using commas or specify the  
; percentage using the following format 'osg, atlas, cms' or 
; 'osg:10, atlas:45, cms:45' 
;

sponsor = SURAgrid 

; The site_policy setting should give an url that lists your site's usage 
; policy (your_site_usage_policy_url)

site_policy = your_site_usage_policy_url 

; The contact setting should give the name (contact_name) of the admin/technical contact 
; for the cluster 
;
; YOU WILL NEED TO CHANGE THIS

contact = contact_name 

; The email setting should give the email address for the technical contact 
; for the cluster 
; 

email = %(admin_email)s 

; The city setting should give the city (city_name) that the cluster is located in 
;
; YOU WILL NEED TO CHANGE THIS

city = city_name 

; The country setting should give the country that the cluster is located in 
;
; YOU WILL NEED TO CHANGE THIS

country = US 

; The longitude setting should give the longitude (longitude_value) for the cluster's location 
; if you are in the US, this should be negative 
; accepted values are between -180 and 180 (e.g., -96.342888) 
;
; YOU WILL NEED TO CHANGE THIS

longitude = longitude_value 

; The latitude setting should give the latitude (latitude_value) for the cluster's location 
; accepted values are between -90 and 90 (e.g., 30.616914) 
;
; YOU WILL NEED TO CHANGE THIS

latitude = latitude_value 

;=================================================================== 
; For the following job manager sections (LSF, SGE, PBS, Condor) 
; you should delete the sections corresponding to job managers that 
; you are NOT using. E.g. if you are just using Condor on your 
; site, you can delete the LSF, SGE and PBS sections. 
;===================================================================


;=================================================================== 
;PBS 
;===================================================================

[PBS] 

; This section has settings for configuring your CE for a PBS job manager

; The enabled setting indicates whether you want your CE to use a PBS job 
; manager 
; valid answers are True or False

enabled = %(enable)s 

; The home setting should give the location of the pbs install directory 
; home = /var/spool/torque 
; home = /var/spool/torque

; The pbs_location setting should give the location of pbs install directory 
; This should be the same as the home setting above

pbs_location = /var/spool/torque 

; We may need to add this for GIP 
; pbs_path = /usr/local/bin

; The job_contact setting should give the contact string for the jobmanager 
; on this CE (e.g. host.name/jobmanager-pbs)

job_contact = %(localhost)s/jobmanager-pbs 

; The util_contact should give the contact string for the default jobmanager 
; on this CE (e.g. host.name/jobmanager)

util_contact = %(localhost)s/jobmanager 

; The wsgram setting should be set to True or False depending on whether you 
; wish to enable wsgram on this CE

wsgram = %(disable)s 

night_whitelist = SURAgrid
weekend_whitelist = SURAgrid 

;=================================================================== 
;Condor 
;===================================================================

[Condor] 

; This section has settings for configuring your CE for a Condor job manager

; The enabled setting indicates whether you want your CE to use a Condor job 
; manager 
; valid answers are True or False

enabled = %(enable)s 

; The condor_location setting should give the location of condor install directory

condor_location = %(unavailable)s 

; The condor_location setting should give the location of condor config file, 
; This is typically etc/condor_config within the condor install directory. 
; If you leave this set to %(unavailable)s, configure-osg will attempt to 
; determine the correct value.

condor_config = %(unavailable)s 

; The job_contact setting should give the contact string for the jobmanager 
; on this CE (e.g. host.name/jobmanager-condor) 

job_contact = %(localhost)s/jobmanager-condor 

; The util_contact should give the contact string for the default jobmanager 
; on this CE (e.g. host.name/jobmanager)

util_contact = %(localhost)s/jobmanager 

; The wsgram setting should be set to True or False depending on whether you 
; wish to enable wsgram on this CE

wsgram = %(disable)s 

;=================================================================== 
;SGE 
;=================================================================== 

[SGE] 

; This section has settings for configuring your CE for a SGE job manager

; The enabled setting indicates whether you want your CE to use a SGE job 
; manager 
; valid answers are True or False

enabled = %(enable)s  

; The sge_root setting should give the location of sge install directory 
;
; The VDT will bootstrap your SGE environment by sourcing 
; $SGE_ROOT/$SGE_CELL/common/settings.sh 
; where $SGE_ROOT and $SGE_CELL are the values given for sge_root and sge_cell.

sge_root = %(unavailable)s 

; The sge_cell setting should be set to the value of $SGE_CELL for your SGE 
; install.

sge_cell = %(unavailable)s 

; The job_contact setting should give the contact string for the jobmanager 
; on this CE (e.g. host.name/jobmanager-sge)

job_contact = %(localhost)s/jobmanager-sge  

; The util_contact should give the contact string for the default jobmanager 
; on this CE (e.g. host.name/jobmanager)

util_contact = %(localhost)s/jobmanager 

; The wsgram setting should be set to True or False depending on whether you 
; wish to enable wsgram on this CE

wsgram = %(disable)s 

;=================================================================== 
;LSF 
;===================================================================

[LSF] 

; This section has settings for configuring your CE for a LSF job manager

; The enabled setting indicates whether you want your CE to use a LSF job 
; manager 
; valid answers are True or False 
; enabled = %(disable)s

enabled = True 

; The home setting should give the location of the lsf install directory

home = %(unavailable)s  

; The lsf_location setting should give the location of lsf install directory 
; This should be the same as the home setting above

lsf_location = %(home)s 

; The job_contact setting should give the contact string for the jobmanager 
; on this CE (e.g. host.name/jobmanager-lsf)

job_contact = %(localhost)s/jobmanager-lsf 

; The util_contact should give the contact string for the default jobmanager 
; on this CE (e.g. host.name/jobmanager)

util_contact = %(localhost)s/jobmanager 

; The wsgram setting should be set to True or False depending on whether you 
; wish to enable wsgram on this CE

wsgram = %(disable)s 

;=================================================================== 
;Managed Fork 
;=================================================================== 

[Managed Fork] 

; The enabled setting indicates whether managed fork is in use on the system 
; or not. You should set this to True or False

enabled = %(disable)s 

;=================================================================== 
;Misc Services 
;===================================================================

[Misc Services] 

; Enter the fully qualified domain name of your GUMS server (gums_server_name)

use_cert_updater = %(enable)s gums_host = gums_server_name 

#GUMS needs to run on port 8443 otherwise major headaches!

authorization_method = xacml 

;=================================================================== 
;Cemon 
;===================================================================

[Cemon] 

# Disable Cemon unless you have real certs for rsvuser.

enabled = %(disable)s 

;=================================================================== 
;Gratia 
;=================================================================== 

[Gratia] 

; Disable Gratia unless you have real certs for rsvuser.

enabled = %(disable)s 

;=================================================================== 
;RSV 
;===================================================================

[RSV] 

# Disable RSV unless you have real certs for rsvuser. 
# We'll configure the rest below, but they won't run 
# unless we change this first line to enable. Substitute ce_host_name 
# with the fully qualified domain name of your compute element 

enabled = %(disable)s
rsv_user = rsvuser 

#decide on whether to use service cert or proxy

use_service_cert = False
rsv_cert_file = /etc/grid-security/rsvcert.pem
rsv_key_file =  /etc/grid-security/rsvkey.pem
enable_ce_probes = %(enable)s 
ce_hosts = ce_host_name
enable_gridftp_probes = True 

# Enter the fully qualified domain names of your GridFTP hosts (gridftp_host_names, separated by commas) 
# Enter the batch system (condor, pbs, sge or lsf) for batch_system_name

gridftp_hosts = gridftp_host_names
gridftp_dir = /tmp
gratia_probes = metric, batch_system_name, gridftp-transfer 

;=================================================================== 
;GIP 
;===================================================================

[GIP] 

# Can we add enabled = %(disable)s to this section?

enabled = %(disable)s 

# Enter the batch system (condor, pbs, sge or lsf) for batch_system_name

batch = batch_system_name 

;=================================================================== 
;Subcluster 
;===================================================================

# Replace XXXXXXX with an OSG-unique name (be creative here) 
# Substitute the other items in red with appropriate values for your compute cluster

[Subcluster XXXXXXX]
name = FACILITY_Resource_XXXXXXX
node_count = 57
ram_mb = 4096
cpu_model = Intel(R) Core(TM)2 Quad CPU Q9550 @ 2.83GHz
cpu_vendor = INTEL
cpu_speed_mhz = 2830
cpu_platform = i686
cpus_per_node = 1
cores_per_node = 4
inbound_network = False
outbound_network = True  

GIP tests

Edit the file $VDT_LOCATION/gip/etc/gip_tests.conf.
1) Change the results_dir to something that exists. Try /tmp to start with.
2) Set your CE's site name for site_names. 3) Comment out site_dns=fnal.gov. The rest should be OK. There's not much documentation on this, so I don't have a great understanding of it.

You can run $VDT_LOCATION/gip/bin/gip_diagnostic to check your GIP configuration. It may complain about user daemon not being able to write to some locations. The diagnostic is actually incorrect. To get rid of the warning make these files/directories group writable by the daemon group.

Files

config.ini for Calclab (RSV/GIP enabled)

Personal tools
Namespaces

Variants
Actions
Navigation
Toolbox