Tags:
create new tag
,
view all tags
<verbatim> ################################### # General configuration variables # ################################### # List of the batch nodes hostnames and optionally the subcluster ID the # WN belongs to. An example file is available in # ${INSTALL_ROOT}/glite/yaim/examples/wn-list.conf # Change the path according to your site settings. #WN_LIST=${INSTALL_ROOT}/glite/yaim/examples/wn-list.conf WN_LIST=/root/wn-list.conf # List of unix users to be created in the service nodes. # The format is as follows: # UID:LOGIN:GID1,GID2,...:GROUP1,GROUP2,...:VO:FLAG: # An example file is available in ${INSTALL_ROOT}/glite/yaim/examples/ig-users.conf # Change the path according to your site settings. # For more information please check ${INSTALL_ROOT}/glite/yaim/examples/users.conf.README #USERS_CONF=${YAIM_ROOT}/glite/yaim/examples/ig-users.conf USERS_CONF=/root/ig-users.conf # List of the local accounts which a user should be mapped to. # The format is as follows: # "VOMS_FQAN":GROUP:GID:FLAG:[VO] # An example file is available in ${INSTALL_ROOT}/glite/yaim/examples/ig-groups.conf # Change the path according to your site settings. # For more information please check ${INSTALL_ROOT}/glite/yaim/examples/groups.conf.README # NOTE: comment out this variable if you want to specify a groups.conf per VO # under the group.d/ directory. #GROUPS_CONF=${YAIM_ROOT}/glite/yaim/examples/ig-groups.conf GROUPS_CONF=/root/ig-groups.conf # Uncomment this variable if you want to specify a local groups.conf # It is similar to GROUPS_CONF but used to specify a separate file # where local accounts specific to the site are defined. # LOCAL_GROUPS_CONF=my_local_groups.conf # Uncomment this variable if you are installing a mysql server # It is the MySQL admin password. MYSQL_PASSWORD=segreto # Uncomment this variable if you want to explicitely use pool # accounts for special users when generating the grid-mapfile. # If not defined, YAIM will decide whether to use special # pool accounts or not automatically # SPECIAL_POOL_ACCOUNTS=yes or no # INFN-GRID: Keep in site-info.def # WARNING! This variable will be removed in a future release JAVA_LOCATION="/usr/java/latest" ################################ # Site configuration variables # ################################ # Human-readable name of your site SITE_NAME=IGI-BOLOGNA ######################################### # ARGUS authorisation framework control # ######################################### # Set USE_ARGUS to yes to enable the configuration of ARGUS USE_ARGUS=no # In case ARGUS is to be used the following should be set # The ARGUS service PEPD endpoints as a space separated list: #ARGUS_PEPD_ENDPOINTS="http://pepd.example.org:8154/authz" ARGUS_PEPD_ENDPOINTS="https://vgrid06.cnaf.infn.it:8154/authz" # ARGUS resource identities: The resource ID can be set # for the cream CE, WMS and other nodes respectively. # If a resource ID is left unset the ARGUS configuration # will be skipped on the associated node. # CREAM_PEPC_RESOURCEID=urn:mysitename.org:resource:ce # WMS_PEPC_RESOURCEID=urn:mysitename.org:resource:wms # GENERAL_PEPC_RESOURCEID=urn:mysitename.org:resource:other CREAM_PEPC_RESOURCEID="http://cnaf.infn.it/cremino" # INFN-GRID: space separed list of the IP addresses of the NTP servers # (preferably set a local ntp server and a public one, ex. pool.ntp.org) #NTP_HOSTS_IP="131.154.1.53 131.154.1.103 pool.ntp.org" NTP_HOSTS_IP="ripe.cnaf.infn.it ntp-3.infn.it ntp-1.infn.it" #--------------------------# # Private network settings # #--------------------------# # INFN-GRID: Set PRIVATE_NETWORK=true to use WN on private network PRIVATE_NETWORK=false # INFN-GRID: If PRIVATE_NETWORK=true, uncomment and write the internal domain name #MY_INT_DOMAIN=your_internal_domain # INFN-GRID: If PRIVATE_NETWORK=true, uncomment and write your internal network #INT_NET=your_intarnal_network # INFN-GRID: If PRIVATE_NETWORK=true, uncomment and write the internal FQDN hostname of CE #CE_INT_HOST=your_internal_ce_host.$MY_INT_DOMAIN # INFN-GRID: If PRIVATE_NETWORK=true, uncomment and write the internal hostname # of the host exporting the directory used to install the application software #INT_HOST_SW_DIR=$CE_INT_HOST ############################## # CE configuration variables # ############################## # Hostname of the CE CE_HOST=cream-01.$MY_DOMAIN ############################ # SubCluster configuration # ############################ # Name of the processor model as defined by the vendor # for the Worker Nodes in a SubCluster. CE_CPU_MODEL=Opteron # Name of the processor vendor # for the Worker Nodes in a SubCluster CE_CPU_VENDOR=AuthenticAMD # Processor clock speed expressed in MHz # for the Worker Nodes in a SubCluster. CE_CPU_SPEED=2589 # For the following variables please check: # https://wiki.egi.eu/wiki/Operations/HOWTO05 # # Operating system name used on the Worker Nodes # part of the SubCluster. CE_OS="ScientificSL" # Operating system release used on the Worker Nodes # part of the SubCluster. CE_OS_RELEASE="5.5" # Operating system version used on the Worker Nodes # part of the SubCluster. CE_OS_VERSION="Boron" # Platform Type of the WN in the SubCluster # Check: https://wiki.egi.eu/wiki/Operations/HOWTO06 CE_OS_ARCH=x86_64 # Total physical memory of a WN in the SubCluster # expressed in Megabytes. CE_MINPHYSMEM=4096 # Total virtual memory of a WN in the SubCluster # expressed in Megabytes. CE_MINVIRTMEM=1024 # Total number of real CPUs/physical chips in # the SubCluster, including the nodes part of the # SubCluster that are temporary down or offline. CE_PHYSCPU=4 # Total number of cores/hyperthreaded CPUs in # the SubCluster, including the nodes part of the # SubCluster that are temporary down or offline CE_LOGCPU=16 # Number of Logical CPUs (cores) of the WN in the # SubCluster CE_SMPSIZE=8 # Performance index of your fabric in SpecInt 2000 CE_SI00=1039 # Performance index of your fabric in SpecFloat 2000 CE_SF00=951 # Set this variable to either TRUE or FALSE to express # the permission for direct outbound connectivity # for the WNs in the SubCluster CE_OUTBOUNDIP=TRUE # Set this variable to either TRUE or FALSE to express # the permission for inbound connectivity # for the WNs in the SubCluster CE_INBOUNDIP=FALSE # Space separated list of software tags supported by the site # e.g. CE_RUNTIMEENV="LCG-2 LCG-2_1_0 LCG-2_1_1 LCG-2_2_0 GLITE-3_0_0 GLITE-3_1_0 R-GMA" # INFN-GRID: Add the following tags to runtime environment: # * If your site belongs to INFN write CITY in capitals (ex. PADOVA) # otherwise INSTITUTE-CITY in capitals (ex. SNS-PISA) # * Write the average value of SpecInt2000 and SpecFloat2000 for your WNs; # please note that now a '_' is used as separator in place of '=' # (see at http://grid-it.cnaf.infn.it/fileadmin/Certification/MetricHowTo.pdf) # SI00MeanPerCPU_<your_value> # SF00MeanPerCPU_<your_value> #CE_RUNTIMEENV="tag1 [tag2 [...]]" CE_RUNTIMEENV=" IGI-BOLOGNA SI00MeanPerCPU_1039 SF00MeanPerCPU_951 GLITE-3_0_0 GLITE-3_1_0 GLITE-3_2_0 " # For the following variables, please check more detailed information in: # https://twiki.cern.ch/twiki/bin/view/LCG/Site-info_configuration_variables#site_info_def # # The following values must be defined by the sys admin: # - CPUScalingReferenceSI00=<referenceCPU-SI00> # - Share=<vo-name>:<vo-share> (optional, multiple definitons) #CE_CAPABILITY="CPUScalingReferenceSI00=value [Share=vo-name1:value [Share=vo-name2:value [...]]]" CE_CAPABILITY="CPUScalingReferenceSI00=1039" # The following values must be defined by the sys admin: # - Cores=value # - value-HEP-SPEC06 (optional), where value is the CPU power computed # using the HEP-SPEC06 benchmark #CE_OTHERDESCR="Cores=value[,Benchmark=value-HEP-SPEC06]" CE_OTHERDESCR="Cores=1,Benchmark=4.156-HEP-SPEC06" ######################################## # Batch server configuration variables # ######################################## #Set it to 'no' if you want to disable the maui configuration in YAIM #Default value: yes CONFIG_MAUI="yes" #Set it to 'no' if you want to disable the /var/torque/server_priv/nodes configuration in YAIM #Default value: yes CONFIG_TORQUE_NODES="yes" # Hostname of the Batch server # Change this if your batch server is not installed # in the same host of the CE # INFN-GRID: Set this variable to the hostname of your batch master server, that # that could be different from current CE (for example in the case of # sites with two CEs). Do not comment it! BATCH_SERVER=batch.cnaf.infn.it # Jobmanager specific settings. Please, define: # lcgpbs, lcglsf, lcgsge or lcgcondor # INFN-GRID: for CREAM CE write it without 'lcg' prefix: pbs, lsf, sge, condor JOB_MANAGER=pbs # torque, lsf, sge or condor # INFN-GRID: write here 'pbs' even if you are using torque (for compatibility # with MPI support on Globus), write 'lsf' if you using LSF. CE_BATCH_SYS=pbs # INFN-GRID: only for torque/pbs; path for batch-system log files # BATCH_LOG_DIR=/var/torque/ BATCH_LOG_DIR=/var/torque/ BATCH_VERSION=2.5.7 # INFN-GRID: only for LSF; path for the batch-system bin files #BATCH_BIN_DIR=my_batch_system_bin_dir # INFN-GRID: only for LSF; path where lsf.conf is located #BATCH_CONF_DIR=lsf_install_path/conf #Path of a file containing the munge key. Munge is required since Torque version 2.5.7. This file will be copied to /etc/munge/munge.key MUNGE_KEY_FILE=/etc/munge/munge.key ################################ # APEL configuration variables # ################################ # Database password for the APEL DB. APEL_DB_PASSWORD=segreto MON_HOST=$CE_HOST ######################## # SE general variables # ######################## STORM_HOST=darkstorm.$MY_DOMAIN # Space separated list of SEs hostnames SE_LIST="$STORM_HOST" # Space separated list of SE hosts from SE_LIST containing # the export directory from the Storage Element and the # mount directory common to the worker nodes that are part # of the Computing Element. If any of the SEs in SE_LIST # does not support the mount concept, do not define # anything for that SE in this variable. If this is the case # for all the SEs in SE_LIST then put the value "none" #SE_MOUNT_INFO_LIST="[SE1:export_dir1,mount_dir1 [SE2:export_dir2,mount_dir2 [...]]|none]" SE_MOUNT_INFO_LIST="none" ################################ # BDII configuration variables # ################################ # Hostname of the top level BDII # INFN-GRID: Default BDII Top for Italy (if you do not have your own) BDII_HOST=egee-bdii.cnaf.infn.it ############################## # VO configuration variables # ############################## # If you are configuring a DNS-like VO, please check # the following URL: https://twiki.cern.ch/twiki/bin/view/LCG/YaimGuide400#vo_d_directory # Space separated list of VOs supported by your site # INFN-GRID: Available VOs (alphabetically sorted - count 44): # alice ams02.cern.ch argo atlas babar # bio biomed cdf cms compassit compchem cyclops # comput-er.it d4science.research-infrastructures.eu dteam enea # enmr.eu esr eticsproject.eu euasia euchina euindia # eumed gilda geant4 gear glast.org # gridit inaf infngrid ingv lhcb # libi lights.infn.it magic omiieurope ops # pacs.infn.it pamela planck superbvo.org theophys # tps.infn.it virgo zeus # INFN-GRID: Your site must support the following certification VOs: # dteam infngrid ops #VOS="vo1 [vo2 [...]]" VOS="comput-er.it dteam igi.italiangrid.it infngrid ops gridit" # Prefix of the experiment software directory in your CE # INFN-GRID: Agreement on software directory name VO_SW_DIR=/opt/exp_soft # Space separated list of queues configured in your CE # INFN-GRID: Your site must configure a queue that support certification VOs #QUEUES="q1 [q2 [...]]" QUEUES="cert prod" # For each queue defined in QUEUES, define a _GROUP_ENABLE variable # which is a space separated list of VO names and VOMS FQANs: # Ex.: MYQUEUE_GROUP_ENABLE="ops atlas cms /cms/Higgs /cms/ROLE=production" # In QUEUE names containing dots and dashes replace them with underscore: # Ex.: QUEUES="my.test-queue" # MY_TEST_QUEUE_GROUP_ENABLE="ops atlas" # INFN-GRID: Here an example of the settings for certification queue: # CERT_GROUP_ENABLE="dteam infngrid ops" #<queue-name>_GROUP_ENABLE="fqan1 [fqan2 [...]]" CERT_GROUP_ENABLE="dteam infngrid ops /dteam/ROLE=lcgadmin /dteam/ROLE=production /ops/ROLE=lcgadmin /ops/ROLE=pilot /infngrid/ROLE=SoftwareManager /infngrid/ROLE=pilot" PROD_GROUP_ENABLE="comput-er.it gridit igi.italiangrid.it /comput-er.it/ROLE=SoftwareManager /gridit/ROLE=SoftwareManager /igi.italiangrid.it/ROLE=SoftwareManager" #CLOUDTF_GROUP_ENABLE="dteam /dteam/ROLE=lcgadmin /dteam/ROLE=production enmr.eu /enmr.eu/ROLE=SoftwareManager" # Optional variable to define the default SE used by the VO. # Define the SE hostname if you want a specific SE to be the default one. # If this variable is not defined, the first SE in SE_LIST will be used # as the default one. # VO_<vo_name>_DEFAULT_SE=vo-default-se # Optional variable to define a list of LBs used by the VO. # Define a space separated list of LB hostnames. # If this variable is not defined LB_HOST will be used. # VO_<vo_name>_LB_HOSTS="vo-lb1 [vo-lb2 [...]]" # Optional variable to automatically add wildcards per FQAN # in the LCMAPS gripmap file and groupmap file. Set it to 'yes' # if you want to add the wildcards in your VO. Do not define it # or set it to 'no' if you do not want to configure wildcards in your VO. # VO_<vo_name>_MAP_WILDCARDS=no # Optional variable to define the Myproxy server supported by the VO. # Define the Myproxy hostname if you want a specific Myproxy server. # If this variable is not defined PX_HOST will be used. # VO_<vo_name>_PX_HOST=vo-myproxy # Optional variable to define a list of RBs used by the VO. # Define a space separated list of RB hostnames. # If this variable is not defined RB_HOST will be used. # VO_<vo_name>_RBS="vo-rb1 [vo-rb2 [...]]" # Area on the WN for the installation of the experiment software. # If on the WNs a predefined shared area has been mounted where # VO managers can pre-install software, then these variable # should point to this area. If instead there is not a shared # area and each job must install the software, then this variables # should contain a dot ( . ). Anyway the mounting of shared areas, # as well as the local installation of VO software is not managed # by yaim and should be handled locally by Site Administrators. # VO_<vo_name>_SW_DIR=wn_exp_soft_dir # This variable contains the vomses file parameters needed # to contact a VOMS server. Multiple VOMS servers can be given # if the parameters are enclosed in single quotes. # VO_<vo_name>_VOMSES="'vo_name voms_server_hostname port voms_server_host_cert_dn vo_name' ['...']" # DN of the CA that signs the VOMS server certificate. # Multiple values can be given if enclosed in single quotes. # Note that there must be as many entries as in the VO_<vo-name>_VOMSES variable. # There is a one to one relationship in the elements of both lists, # so the order must be respected # VO_<vo_name>_VOMS_CA_DN="'voms_server_ca_dn' ['...']" # A list of the VOMS servers used to create the DN grid-map file. # Multiple values can be given if enclosed in single quotes. # VO_<vo_name>_VOMS_SERVERS="'vomss://<host-name>:8443/voms/<vo-name>?/<vo-name>' ['...']" # Optional variable to define a list of WMSs used by the VO. # Define a space separated list of WMS hostnames. # If this variable is not defined WMS_HOST will be used. # VO_<vo_name>_WMS_HOSTS="vo-wms1 [vo-wms2 [...]]" # Optional variable to create a grid-mapfile with mappings to ordinary # pool accounts, not containing mappings to special users. # - UNPRIVILEGED_MKGRIDMAP=no or undefined, will contain # special users if defined in groups.conf # - UNPRIVILEGED_MKGRIDMAP=yes, will create a grid-mapfile # containing only mappings to ordinary pool accounts. # VO_<vo_name>_UNPRIVILEGED_MKGRIDMAP=no # gLite pool account home directory for the user accounts specified in USERS_CONF. # Define this variable if you would like to use a directory different than /home. # VO_<vo_name>_USER_HOME_PREFIX=/pool_account_home_dir # Examples for the following VOs are included below: # # atlas # alice # lhcb # cms # dteam # biomed # ops # # VOs should check the CIC portal http://cic.in2p3.fr for the VO ID card information # # # INFN-GRID: aux variable: default SE per VO #SE_HOST=my-close-se.${MY_DOMAIN} SE_HOST=$STORM_HOST ######### # atlas # ######### VO_ATLAS_SW_DIR=$VO_SW_DIR/atlas VO_ATLAS_DEFAULT_SE=$SE_HOST VO_ATLAS_STORAGE_DIR=$CLASSIC_STORAGE_DIR/atlas VO_ATLAS_VOMS_SERVERS='vomss://voms.cern.ch:8443/voms/atlas?/atlas/' # INFN-GRID: added BNL replica VO_ATLAS_VOMSES="'atlas lcg-voms.cern.ch 15001 /DC=ch/DC=cern/OU=computers/CN=lcg-voms.cern.ch atlas 24' 'atlas voms.cern.ch 15001 /DC=ch/DC=cern/OU=computers/CN=voms.cern.ch atlas 24' 'atlas vo.racf.bnl.gov 15003 /DC=org/DC=doegrids/OU=Services/CN=vo.racf.bnl.gov atlas 24'" # INFN-GRID: added BNL replica VO_ATLAS_VOMS_CA_DN="'/DC=ch/DC=cern/CN=CERN Trusted Certification Authority' '/DC=ch/DC=cern/CN=CERN Trusted Certification Authority' '/DC=org/DC=DOEGrids/OU=Certificate Authorities/CN=DOEGrids CA 1'" VO_ATLAS_RBS="atlasrb1.cern.ch atlasrb2.cern.ch" ########## # alice # ########## VO_ALICE_SW_DIR=$VO_SW_DIR/alice VO_ALICE_DEFAULT_SE=$SE_HOST VO_ALICE_STORAGE_DIR=$CLASSIC_STORAGE_DIR/alice VO_ALICE_VOMS_SERVERS='vomss://voms.cern.ch:8443/voms/alice?/alice/' VO_ALICE_VOMSES="'alice lcg-voms.cern.ch 15000 /DC=ch/DC=cern/OU=computers/CN=lcg-voms.cern.ch alice 24' 'alice voms.cern.ch 15000 /DC=ch/DC=cern/OU=computers/CN=voms.cern.ch alice 24'" VO_ALICE_VOMS_CA_DN="'/DC=ch/DC=cern/CN=CERN Trusted Certification Authority' '/DC=ch/DC=cern/CN=CERN Trusted Certification Authority'" ####### # cms # ####### VO_CMS_SW_DIR=$VO_SW_DIR/cms VO_CMS_DEFAULT_SE=$SE_HOST VO_CMS_STORAGE_DIR=$CLASSIC_STORAGE_DIR/cms VO_CMS_VOMS_SERVERS='vomss://voms.cern.ch:8443/voms/cms?/cms/' VO_CMS_VOMSES="'cms lcg-voms.cern.ch 15002 /DC=ch/DC=cern/OU=computers/CN=lcg-voms.cern.ch cms 24' 'cms voms.cern.ch 15002 /DC=ch/DC=cern/OU=computers/CN=voms.cern.ch cms 24'" VO_CMS_VOMS_CA_DN="'/DC=ch/DC=cern/CN=CERN Trusted Certification Authority' '/DC=ch/DC=cern/CN=CERN Trusted Certification Authority'" ######## # lhcb # ######## VO_LHCB_SW_DIR=$VO_SW_DIR/lhcb VO_LHCB_DEFAULT_SE=$SE_HOST VO_LHCB_STORAGE_DIR=$CLASSIC_STORAGE_DIR/lhcb VO_LHCB_VOMS_SERVERS='vomss://voms.cern.ch:8443/voms/lhcb?/lhcb/' VO_LHCB_VOMSES="'lhcb lcg-voms.cern.ch 15003 /DC=ch/DC=cern/OU=computers/CN=lcg-voms.cern.ch lhcb 24' 'lhcb voms.cern.ch 15003 /DC=ch/DC=cern/OU=computers/CN=voms.cern.ch lhcb 24'" VO_LHCB_VOMS_CA_DN="'/DC=ch/DC=cern/CN=CERN Trusted Certification Authority' '/DC=ch/DC=cern/CN=CERN Trusted Certification Authority'" ######### # dteam # ######### VO_DTEAM_SW_DIR=$VO_SW_DIR/dteam VO_DTEAM_DEFAULT_SE=$SE_HOST VO_DTEAM_STORAGE_DIR=$CLASSIC_STORAGE_DIR/dteam VO_DTEAM_VOMS_SERVERS='vomss://voms.hellasgrid.gr:8443/voms/dteam?/dteam/' VO_DTEAM_VOMSES="'dteam lcg-voms.cern.ch 15004 /DC=ch/DC=cern/OU=computers/CN=lcg-voms.cern.ch dteam 24' 'dteam voms.cern.ch 15004 /DC=ch/DC=cern/OU=computers/CN=voms.cern.ch dteam 24' 'dteam voms.hellasgrid.gr 15004 /C=GR/O=HellasGrid/OU=hellasgrid.gr/CN=voms.hellasgrid.gr dteam 24' 'dteam voms2.hellasgrid.gr 15004 /C=GR/O=HellasGrid/OU=hellasgrid.gr/CN=voms2.hellasgrid.gr dteam 24'" VO_DTEAM_VOMS_CA_DN="'/DC=ch/DC=cern/CN=CERN Trusted Certification Authority' '/DC=ch/DC=cern/CN=CERN Trusted Certification Authority' '/C=GR/O=HellasGrid/OU=Certification Authorities/CN=HellasGrid CA 2006' '/C=GR/O=HellasGrid/OU=Certification Authorities/CN=HellasGrid CA 2006'" ########## # biomed # ########## VO_BIOMED_SW_DIR=$VO_SW_DIR/biomed VO_BIOMED_DEFAULT_SE=$SE_HOST VO_BIOMED_STORAGE_DIR=$CLASSIC_STORAGE_DIR/biomed VO_BIOMED_VOMS_SERVERS="vomss://cclcgvomsli01.in2p3.fr:8443/voms/biomed?/biomed/" VO_BIOMED_VOMSES="'biomed cclcgvomsli01.in2p3.fr 15000 /O=GRID-FR/C=FR/O=CNRS/OU=CC-LYON/CN=cclcgvomsli01.in2p3.fr biomed 24'" # INFN-GRID: Fix CA DN VO_BIOMED_VOMS_CA_DN="'/C=FR/O=CNRS/CN=GRID2-FR'" ####### # ops # ####### VO_OPS_SW_DIR=$VO_SW_DIR/ops VO_OPS_DEFAULT_SE=$SE_HOST VO_OPS_STORAGE_DIR=$CLASSIC_STORAGE_DIR/ops VO_OPS_VOMS_SERVERS="vomss://voms.cern.ch:8443/voms/ops?/ops/" VO_OPS_VOMSES="'ops lcg-voms.cern.ch 15009 /DC=ch/DC=cern/OU=computers/CN=lcg-voms.cern.ch ops 24' 'ops voms.cern.ch 15009 /DC=ch/DC=cern/OU=computers/CN=voms.cern.ch ops 24'" VO_OPS_VOMS_CA_DN="'/DC=ch/DC=cern/CN=CERN Trusted Certification Authority' '/DC=ch/DC=cern/CN=CERN Trusted Certification Authority'" #====================== # INFN-GRID: Added VOs #====================== ######## # argo # ######## VO_ARGO_SW_DIR=$VO_SW_DIR/argo VO_ARGO_DEFAULT_SE=$SE_HOST VO_ARGO_STORAGE_DIR=$CLASSIC_STORAGE_DIR/argo VO_ARGO_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/argo?/argo' 'vomss://voms-01.pd.infn.it:8443/voms/argo?/argo'" VO_ARGO_VOMSES="'argo voms.cnaf.infn.it 15012 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it argo' 'argo voms-01.pd.infn.it 15012 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it argo'" VO_ARGO_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ######### # babar # ######### VO_BABAR_SW_DIR=$VO_SW_DIR/babar VO_BABAR_DEFAULT_SE=$SE_HOST VO_BABAR_STORAGE_DIR=$CLASSIC_STORAGE_DIR/babar VO_BABAR_VOMS_SERVERS="'vomss://voms.gridpp.ac.uk:8443/voms/babar?/babar'" VO_BABAR_VOMSES="'babar voms.gridpp.ac.uk 15002 /C=UK/O=eScience/OU=Manchester/L=HEP/CN=voms.gridpp.ac.uk/emailAddress=ops@tier2.hep.manchester.ac.uk babar'" VO_BABAR_VOMS_CA_DN="'/C=UK/O=eScienceCA/OU=Authority/CN=UK e-Science CA'" ####### # bio # ####### VO_BIO_SW_DIR=$VO_SW_DIR/bio VO_BIO_DEFAULT_SE=$SE_HOST VO_BIO_STORAGE_DIR=$CLASSIC_STORAGE_DIR/bio VO_BIO_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/bio?/bio' 'vomss://voms-01.pd.infn.it:8443/voms/bio?/bio'" VO_BIO_VOMSES="'bio voms.cnaf.infn.it 15007 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it bio' 'bio voms-01.pd.infn.it 15007 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it bio'" VO_BIO_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ####### # cdf # ####### VO_CDF_SW_DIR=$VO_SW_DIR/cdf VO_CDF_DEFAULT_SE=$SE_HOST VO_CDF_STORAGE_DIR=$CLASSIC_STORAGE_DIR/cdf VO_CDF_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/cdf?/cdf' 'vomss://voms-01.pd.infn.it:8443/voms/cdf?/cdf'" VO_CDF_VOMSES="'cdf voms.cnaf.infn.it 15001 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it cdf' 'cdf voms-01.pd.infn.it 15001 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it cdf' 'cdf voms.fnal.gov 15020 /DC=org/DC=doegrids/OU=Services/CN=http/voms.fnal.gov cdf'" VO_CDF_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA' '/DC=org/DC=DOEGrids/OU=Certificate Authorities/CN=DOEGrids CA 1'" ############# # compassit # ############# VO_COMPASSIT_SW_DIR=$VO_SW_DIR/compassit VO_COMPASSIT_DEFAULT_SE=$SE_HOST VO_COMPASSIT_STORAGE_DIR=$CLASSIC_STORAGE_DIR/compassit VO_COMPASSIT_VOMS_SERVERS="'vomss://voms2.cnaf.infn.it:8443/voms/compassit?/compassit' 'vomss://voms-02.pd.infn.it:8443/voms/compassit?/compassit'" VO_COMPASSIT_VOMSES="'compassit voms2.cnaf.infn.it 15012 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms2.cnaf.infn.it compassit' 'compassit voms-02.pd.infn.it 15012 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-02.pd.infn.it compassit'" VO_COMPASSIT_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ############ # compchem # ############ VO_COMPCHEM_SW_DIR=$VO_SW_DIR/compchem VO_COMPCHEM_DEFAULT_SE=$SE_HOST VO_COMPCHEM_STORAGE_DIR=$CLASSIC_STORAGE_DIR/compchem VO_COMPCHEM_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/compchem?/compchem' 'vomss://voms-01.pd.infn.it:8443/voms/compchem?/compchem'" VO_COMPCHEM_VOMSES="'compchem voms.cnaf.infn.it 15003 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it compchem' 'compchem voms-01.pd.infn.it 15003 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it compchem'" VO_COMPCHEM_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ########### # cyclops # ########### VO_CYCLOPS_SW_DIR=$VO_SW_DIR/cyclops VO_CYCLOPS_DEFAULT_SE=$SE_HOST VO_CYCLOPS_STORAGE_DIR=$CLASSIC_STORAGE_DIR/cyclops VO_CYCLOPS_VOMS_SERVERS="'vomss://voms2.cnaf.infn.it:8443/voms/cyclops?/cyclops' 'vomss://voms-02.pd.infn.it:8443/voms/cyclops?/cyclops'" VO_CYCLOPS_VOMSES="'cyclops voms2.cnaf.infn.it 15011 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms2.cnaf.infn.it cyclops' 'cyclops voms-02.pd.infn.it 15011 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-02.pd.infn.it cyclops'" VO_CYCLOPS_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ######## # enea # ######## VO_ENEA_SW_DIR=$VO_SW_DIR/enea VO_ENEA_DEFAULT_SE=$SE_HOST VO_ENEA_STORAGE_DIR=$CLASSIC_STORAGE_DIR/enea VO_ENEA_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/enea?/enea' 'vomss://voms-01.pd.infn.it:8443/voms/enea?/enea'" VO_ENEA_VOMSES="'enea voms.cnaf.infn.it 15005 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it enea' 'enea voms-01.pd.infn.it 15005 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it enea'" VO_ENEA_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ####### # esr # ####### VO_ESR_SW_DIR=$VO_SW_DIR/esr VO_ESR_DEFAULT_SE=$SE_HOST VO_ESR_STORAGE_DIR=$CLASSIC_STORAGE_DIR/esr VO_ESR_VOMS_SERVERS="'vomss://voms.grid.sara.nl:8443/voms/esr?/esr/'" VO_ESR_VOMSES="'esr voms.grid.sara.nl 30001 /O=dutchgrid/O=hosts/OU=sara.nl/CN=voms.grid.sara.nl esr'" VO_ESR_VOMS_CA_DN="'/C=NL/O=NIKHEF/CN=NIKHEF medium-security certification auth'" ######### # euasia# ######### VO_EUASIA_SW_DIR=$VO_SW_DIR/euasia VO_EUASIA_DEFAULT_SE=$SE_HOST #VO_EUASIA_STORAGE_DIR=$CLASSIC_STORAGE_DIR/euasia VO_EUASIA_VOMS_SERVERS="'vomss://voms.grid.sinica.edu.tw:8443/voms/euasia?/euasia'" VO_EUASIA_VOMSES="'euasia voms.grid.sinica.edu.tw 15015 /C=TW/O=AS/OU=GRID/CN=voms.grid.sinica.edu.tw euasia'" VO_EUASIA_VOMS_CA_DN="'/C=TW/O=AS/CN=Academia Sinica Grid Computing Certification Authority Mercury'" ########### # euchina # ########### VO_EUCHINA_SW_DIR=$VO_SW_DIR/euchina VO_EUCHINA_DEFAULT_SE=$SE_HOST VO_EUCHINA_STORAGE_DIR=$CLASSIC_STORAGE_DIR/euchina VO_EUCHINA_VOMS_SERVERS="'vomss://voms2.cnaf.infn.it:8443/voms/euchina?/euchina' 'vomss://voms-02.pd.infn.it:8443/voms/euchina?/euchina'" VO_EUCHINA_VOMSES="'euchina voms2.cnaf.infn.it 15017 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms2.cnaf.infn.it euchina' 'euchina voms-02.pd.infn.it 15017 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-02.pd.infn.it euchina'" VO_EUCHINA_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ########### # euindia # ########### VO_EUINDIA_SW_DIR=$VO_SW_DIR/euindia VO_EUINDIA_DEFAULT_SE=$SE_HOST VO_EUINDIA_STORAGE_DIR=$CLASSIC_STORAGE_DIR/euindia VO_EUINDIA_VOMS_SERVERS="'vomss://voms2.cnaf.infn.it:8443/voms/euindia?/euindia' 'vomss://voms-02.pd.infn.it:8443/voms/euindia?/euindia'" VO_EUINDIA_VOMSES="'euindia voms2.cnaf.infn.it 15010 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms2.cnaf.infn.it euindia' 'euindia voms-02.pd.infn.it 15010 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-02.pd.infn.it euindia'" VO_EUINDIA_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ######### # eumed # ######### VO_EUMED_SW_DIR=$VO_SW_DIR/eumed VO_EUMED_DEFAULT_SE=$SE_HOST VO_EUMED_STORAGE_DIR=$CLASSIC_STORAGE_DIR/eumed VO_EUMED_VOMS_SERVERS="'vomss://voms2.cnaf.infn.it:8443/voms/eumed?/eumed' 'vomss://voms-02.pd.infn.it:8443/voms/eumed?/eumed'" VO_EUMED_VOMSES="'eumed voms2.cnaf.infn.it 15016 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms2.cnaf.infn.it eumed' 'eumed voms-02.pd.infn.it 15016 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-02.pd.infn.it eumed'" VO_EUMED_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ########## # geant4 # ########## VO_GEANT4_SW_DIR=$VO_SW_DIR/geant4 VO_GEANT4_DEFAULT_SE=$SE_HOST VO_GEANT4_STORAGE_DIR=$CLASSIC_STORAGE_DIR/geant4 VO_GEANT4_VOMS_SERVERS="'vomss://voms.cern.ch:8443/voms/geant4?/geant4'" VO_GEANT4_VOMSES="'geant4 voms.cern.ch 15007 /DC=ch/DC=cern/OU=computers/CN=voms.cern.ch geant4'" VO_GEANT4_VOMS_CA_DN="'/DC=ch/DC=cern/CN=CERN Trusted Certification Authority'" ######## # gear # ######## VO_GEAR_SW_DIR=$VO_SW_DIR/gear VO_GEAR_DEFAULT_SE=$SE_HOST VO_GEAR_STORAGE_DIR=$CLASSIC_STORAGE_DIR/gear VO_GEAR_VOMS_SERVERS="'vomss://voms.cern.ch:8443/voms/geant4?/geant4/gear'" VO_GEAR_VOMS_CA_DN="'/DC=ch/DC=cern/CN=CERN Trusted Certification Authority'" ######### # gilda # ######### # GILDA VO: contact Gilda Team for Gilda settings ########## # gridit # ########## VO_GRIDIT_SW_DIR=$VO_SW_DIR/gridit VO_GRIDIT_DEFAULT_SE=$SE_HOST VO_GRIDIT_STORAGE_DIR=$CLASSIC_STORAGE_DIR/gridit VO_GRIDIT_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/gridit?/gridit' 'vomss://voms-01.pd.infn.it:8443/voms/gridit?/gridit'" VO_GRIDIT_VOMSES="'gridit voms.cnaf.infn.it 15008 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it gridit' 'gridit voms-01.pd.infn.it 15008 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it gridit'" VO_GRIDIT_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ######## # inaf # ######## VO_INAF_SW_DIR=$VO_SW_DIR/inaf VO_INAF_DEFAULT_SE=$SE_HOST VO_INAF_STORAGE_DIR=$CLASSIC_STORAGE_DIR/inaf VO_INAF_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/inaf?/inaf' 'vomss://voms-01.pd.infn.it:8443/voms/inaf?/inaf'" VO_INAF_VOMSES="'inaf voms.cnaf.infn.it 15010 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it inaf' 'inaf voms-01.pd.infn.it 15010 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it inaf'" VO_INAF_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ############ # infngrid # ############ VO_INFNGRID_SW_DIR=$VO_SW_DIR/infngrid VO_INFNGRID_DEFAULT_SE=$SE_HOST VO_INFNGRID_STORAGE_DIR=$CLASSIC_STORAGE_DIR/infngrid VO_INFNGRID_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/infngrid?/infngrid' 'vomss://voms-01.pd.infn.it:8443/voms/infngrid?/infngrid'" VO_INFNGRID_VOMSES="'infngrid voms.cnaf.infn.it 15000 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it infngrid' 'infngrid voms-01.pd.infn.it 15000 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it infngrid'" VO_INFNGRID_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ######## # ingv # ######## VO_INGV_SW_DIR=$VO_SW_DIR/ingv VO_INGV_DEFAULT_SE=$SE_HOST VO_INGV_STORAGE_DIR=$CLASSIC_STORAGE_DIR/ingv VO_INGV_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/ingv?/ingv' 'vomss://voms-01.pd.infn.it:8443/voms/ingv?/ingv'" VO_INGV_VOMSES="'ingv voms.cnaf.infn.it 15011 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it ingv' 'ingv voms-01.pd.infn.it 15011 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it ingv'" VO_INGV_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ######## # libi # ######## VO_LIBI_SW_DIR=$VO_SW_DIR/libi VO_LIBI_DEFAULT_SE=$SE_HOST VO_LIBI_STORAGE_DIR=$CLASSIC_STORAGE_DIR/libi VO_LIBI_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/libi?/libi' 'vomss://voms-01.pd.infn.it:8443/voms/libi?/libi'" VO_LIBI_VOMSES="'libi voms.cnaf.infn.it 15015 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it libi' 'libi voms-01.pd.infn.it 15015 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it libi'" VO_LIBI_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ######### # magic # ######### VO_MAGIC_SW_DIR=$VO_SW_DIR/magic VO_MAGIC_DEFAULT_SE=$SE_HOST VO_MAGIC_STORAGE_DIR=$CLASSIC_STORAGE_DIR/magic VO_MAGIC_VOMS_SERVERS="'vomss://voms01.pic.es:8443/voms/magic?/magic' 'vomss://voms02.pic.es:8443/voms/magic?/magic'" VO_MAGIC_VOMSES="'magic voms01.pic.es 15003 /DC=es/DC=irisgrid/O=pic/CN=voms01.pic.es magic' 'magic voms02.pic.es 15003 /DC=es/DC=irisgrid/O=pic/CN=voms02.pic.es magic'" VO_MAGIC_VOMS_CA_DN="'/DC=es/DC=irisgrid/CN=IRISGridCA' '/DC=es/DC=irisgrid/CN=IRISGridCA'" ######## # na48 # ######## VO_NA48_SW_DIR=$VO_SW_DIR/na48 VO_NA48_DEFAULT_SE=$SE_HOST VO_NA48_STORAGE_DIR=$CLASSIC_STORAGE_DIR/na48 VO_NA48_VOMS_SERVERS="'vomss://lcg-voms.cern.ch:8443/voms/na48?/na48' 'vomss://voms.cern.ch:8443/voms/na48?/na48'" VO_NA48_VOMSES="'na48 lcg-voms.cern.ch 15011 /DC=ch/DC=cern/OU=computers/CN=lcg-voms.cern.ch na48' 'na48 voms.cern.ch 15011 DC=ch/DC=cern/OU=computers/CN=voms.cern.ch na48'" VO_NA48_VOMS_CA_DN="'/DC=ch/DC=cern/CN=CERN Trusted Certification Authority' '/DC=ch/DC=cern/CN=CERN Trusted Certification Authority'" ########## # pamela # ########## VO_PAMELA_SW_DIR=$VO_SW_DIR/pamela VO_PAMELA_DEFAULT_SE=$SE_HOST VO_PAMELA_STORAGE_DIR=$CLASSIC_STORAGE_DIR/pamela VO_PAMELA_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/pamela?/pamela' 'vomss://voms-01.pd.infn.it:8443/voms/pamela?/pamela'" VO_PAMELA_VOMSES="'pamela voms.cnaf.infn.it 15013 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it pamela' 'pamela voms-01.pd.infn.it 15013 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it pamela'" VO_PAMELA_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ########## # planck # ########## VO_PLANCK_SW_DIR=$VO_SW_DIR/planck VO_PLANCK_DEFAULT_SE=$SE_HOST VO_PLANCK_STORAGE_DIR=$CLASSIC_STORAGE_DIR/planck VO_PLANCK_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/planck?/planck' 'vomss://voms-01.pd.infn.it:8443/voms/planck?/planck'" VO_PLANCK_VOMSES="'planck voms.cnaf.infn.it 15002 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it planck' 'planck voms-01.pd.infn.it 15002 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it planck'" VO_PLANCK_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ######## # sixt # ######## VO_SIXT_SW_DIR=$VO_SW_DIR/sixt VO_SIXT_DEFAULT_SE=$SE_HOST VO_SIXT_STORAGE_DIR=$CLASSIC_STORAGE_DIR/sixt VO_SIXT_VOMS_SERVERS="'vomss://voms.cern.ch:8443/voms/sixt?/sixt'" VO_SIXT_VOMSES="'sixt voms.cern.ch 15005 /DC=ch/DC=cern/OU=computers/CN=voms.cern.ch sixt'" VO_SIXT_VOMS_CA_DN="'/DC=ch/DC=cern/CN=CERN Trusted Certification Authority'" ############ # theophys # ############ VO_THEOPHYS_SW_DIR=$VO_SW_DIR/theophys VO_THEOPHYS_DEFAULT_SE=$SE_HOST VO_THEOPHYS_STORAGE_DIR=$CLASSIC_STORAGE_DIR/theophys VO_THEOPHYS_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/theophys?/theophys' 'vomss://voms-01.pd.infn.it:8443/voms/theophys?/theophys'" VO_THEOPHYS_VOMSES="'theophys voms.cnaf.infn.it 15006 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it theophys' 'theophys voms-01.pd.infn.it 15006 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it theophys'" VO_THEOPHYS_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ######### # virgo # ######### VO_VIRGO_SW_DIR=$VO_SW_DIR/virgo VO_VIRGO_DEFAULT_SE=$SE_HOST VO_VIRGO_STORAGE_DIR=$CLASSIC_STORAGE_DIR/virgo VO_VIRGO_VOMS_SERVERS="'vomss://voms.cnaf.infn.it:8443/voms/virgo?/virgo' 'vomss://voms-01.pd.infn.it:8443/voms/virgo?/virgo'" VO_VIRGO_VOMSES="'virgo voms.cnaf.infn.it 15009 /C=IT/O=INFN/OU=Host/L=CNAF/CN=voms.cnaf.infn.it virgo' 'virgo voms-01.pd.infn.it 15009 /C=IT/O=INFN/OU=Host/L=Padova/CN=voms-01.pd.infn.it virgo'" VO_VIRGO_VOMS_CA_DN="'/C=IT/O=INFN/CN=INFN CA' '/C=IT/O=INFN/CN=INFN CA'" ######## # zeus # ######## VO_ZEUS_SW_DIR=$VO_SW_DIR/zeus VO_ZEUS_DEFAULT_SE=$SE_HOST VO_ZEUS_STORAGE_DIR=$CLASSIC_STORAGE_DIR/zeus VO_ZEUS_VOMS_SERVERS="'vomss://grid-voms.desy.de:8443/voms/zeus?/zeus'" VO_ZEUS_VOMSES="'zeus grid-voms.desy.de 15112 /C=DE/O=GermanGrid/OU=DESY/CN=host/grid-voms.desy.de zeus'" VO_ZEUS_VOMS_CA_DN="'/C=DE/O=GermanGrid/CN=GridKa-CA'" </verbatim> -- Main.AlessandroPaolini - 2012-05-25
E
dit
|
A
ttach
|
PDF
|
H
istory
: r2
<
r1
|
B
acklinks
|
V
iew topic
|
M
ore topic actions
Topic revision: r2 - 2012-05-25
-
AlessandroPaolini
Home
Site map
CEMon web
CREAM web
Cloud web
Cyclops web
DGAS web
EgeeJra1It web
Gows web
GridOversight web
IGIPortal web
IGIRelease web
MPI web
Main web
MarcheCloud web
MarcheCloudPilotaCNAF web
Middleware web
Operations web
Sandbox web
Security web
SiteAdminCorner web
TWiki web
Training web
UserSupport web
VOMS web
WMS web
WMSMonitor web
WeNMR web
SiteAdminCorner Web
Create New Topic
Index
Search
Changes
Notifications
RSS Feed
Statistics
Preferences
P
View
Raw View
Print version
Find backlinks
History
More topic actions
Edit
Raw edit
Attach file or image
Edit topic preference settings
Set new parent
More topic actions
Account
Log In
E
dit
A
ttach
Copyright © 2008-2024 by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding TWiki?
Send feedback