Regression Tests

bug #33342: separate retry policies for ISB and OSB Not implemented

Description of the test.

Of course we're speaking of submission to the lcg-CE, as CREAM uses its own jobwrapper.

ISB: https://devel11.cnaf.infn.it:9000/a...

submitted a job and then removed its ISB

[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf lsisb.jdl
Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/-...

==========================================================================

server side:
[root@devel11 input]# rm -f a
[root@devel11 input]# pwd
/var/SandboxDir/-h/https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2f-h4MRDzYkufRu71MKfF1pw/input

after a while, maradona file reports:
[root@devel11 https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2f-h4MRDzYkufRu71MKfF1pw]# cat Maradona.output
LM_log_done_begin
Wed Apr 20 22:08:51 CEST 2011: lcg-jobwrapper-hook.sh not readable or not present
Wed Apr 20 22:08:52 CEST 2011: Error during transfer
Wed Apr 20 22:09:53 CEST 2011: Error during transfer
Wed Apr 20 22:11:54 CEST 2011: Error during transfer

LM_log_done_end
Cannot download a from gsiftp://devel11.cnaf.infn.it:2811/var...
Killing log watchdog (pid=21047)...
jw exit status = 1

OSB:
[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf lsosb.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/3...

==========================================================================

Considered successful if: after more than twenty minutes Maradona file hasn't returned yet and the job is running, meaning that other defaults are in place (the ones previously used for both ISB and OSB)

bug #36292: Not all attributes of a SA/SE coul be used in a gangmatching Not implemented

Description of the test.

Fix certified doing a listmatch with the following expression in the jdl:

Requirements = regexp(".in2p3.fr:2119.*",other.GlueCEUniqueID) && anyMatch(other.storage.CloseSEs,target.GlueSEImplementationVersion=="1.9.5-24");

which returns:

- cclcgceli02.in2p3.fr:2119/jobmanager-bqs-short
- cclcgceli04.in2p3.fr:2119/jobmanager-bqs-short
- cclcgceli06.in2p3.fr:2119/jobmanager-bqs-short
- cclcgceli09.in2p3.fr:2119/jobmanager-bqs-short
- cclcgceli02.in2p3.fr:2119/jobmanager-bqs-medium
- cclcgceli04.in2p3.fr:2119/jobmanager-bqs-long
- cclcgceli06.in2p3.fr:2119/jobmanager-bqs-medium
- cclcgceli09.in2p3.fr:2119/jobmanager-bqs-medium
- cclcgceli02.in2p3.fr:2119/jobmanager-bqs-long
- cclcgceli06.in2p3.fr:2119/jobmanager-bqs-long
- cclcgceli04.in2p3.fr:2119/jobmanager-bqs-medium
- cclcgceli09.in2p3.fr:2119/jobmanager-bqs-long

Double checking that the correct "GlueSEImplementationVersion" is picked up:

lcg-infosites --vo dteam closeSE >closeses.txt

gives the following closeSEs:

cclcgceli02.in2p3.fr:2119/jobmanager-bqs-long
ccsrm.in2p3.fr
ccsrm02.in2p3.fr

and ldapsearch -x -H ldap://lcg-bdii.cern.ch:2170 -b 'Mds-vo-name=local,o=Grid' '(GlueSEUniqueId=ccsrm.in2p3.fr)'
returns:

...
GlueSEImplementationVersion: 1.9.5-24
...

bug #40982: When a collection is aborted the "Abort" event should be logged for the sub-nodes as well /2 Implemented

Description of the test.

coll_10.jdl is a ten nodes collection, only the first node having non empty ISB.

[mcecchi@cert-19 ~]$ head -25 coll_10.jdl
[
Type = "collection";
InputSandbox = {"/home/mcecchi/Test.sh"};
RetryCount = 1;
Requirements = ( random(1.0) < 0.5 );
ShallowRetryCount = 2;
nodes = {
[
JobType = "Normal";
Zippedisb=true;
Executable = "Test.sh";
StdOutput = "test.out";
StdError = "test.err";
InputSandbox = {"a"};
OutputSandbox = {};
],
[
JobType = "Normal";
Executable = "Test.sh";
StdOutput = "test.out";
StdError = "test.err";
OutputSandbox = {};
],
[
JobType = "Normal";

We register the collection:

[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf --register-only coll_10.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully registered to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/M...

==========================================================================

To complete the operation, the following file containing the InputSandbox of the job needs to be transferred:
==========================================================================================================
ISB ZIP file : /tmp/ISBfiles_aoIPOxSR3GFuEcTxqJ6_Mg_0.tar.gz
Destination : gsiftp://devel11.cnaf.infn.it:2811/var...

We do NOT transfer ISB for the first node and start the job.

[mcecchi@cert-19 ~]$ glite-wms-job-submit --start https://devel11.cnaf.infn.it:9000/M...
Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully started to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/M...

==========================================================================

Considered successful if:

[mcecchi@cert-19 ~]$ glite-wms-job-status https://devel11.cnaf.infn.it:9000/MKfaYgbm4wDSKh0iO191rA|grep Aborted|wc -l
11

bug #44599: WMS should consider "MaxTotalJobs" Implemented

Description of the test.

The wmp configuration section has a new entry:

WmsRequirements = ((ShortDeadlineJob =?= TRUE) ? RegExp(".sdj$", other.GlueCEUniqueID) : !RegExp(".sdj$", other.GlueCEUniqueID)) && (other.GlueCEPolicyMaxTotalJobs == 0 || other.GlueCEStateTotalJobs < other.GlueCEPolicyMaxTotalJobs) && (EnableWmsFeedback =?= TRUE ? RegExp("cream", other.GlueCEImplementationName, "i") : true);

WmsRequirements value is appended in && to the user requirements

UI:

[mcecchi@cert-19 ~]$ cat job.jdl
[
Executable = "/bin/true";
Arguments = "";
StdOutput = "out.log";
StdError = "err.log";
OutputSandbox = {};
InputSandbox = {"a"};
rank = - other.GlueCEStateEstimatedResponseTime ;
//Requirements = !RegExp("cream.*", other.GlueCEUniqueID);
Requirements = true;
RetryCount = 0;
ShallowRetryCount = -1;
]

[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf job.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/X...

==========================================================================

WMS, wmproxy.log:21 Apr, 00:16:45 -D- PID: 4336 - "wmp2wm::submit": LB Logged jdl/path: [ OutputSandboxPath = "/var/SandboxDir/Xm/https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2fXmA_5f5p51xWR_5f5N2B8MDxGw/output"; StdOutput = "out.log"; ShallowRetryCount = -1; SignificantAttributes = { "Requirements","Rank","FuzzyRank" }; RetryCount = 0; Executable = "/bin/true"; Type = "job"; LB_sequence_code = "UI=000000:NS=0000000004:WM=000000:BH=0000000000:JSS=000000:LM=000000:LRMS=000000:APP=000000:LBS=000000"; Arguments = ""; AllowZippedISB = true; VirtualOrganisation = "dteam"; JobType = "normal"; DefaultRank = -other.GlueCEStateEstimatedResponseTime; ZippedISB = { "ISBfiles_rmRmJ25HMnGWbGATaFD7cw_0.tar.gz" }; OutputSandbox = { }; InputSandbox = { "gsiftp://devel11.cnaf.infn.it:2811/var..." }; edg_jobid = "https://devel11.cnaf.infn.it:9000/X..."; VOMS_FQAN = "/dteam/Role=NULL/Capability=NULL"; CertificateSubject = "/C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi"; StdError = "err.log"; InputSandboxPath = "/var/SandboxDir/Xm/https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2fXmA_5f5p51xWR_5f5N2B8MDxGw/input"; rank = -other.GlueCEStateEstimatedResponseTime; MyProxyServer = "myproxy.cnaf.infn.it"; X509UserProxy = "/var/SandboxDir/Xm/https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2fXmA_5f5p51xWR_5f5N2B8MDxGw/user.proxy"; InputSandboxDestFileName = { "a" }; requirements = ( ( true ) && ( other.GlueCEStateStatus == "Production" ) ) && ( ( ( ShortDeadlineJob is true ) ? RegExp(".sdj$",other.GlueCEUniqueID) : !RegExp(".sdj$",other.GlueCEUniqueID) ) && ( other.GlueCEPolicyMaxTotalJobs == 0 || other.GlueCEStateTotalJobs < other.GlueCEPolicyMaxTotalJobs ) && ( EnableWmsFeedback is true ? RegExp("cream",other.GlueCEImplementationName,"i") : true ) ); WMPInputSandboxBaseURI = "gsiftp://devel11.cnaf.infn.it:2811/var..." ]

Considered successful if "requirements" becomes:

requirements = ( ( true ) && ( other.GlueCEStateStatus == "Production" ) ) && ( ( ( ShortDeadlineJob is true ) ? RegExp(".sdj$",other.GlueCEUniqueID) : !RegExp(".sdj$",other.GlueCEUniqueID) ) && ( other.GlueCEPolicyMaxTotalJobs == 0 || other.GlueCEStateTotalJobs < other.GlueCEPolicyMaxTotalJobs ) && ( EnableWmsFeedback is true ? RegExp("cream",other.GlueCEImplementationName,"i") : true ) ); 

bug #45883: Optimization of resubmission Implemented

A proper way to test this bug was not found.

bug #48636: job wrapper should log events for truncated files Not implemented

Description of the bug.

WMS:

[root@devel11 SandboxDir]# grep MaxOutp /etc/glite-wms/glite_wms.conf
MaxOutputSandboxSize = 50M;
[root@devel11 SandboxDir]# /etc/init.d/glite-wms-wm restart

UI:

[mcecchi@cert-19 ~]$ head -5 split_odb.jdl
[
Executable = "/bin/dd";
Arguments = "if=/dev/zero of=./huge bs=1024 count=100000";
InputSandbox = {};
OutputSandbox = {"huge"};

WMS, jobwrapper reports:

[root@devel11 SandboxDir]# grep max_out ../jobcontrol/submit/Y6/JobWrapper.https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2fY6mVGB9jweLcWK0ECpcRKA.sh
__max_outputsandbox_size=52428800
max_osb_size=${__max_outputsandbox_size}

Considered successful if on the UI the job status reports:

<...>
Event: UserTag
- Arrived = Thu Apr 21 11:31:25 2011 CEST
- Host = gs-03.uaic.ro
- Level = SYSTEM
- Name = notice
- Priority = synchronous
- Seqcode = UI=000000:NS=0000000004:WM=000005:BH=0000000000:JSS=000003:LM=000000:LRMS=000009:APP=000000:LBS=000000
- Source = LRMS
- Timestamp = Thu Apr 21 11:23:08 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi
- Value = OSB quota exceeded for /home/dteam034/globus-tmp.gs-03.9232.0/https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2fY6mVGB9jweLcWK0ECpcRKA/huge, truncating needed
---
Event: UserTag
- Arrived = Thu Apr 21 11:31:25 2011 CEST
- Host = gs-03.uaic.ro
- Level = SYSTEM
- Name = notice
- Priority = synchronous
- Seqcode = UI=000000:NS=0000000004:WM=000005:BH=0000000000:JSS=000003:LM=000000:LRMS=000011:APP=000000:LBS=000000
- Source = LRMS
- Timestamp = Thu Apr 21 11:24:10 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi
- Value = Truncated last 52428800 bytes for file huge
---
<...>

bug #48640: glite-wms-wmproxy to support graceful command Not implemented

The test is considered successful if the following processes:

root     16366     1  0 13:14 ?        00:00:01 /usr/sbin/httpd -k start -f /etc/glite-wms/glite_wms_wmproxy_httpd.conf
glite    16641 16366  0 13:16 ?        00:00:00 /usr/sbin/httpd -k start -f /etc/glite-wms/glite_wms_wmproxy_httpd.conf
glite    16642 16366  0 13:16 ?        00:00:00 /usr/sbin/httpd -k start -f /etc/glite-wms/glite_wms_wmproxy_httpd.conf
glite    16643 16366  0 13:16 ?        00:00:00 /usr/sbin/httpd -k start -f /etc/glite-wms/glite_wms_wmproxy_httpd.conf
glite    16644 16366  0 13:16 ?        00:00:00 /usr/sbin/httpd -k start -f /etc/glite-wms/glite_wms_wmproxy_httpd.conf
glite    16645 16366  0 13:16 ?        00:00:00 /usr/sbin/httpd -k start -f /etc/glite-wms/glite_wms_wmproxy_httpd.conf
glite    16646 16366  0 13:16 ?        00:00:00 /usr/sbin/httpd -k start -f /etc/glite-wms/glite_wms_wmproxy_httpd.conf

are restarted gracefully after issuing the command:

/etc/init.d/glite-wms-wmproxy graceful

bug #49844: WMProxy does not catch signal 25 Not implemented

Description of the test.

WMS: [root@devel11 wms]# ps aux | grep wmproxy_server glite 18618 5.2 0.4 322316 19184 ? Sl 23:17 0:01 /usr/bin/glite_wms_wmproxy_server

UI: [mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf coll_100_failing.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

WMS: [root@devel11 wms]# kill -25 18618

UI:

================== glite-wms-job-submit Success ==================

The job has been successfully submitted to the WMProxy Your job identifier is:

https://devel11.cnaf.infn.it:9000/3...

======================================================================

Considered successful if the job is submitted.

bug #50009: wmproxy.gacl person record allows anyone to pass Implemented

Description of the test.

WMS gacl file about dteam:

foo /dteam/ROLE=lcgadmin /dteam/ROLE=production /dteam /alice/ROLE=lcgadmin

UI:

[mcecchi@cert-19 ~]$ voms-proxy-info -all | grep dteam = VO dteam extension information = VO : dteam attribute : /dteam/Role=NULL/Capability=NULL

[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf job.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

Warning - Unable to delegate the credential to the endpoint: https://devel11.cnaf.infn.it:7443/g... AuthorizationException: Authorization error: user not authorized

Considered successful if the user is not authorized.

bug #52617: [ yaim-wms ] host{cert,key}.pem in /home/glite Implemented

Description of the test:

WMS:

[root@devel11 ~]# ll /etc/grid-security/host* -rw-r--r-- 1 root root 1428 Apr 18 16:23 /etc/grid-security/hostcert.pem -r-------- 1 root root 887 Apr 18 16:23 /etc/grid-security/hostkey.pem

[root@devel11 ~]# ll /home/glite/ total 0

Considered successful if there are no certificates under /home/glite

bug #53294: WMS 3.2 WMProxy logs are useless below level 6 Not implemented

Test description.

Setting wmp loglevel to 5:

UI:

[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf lsisb.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/r...

WMS:

21 Apr, 00:02:07 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": ================== Incoming Request ==================
21 Apr, 00:02:07 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": Called Operation: jobRegister
21 Apr, 00:02:07 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": Remote Host Name: cert-19.pd.infn.it
21 Apr, 00:02:07 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": Remote CLIENT S DN: /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi/CN=proxy
21 Apr, 00:02:07 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": Remote GRST CRED: VOMS 47918880745312 1303374407 0 /dteam/Role=NULL/Capability=NULL
21 Apr, 00:02:07 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": Service GRST PROXY LIMIT: 6
21 Apr, 00:02:07 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": WMProxy instance serving core request N.: 7
21 Apr, 00:02:08 -I- PID: 5877 - "wmpcoreoperations::jobRegister": Delegation ID: 7DvOGpVUCAkzYjZQLqWzJg
21 Apr, 00:02:09 -I- PID: 5877 - "wmpcommon::getType": JDL Type: job
21 Apr, 00:02:09 -I- PID: 5877 - "wmpcoreoperations::regist JOB": Registering id: https://devel11.cnaf.infn.it:9000/r...

21 Apr, 00:02:13 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": ================== Incoming Request ==================
21 Apr, 00:02:13 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": Called Operation: jobStart
21 Apr, 00:02:13 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": Remote Host Name: cert-19.pd.infn.it
21 Apr, 00:02:13 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": Remote CLIENT S DN: /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi/CN=proxy
21 Apr, 00:02:13 -I- PID: 5877 - "wmpcommon::initWMProxyOperation": Remote GRST CRED: VOMS 47918880745312 1303374407 0 /dteam/Role=NULL/Capability=NULL 

Considered successful if the log entries contain enough information

bug #54728: WMP finds FQAN inconsistency only if GROUPS are different, not ROLES Not implemented

Test description.

[apaolini@ui ~]$ voms-proxy-init -voms dteam
Enter GRID pass phrase:
Your identity: /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Alessandro Paolini
Creating temporary proxy ........................................................ Done
Contacting voms.hellasgrid.gr:15004 [/C=GR/O=HellasGrid/OU=hellasgrid.gr/CN=voms.hellasgrid.gr] "dteam" Done
Creating proxy ............................... Done
Your proxy is valid until Fri Apr 22 02:18:32 2011

[apaolini@ui ~]$ glite-wms-job-delegate-proxy -d paulaner -e https://devel11.cnaf.infn.it:7443/g...

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

================== glite-wms-job-delegate-proxy Success ==================

Your proxy has been successfully delegated to the WMProxy(s):
https://devel11.cnaf.infn.it:7443/g...
with the delegation identifier: paulaner

==========================================================================

[apaolini@ui ~]$ voms-proxy-init -voms dteam:/dteam/Role=lcgadmin
Enter GRID pass phrase:
Your identity: /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Alessandro Paolini
Creating temporary proxy .................... Done
Contacting voms.hellasgrid.gr:15004 [/C=GR/O=HellasGrid/OU=hellasgrid.gr/CN=voms.hellasgrid.gr] "dteam" Done
Creating proxy ............................................................................................................... Done
Your proxy is valid until Fri Apr 22 02:20:51 2011

[apaolini@ui ~]$ glite-wms-job-submit -d paulaner -e https://devel11.cnaf.infn.it:7443/g... test.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

Warning - Unable to register the job to the service: https://devel11.cnaf.infn.it:7443/g...
Client proxy FQAN (/dteam/Role=lcgadmin/Capability=NULL) does not match delegated proxy FQAN (/dteam/Role=NULL/Capability=NULL)

Method: jobRegister 

Considered successful if the job is not submitted

bug #55814: the amount of information logged to the LB needs to be reviewed Implemented

Test description.

[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf job.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/U...

==========================================================================

Peeking at a logging-info:

Event: Transfer
- Arrived = Thu Apr 21 15:19:18 2011 CEST
- Dest host = localhost
- Dest instance = /var/logmonitor/CondorG.log/CondorG.1303347530.log
- Dest jobid = unavailable
- Destination = LogMonitor
- Host = devel11.cnaf.infn.it
- Level = SYSTEM
- Priority = synchronous
- Reason = unavailable
- Result = START
- Seqcode = UI=000000:NS=0000000004:WM=000005:BH=0000000000:JSS=000002:LM=000000:LRMS=000000:APP=000000:LBS=000000
- Source = JobController
- Src instance = unique
- Timestamp = Thu Apr 21 15:19:18 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi/CN=proxy/CN=proxy
- Job = unavailable

Event: Transfer
- Arrived = Thu Apr 21 15:19:18 2011 CEST
- Dest host = localhost
<...>
- Job = (unavailable)

As a side note, also the classad file is not written anymore

[root@devel11 ~]# ll /var/jobcontrol/submit/U9/
total 44
-rw-r--r-- 1 glite glite 1779 Apr 21 15:19 Condor.https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2fU9uageRLsGUgszr2Ix3y9A.submit
-rw-r--r-- 1 glite glite 40595 Apr 21 15:19 JobWrapper.https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2fU9uageRLsGUgszr2Ix3y9A.sh 

Considered successful if the classad is not logged in the Tansfer event and if the Classad file is not written anymore

bug #56734: ListMatch should consider also SDJ specification Not implemented

Test description.

After inserting the following attribute into the jdl: ShortDeadlineJob = true; a listmatch is executed. Considered successful if the listmatch returns only a subset of the available CEs and a submitted job is executed and comes to succesful completion.

bug #56933: WMProxy Server: gSoap needs to be built with WITH_IPV6 flag Not implemented

Test description.

emi.wms.wmproxy workspace:

[mcecchi@devel10 emi.wms.wmproxy]$ grep -2r IPV6 .
./src/server/Makefile.am- -DWANT_NAMESPACES \
./src/server/Makefile.am- -DWITH_FASTCGI \
./src/server/Makefile.am: -DWITH_IPV6
./src/server/Makefile.am-
./src/server/Makefile.am-MAINTAINERCLEANFILES = Makefile.in
-- 
./configure.ac-AC_STRUCT_TM
./configure.ac-
./configure.ac:AH_TEMPLATE([WITH_IPV6], [For Ipv6 support])
./configure.ac:AC_DEFINE(WITH_IPV6)
./configure.ac-
./configure.ac- 

Considered successful if the flag -DWITH_IPV6 is found in the Makefile.

bug #58878: Request for a feature allowing propagation of generic parameters from JDL to LRMs Not implemented

Test description.

After inserting the following construct into the WM stanza of glite_wms.conf:

PropagateToLRMS= {
[ name = "smpgranularity"; value = jdl.SMPGranularity ],
[ name = "wholenodes"; value = jdl.WholeNodes ; requires = jdl.WholeNodes == true; ],
[ name = "hostsmpsize"; value = ce.GlueHostArchitectureSMPSize ],
[ name = "mpi_type"; value = jdl.MpiType; requires = ce.GlueCEInfoLRMSType == "lsf"; ],
[ name = "hostmainmem"; value = ce.GlueHostMainMemoryRAMSize; requires = ce.GlueCEInfoLRMSType == "pbs"; ]
};

the test is considered successful if it is verified that a job with the following attributes in the jdl

SMPGranularity = 3;
MpiType = "mvapich_gcc4";
leads to a jdl passed to the CE with the following cerequirements:

CeRequirements = "true && ( true && ( ( ShortDeadlineJob is true ) ? false : true ) && ( EnableWmsFeedback is true ? true : true ) )&&smpgranularity==3&&hostsmpsize==8&&hostmainmem==16384";

with lrms_type = "pbs";

bug #58968: Request for handling SMPGranularity attribute in the JDL Implemented

Test description.

[mcecchi@cert-19 ~]$ head -5 job.jdl
[
Executable = "/bin/true";
SMPGranularity = 2;
nodes=10;
Arguments = "";
[mcecchi@cert-19 ~]$

[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf job.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/i...

==========================================================================

[mcecchi@cert-19 ~]$ glite-wms-job-logging-info -v 3 https://devel11.cnaf.infn.it:9000/i... | grep equirements
<...>
requirements = ( ( ( ( true ) && ( other.GlueCEStateStatus == "Production" ) ) && other.GlueHostArchitectureSMPSize >= SMPGranularity ) && other.GlueHostArchitectureSMPSize >= SMPGranularity ) && ( ( ( ShortDeadlineJob is true ) ? RegExp(".sdj$",other.GlueCEUniqueID) : !RegExp(".sdj$",other.GlueCEUniqueID) ) && ( other.GlueCEPolicyMaxTotalJobs == 0 || other.GlueCEStateTotalJobs < other.GlueCEPolicyMaxTotalJobs ) && ( EnableWmsFeedback is true ? RegExp("cream",other.GlueCEImplementationName,"i") : true ) );
<...>

Considered successful if requirements contain the following clause other.GlueHostArchitectureSMPSize >= SMPGranularity

bug #55122: WM running but not processing jobs Not implemented

A proper way to test this bug was not found.

bug #59781: limit maximum sleep time in job wrapper Not implemented

Test description.

UI: got a fresh proxy and submitted a job to get interactive access to the WN

[mcecchi@cert-19 ~]$ voms-proxy-info
subject : /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi/CN=proxy
issuer : /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi
identity : /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi
type : proxy
strength : 1024 bits
path : /tmp/x509up_u500
timeleft : 11:56:23

[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf cloud.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/K...

==========================================================================

WN:
bash-3.2$ ps ux | grep sleep
dte129 20959 0.0 0.0 58932 500 ? S 13:29 0:00 sleep 1000 

Considered successful if the sleep argument is 1000 seconds

bug #61557: user job is not killed when proxy expires Not implemented

Test description.

submitted a job that provides interactive access.

[mcecchi@ui ~]$ cat interactive.sh
#!/bin/sh
mkfifo fifo; nc wms009.cnaf.infn.it 20001 <fifo | /bin/bash -i &> fifo; rm fifo

[root@wms009 mcecchi]# nc -l 20001
WN:
bash: no job control in this shell
bash-3.2$ ls
err.log
fifo
https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2fRt6CVAJzhpw4TZWlgrw78Q.output
interactive.sh
out.log
tmp.bsXSl20922
tmp.FaMJj20924
tmp.NjfKR20923
tmp.ojneI20911
tmp.PMecA20910
bash-3.2$ pwd
/home/dteam034/globus-tmp.grwn213.20535.0/https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2fRt6CVAJzhpw4TZWlgrw78Q
bash-3.2$ ps xo pgrp,pid,stat,cmd
PGRP PID STAT CMD
20424 20424 Ss -sh
20424 20531 S /bin/sh /var/torque/mom_priv/jobs/821801.grce001.inr.troitsk.ru.SC
20424 20535 S /usr/bin/perl -w /tmp/bootstrap.r20532 /home/dteam034/ grce001.inr.troitsk.ru /home/dteam034/.globus/job/grce001.inr.troitsk.ru/2992.1303461286/x509_up X509GPG:globus-cache-export.BU3427.gpg /dev/null /home/dteam034/.globus/job/grce001.inr.troitsk.ru/2992.1303461286/stdout stdoutftp /home/dteam034/.globus/job/grce001.inr.troitsk.ru/2992.1303461286/stderr stderrftp /home/dteam034/.lcgjm/globus-cache-export.BU3427 https://grce001.inr.troitsk.ru:2002... /home/dteam034/ NONE /home/dteam034/.globus/.gass_cache/local/md5/33/217dcbdc0a0cc3de7c87eaeac07fd8/md5/e0/fbf2948266acc97e62b731ff1c259e/data hostsmpsize=8
20424 20541 S /usr/bin/perl -w /tmp/bootstrap.r20532 /home/dteam034/ grce001.inr.troitsk.ru /home/dteam034/.globus/job/grce001.inr.troitsk.ru/2992.1303461286/x509_up X509GPG:globus-cache-export.BU3427.gpg /dev/null /home/dteam034/.globus/job/grce001.inr.troitsk.ru/2992.1303461286/stdout stdoutftp /home/dteam034/.globus/job/grce001.inr.troitsk.ru/2992.1303461286/stderr stderrftp /home/dteam034/.lcgjm/globus-cache-export.BU3427 https://grce001.inr.troitsk.ru:2002... /home/dteam034/ NONE /home/dteam034/.globus/.gass_cache/local/md5/33/217dcbdc0a0cc3de7c87eaeac07fd8/md5/e0/fbf2948266acc97e62b731ff1c259e/data hostsmpsize=8
20424 20648 S /usr/bin/perl -w /tmp/bootstrap.r20532 /home/dteam034/ grce001.inr.troitsk.ru /home/dteam034/.globus/job/grce001.inr.troitsk.ru/2992.1303461286/x509_up X509GPG:globus-cache-export.BU3427.gpg /dev/null /home/dteam034/.globus/job/grce001.inr.troitsk.ru/2992.1303461286/stdout stdoutftp /home/dteam034/.globus/job/grce001.inr.troitsk.ru/2992.1303461286/stderr stderrftp /home/dteam034/.lcgjm/globus-cache-export.BU3427 https://grce001.inr.troitsk.ru:2002... /home/dteam034/ NONE /home/dteam034/.globus/.gass_cache/local/md5/33/217dcbdc0a0cc3de7c87eaeac07fd8/md5/e0/fbf2948266acc97e62b731ff1c259e/data hostsmpsize=8
20424 20682 S sh -c if [ -x ${LCG_LOCATION:-/opt/lcg}/libexec/jobwrapper ]; then ${LCG_LOCATION:-/opt/lcg}/libexec/jobwrapper /home/dteam034/globus-tmp.grwn213.20535.0/globus-tmp.grwn213.20535.2 hostsmpsize=8; else /home/dteam034/globus-tmp.grwn213.20535.0/globus-tmp.grwn213.20535.2 hostsmpsize=8; fi
20424 20683 S /bin/sh /opt/lcg/libexec/jobwrapper /home/dteam034/globus-tmp.grwn213.20535.0/globus-tmp.grwn213.20535.2 hostsmpsize=8
20424 20745 S /bin/sh /home/dteam034/globus-tmp.grwn213.20535.0/globus-tmp.grwn213.20535.2 hostsmpsize=8
20424 20950 S /bin/sh /home/dteam034/globus-tmp.grwn213.20535.0/globus-tmp.grwn213.20535.2 hostsmpsize=8
20951 20951 S /bin/sh /home/dteam034/globus-tmp.grwn213.20535.0/globus-tmp.grwn213.20535.2 hostsmpsize=8
20424 20952 S /bin/sh /home/dteam034/globus-tmp.grwn213.20535.0/globus-tmp.grwn213.20535.2 hostsmpsize=8
20951 20955 S /usr/bin/time -p -o tmp.bsXSl20922 ./interactive.sh
20951 20956 S /bin/sh ./interactive.sh
20951 20958 S nc wms009.cnaf.infn.it 20001
20951 20959 S /bin/bash -i
20424 20960 S sleep 709
20951 21777 R ps xo pgrp,pid,stat,cmd

/usr/bin/time -p -o tmp.bsXSl20922 ./interactive.sh runs in bg with the same process group 20951 of /bin/sh ./interactive.sh and its children

More specifically: UI:

[mcecchi@ui ~]$ voms-proxy-info
subject : /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi/CN=proxy
issuer : /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi
identity : /C=IT/O=INFN/OU=Personal Certificate/L=CNAF/CN=Marco Cecchi
type : proxy
strength : 1024 bits
path : /tmp/x509up_u40121
timeleft : 0:14:47

[mcecchi@ui ~]$ glite-wms-job-submit -a -c devel11.conf interactive.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/U...

===========================================================

[root@wms009 mcecchi]# nc -l 20001
bash: no job control in this shell
bash-3.2$ pwd
/var/spool/pbs/tmpdir/52994.gridgate.ucc.ie/https_3a_2f_2fdevel11.cnaf.infn.it_3

a9000_2fZLKFE55HSB9gKxcbqhTaeg
bash-3.2$ ls
interactive.sh
err.log
fifo
https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2fZLKFE55HSB9gKxcbqhTaeg.output
out.log
tmp.dJPnX32021
tmp.IWdDT32037
tmp.LNLGJ32022
tmp.udGTb32036
tmp.ZEXBZ32035
bash-3.2$ ps x
PID TTY STAT TIME COMMAND
31878 ? Ss 0:00 -sh
31911 ? S 0:00 /bin/sh /var/spool/pbs/mom_priv/jobs/52994.gridgate.u

cc.ie.SC
31912 ? S 0:00 /bin/sh /home/dte150/.globus/.gass_cache/local/md5/55

/adef21944b1300765235bd50f58cb0/md5/d7/71611fd2c715ec1fcd4ddb5c8cf3b3/data hosts

mpsize=2,hostmainmem=200
32047 ? S 0:00 /bin/sh /home/dte150/.globus/.gass_cache/local/md5/55

/adef21944b1300765235bd50f58cb0/md5/d7/71611fd2c715ec1fcd4ddb5c8cf3b3/data hosts

mpsize=2,hostmainmem=200
32048 ? S 0:00 /bin/sh /home/dte150/.globus/.gass_cache/local/md5/55

/adef21944b1300765235bd50f58cb0/md5/d7/71611fd2c715ec1fcd4ddb5c8cf3b3/data hosts

mpsize=2,hostmainmem=200
32049 ? S 0:00 /bin/sh /home/dte150/.globus/.gass_cache/local/md5/55

/adef21944b1300765235bd50f58cb0/md5/d7/71611fd2c715ec1fcd4ddb5c8cf3b3/data hosts

mpsize=2,hostmainmem=200
32050 ? S 0:00 /usr/bin/time -p -o tmp.ZEXBZ32035 ./interactive.sh
32053 ? S 0:00 /bin/sh ./interactive.sh
32055 ? S 0:00 nc wms009.cnaf.infn.it 20001
32056 ? S 0:00 /bin/bash -i
32057 ? S 0:00 sleep 818
32065 ? R 0:00 ps x
bash-3.2$

818 seconds later the job exits...

PID TTY STAT TIME COMMAND
802 ? R 0:00 ps x
31878 ? Ss 0:00 -sh
31911 ? S 0:00 /bin/sh /var/spool/pbs/mom_priv/jobs/52994.gridgate.ucc.ie.SC
31912 ? S 0:00 /bin/sh /home/dte150/.globus/.gass_cache/local/md5/55/adef21944b1300765235bd50f58cb0/md5/d7/71611fd2c715ec1fcd4ddb5c8cf3b3/data hostsmpsize=2,hostmainmem=200
32047 ? S 0:00 /bin/sh /home/dte150/.globus/.gass_cache/local/md5/55/adef21944b1300765235bd50f58cb0/md5/d7/71611fd2c715ec1fcd4ddb5c8cf3b3/data hostsmpsize=2,hostmainmem=200
32048 ? S 0:00 /bin/sh /home/dte150/.globus/.gass_cache/local/md5/55/adef21944b1300765235bd50f58cb0/md5/d7/71611fd2c715ec1fcd4ddb5c8cf3b3/data hostsmpsize=2,hostmainmem=200
32049 ? S 0:00 /bin/sh /home/dte150/.globus/.gass_cache/local/md5/55/adef21944b1300765235bd50f58cb0/md5/d7/71611fd2c715ec1fcd4ddb5c8cf3b3/data hostsmpsize=2,hostmainmem=200
32050 ? S 0:00 /usr/bin/time -p -o tmp.ZEXBZ32035 ./cloud.sh
32053 ? S 0:00 /bin/sh ./cloud.sh
32055 ? S 0:00 nc wms009.cnaf.infn.it 20001
32056 ? S 0:00 /bin/bash -i
32057 ? S 0:00 sleep 818
bash-3.2$ [root@wms009 mcecchi]# 

Considered successful if the job is killed.

bug #62211: [ yaim-wms ] Enable Glue 2.0 publishing Implemented

Test description.

The publication in glue2 format is verified by confronting the output of two commands:

ldapsearch -x -H ldap://devel09.cnaf.infn.it:2170 -b mds-vo-name=resource,o=grid

for glue1 and

ldapsearch -x -H ldap://devel09.cnaf.infn.it:2170 -b o=glue

for glue2. These give respectively the following output:

# extended LDIF
#
# LDAPv3
# base <mds-vo-name=resource,o=grid> with scope subtree
# filter: (objectclass=*)
# requesting: ALL
#

# resource, grid
dn: Mds-Vo-name=resource,o=grid
Mds-Vo-name: resource
objectClass: Mds

# devel09.cnaf.infn.it_org.glite.wms.WMProxy_3504844140, resource, grid
dn: GlueServiceUniqueID=devel09.cnaf.infn.it_org.glite.wms.WMProxy_3504844140,
 Mds-Vo-name=resource,o=grid
GlueServiceAccessControlBaseRule: VO:atlas
GlueServiceAccessControlBaseRule: VO:alice
GlueServiceAccessControlBaseRule: VO:lhcb
GlueServiceAccessControlBaseRule: VO:cms
GlueServiceAccessControlBaseRule: VO:dteam
GlueServiceAccessControlBaseRule: VO:biomed
GlueServiceAccessControlBaseRule: VO:ops
GlueServiceAccessControlBaseRule: VO:infngrid
GlueServiceStatus: Other
GlueServiceStatusInfo: ??
objectClass: GlueTop
objectClass: GlueService
objectClass: GlueKey
objectClass: GlueSchemaVersion
GlueServiceUniqueID: devel09.cnaf.infn.it_org.glite.wms.WMProxy_3504844140
GlueServiceAccessControlRule: atlas
GlueServiceAccessControlRule: alice
GlueServiceAccessControlRule: lhcb
GlueServiceAccessControlRule: cms
GlueServiceAccessControlRule: dteam
GlueServiceAccessControlRule: biomed
GlueServiceAccessControlRule: ops
GlueServiceAccessControlRule: infngrid
GlueServiceStartTime: 1970-01-01T01:00:00+01:00
GlueServiceEndpoint: https://devel09.cnaf.infn.it:7443/glite_wms_wmproxy_serve
 r
GlueServiceVersion: 3.3.99
GlueSchemaVersionMinor: 3
GlueServiceName: INFN-CREAMTEST-WMProxy
GlueServiceType: org.glite.wms.WMProxy
GlueServiceWSDL: http://trinity.datamat.it/projects/EGEE/WMProxy/WMProxy.wsdl
GlueServiceSemantics: https://edms.cern.ch/file/674643/1/EGEE-JRA1-TEC-674643-
 WMPROXY-guide-v0-3.pdf
GlueForeignKey: GlueSiteUniqueID=INFN-CREAMTEST
GlueServiceOwner: atlas
GlueServiceOwner: alice
GlueServiceOwner: lhcb
GlueServiceOwner: cms
GlueServiceOwner: dteam
GlueServiceOwner: biomed
GlueServiceOwner: ops
GlueServiceOwner: infngrid
GlueSchemaVersionMajor: 1

# DN, devel09.cnaf.infn.it_org.glite.wms.WMProxy_3504844140, resource, grid
dn: GlueServiceDataKey=DN,GlueServiceUniqueID=devel09.cnaf.infn.it_org.glite.w
 ms.WMProxy_3504844140,Mds-Vo-name=resource,o=grid
GlueSchemaVersionMajor: 1
objectClass: GlueTop
objectClass: GlueServiceData
objectClass: GlueKey
objectClass: GlueSchemaVersion
GlueSchemaVersionMinor: 3
GlueServiceDataKey: DN
GlueChunkKey: GlueServiceUniqueID=devel09.cnaf.infn.it_org.glite.wms.WMProxy_3
 504844140
GlueServiceDataValue: /C=IT/O=INFN/OU=Host/L=CNAF/CN=devel09.cnaf.infn.it

# glite-info-service_version, devel09.cnaf.infn.it_org.glite.wms.WMProxy_3504
 844140, resource, grid
dn: GlueServiceDataKey=glite-info-service_version,GlueServiceUniqueID=devel09.
 cnaf.infn.it_org.glite.wms.WMProxy_3504844140,Mds-Vo-name=resource,o=grid
GlueSchemaVersionMajor: 1
objectClass: GlueTop
objectClass: GlueServiceData
objectClass: GlueKey
objectClass: GlueSchemaVersion
GlueSchemaVersionMinor: 3
GlueServiceDataKey: glite-info-service_version
GlueChunkKey: GlueServiceUniqueID=devel09.cnaf.infn.it_org.glite.wms.WMProxy_3
 504844140
GlueServiceDataValue: 1.6

# glite-info-service_hostname, devel09.cnaf.infn.it_org.glite.wms.WMProxy_350
 4844140, resource, grid
dn: GlueServiceDataKey=glite-info-service_hostname,GlueServiceUniqueID=devel09
 .cnaf.infn.it_org.glite.wms.WMProxy_3504844140,Mds-Vo-name=resource,o=grid
GlueSchemaVersionMajor: 1
objectClass: GlueTop
objectClass: GlueServiceData
objectClass: GlueKey
objectClass: GlueSchemaVersion
GlueSchemaVersionMinor: 3
GlueServiceDataKey: glite-info-service_hostname
GlueChunkKey: GlueServiceUniqueID=devel09.cnaf.infn.it_org.glite.wms.WMProxy_3
 504844140
GlueServiceDataValue: devel09.cnaf.infn.it

# search result
search: 2
result: 0 Success

# numResponses: 6
# numEntries: 5

and

# extended LDIF
#
# LDAPv3
# base <o=glue> with scope subtree
# filter: (objectclass=*)
# requesting: ALL
#

# glue
dn: o=glue
objectClass: organization
o: glue

# grid, glue
dn: GLUE2GroupID=grid,o=glue
objectClass: GLUE2Group
GLUE2GroupID: grid

# resource, glue
dn: GLUE2GroupID=resource,o=glue
objectClass: GLUE2Group
GLUE2GroupID: resource

# devel09.cnaf.infn.it_3504844140, resource, glue
dn: GLUE2ServiceID=devel09.cnaf.infn.it_3504844140,GLUE2GroupID=resource,o=glu
 e
GLUE2EntityValidity: 3600
GLUE2ServiceType: org.glite.wms.WMProxy
GLUE2ServiceID: devel09.cnaf.infn.it_3504844140
objectClass: GLUE2Entity
objectClass: GLUE2Service
GLUE2ServiceQualityLevel: production
GLUE2EntityOtherInfo: InfoProviderName=glite-info-glue2-service
GLUE2EntityOtherInfo: InfoProviderVersion=1.0
GLUE2EntityOtherInfo: InfoProviderHost=devel09.cnaf.infn.it
GLUE2ServiceComplexity: endpointType=1, share=0, resource=0
GLUE2EntityName: INFN-CREAMTEST-WMProxy
GLUE2ServiceAdminDomainForeignKey: INFN-CREAMTEST
GLUE2EntityCreationTime: 2011-04-21T14:43:18Z

# devel09.cnaf.infn.it_3504844140_org.glite.wms.WMProxy_3504844140, devel09.c
 naf.infn.it_3504844140, resource, glue
dn: GLUE2EndpointID=devel09.cnaf.infn.it_3504844140_org.glite.wms.WMProxy_3504
 844140,GLUE2ServiceID=devel09.cnaf.infn.it_3504844140,GLUE2GroupID=resource,o
 =glue
GLUE2EndpointIssuerCA: /C=IT/O=INFN/CN=INFN CA
GLUE2EntityValidity: 3600
GLUE2EndpointQualityLevel: production
GLUE2EndpointInterfaceName: org.glite.wms.WMProxy
GLUE2EndpointTrustedCA: IGTF
objectClass: GLUE2Entity
objectClass: GLUE2Endpoint
GLUE2EndpointID: devel09.cnaf.infn.it_3504844140_org.glite.wms.WMProxy_3504844
 140
GLUE2EndpointInterfaceVersion: 3.3.99
GLUE2EndpointURL: https://devel09.cnaf.infn.it:7443/glite_wms_wmproxy_server
GLUE2EntityOtherInfo: InfoProviderName=glite-info-glue2-endpoint
GLUE2EntityOtherInfo: InfoProviderVersion=1.0
GLUE2EntityOtherInfo: InfoProviderHost=devel09.cnaf.infn.it
GLUE2EndpointDowntimeInfo: See the GOC DB for downtimes: https://goc.gridops.o
 rg/
GLUE2EndpointWSDL: http://trinity.datamat.it/projects/EGEE/WMProxy/WMProxy.wsd
 l
GLUE2EndpointServingState: production
GLUE2EndpointHealthState: other
GLUE2EndpointSemantics: https://edms.cern.ch/file/674643/1/EGEE-JRA1-TEC-67464
 3-WMPROXY-guide-v0-3.pdf
GLUE2EndpointServiceForeignKey: devel09.cnaf.infn.it_3504844140
GLUE2EntityName: WMProxy endpoint for Service devel09.cnaf.infn.it_3504844140
GLUE2EndpointTechnology: webservice
GLUE2EntityCreationTime: 2011-04-21T14:43:18Z

# DN_1, devel09.cnaf.infn.it_3504844140_org.glite.wms.WMProxy_3504844140, dev
 el09.cnaf.infn.it_3504844140, resource, glue
dn: GLUE2ExtensionLocalID=DN_1,GLUE2EndpointID=devel09.cnaf.infn.it_3504844140
 _org.glite.wms.WMProxy_3504844140,GLUE2ServiceID=devel09.cnaf.infn.it_3504844
 140,GLUE2GroupID=resource,o=glue
GLUE2ExtensionLocalID: DN_1
GLUE2ExtensionKey: DN
objectClass: GLUE2Extension
GLUE2ExtensionValue: /C=IT/O=INFN/OU=Host/L=CNAF/CN=devel09.cnaf.infn.it
GLUE2ExtensionEntityForeignKey: devel09.cnaf.infn.it_3504844140_org.glite.wms.
 WMProxy_3504844140

# devel09.cnaf.infn.it_3504844140_org.glite.wms.WMProxy_3504844140_Policy, de
 vel09.cnaf.infn.it_3504844140_org.glite.wms.WMProxy_3504844140, devel09.cnaf
 .infn.it_3504844140, resource, glue
dn: GLUE2PolicyID=devel09.cnaf.infn.it_3504844140_org.glite.wms.WMProxy_350484
 4140_Policy,GLUE2EndpointID=devel09.cnaf.infn.it_3504844140_org.glite.wms.WMP
 roxy_3504844140,GLUE2ServiceID=devel09.cnaf.infn.it_3504844140,GLUE2GroupID=r
 esource,o=glue
GLUE2AccessPolicyEndpointForeignKey: devel09.cnaf.infn.it_3504844140_org.glite
 .wms.WMProxy_3504844140
GLUE2EntityValidity: 3600
objectClass: GLUE2Entity
objectClass: GLUE2Policy
objectClass: GLUE2AccessPolicy
GLUE2PolicyRule: VO:atlas
GLUE2PolicyRule: VO:alice
GLUE2PolicyRule: VO:lhcb
GLUE2PolicyRule: VO:cms
GLUE2PolicyRule: VO:dteam
GLUE2PolicyRule: VO:biomed
GLUE2PolicyRule: VO:ops
GLUE2PolicyRule: VO:infngrid
GLUE2PolicyScheme: org.glite.standard
GLUE2EntityOtherInfo: InfoProviderName=glite-info-glue2-endpoint
GLUE2EntityOtherInfo: InfoProviderVersion=1.0
GLUE2EntityOtherInfo: InfoProviderHost=devel09.cnaf.infn.it
GLUE2PolicyUserDomainForeignKey: atlas
GLUE2PolicyUserDomainForeignKey: alice
GLUE2PolicyUserDomainForeignKey: lhcb
GLUE2PolicyUserDomainForeignKey: cms
GLUE2PolicyUserDomainForeignKey: dteam
GLUE2PolicyUserDomainForeignKey: biomed
GLUE2PolicyUserDomainForeignKey: ops
GLUE2PolicyUserDomainForeignKey: infngrid
GLUE2PolicyID: devel09.cnaf.infn.it_3504844140_org.glite.wms.WMProxy_350484414
 0_Policy
GLUE2EntityName: Access control rules for Endpoint devel09.cnaf.infn.it_350484
 4140_org.glite.wms.WMProxy_3504844140
GLUE2EntityCreationTime: 2011-04-21T14:43:18Z

# search result
search: 2
result: 0 Success

# numResponses: 8
# numEntries: 7

Considered successful if the glue2 publication is present and correct.

bug #62709: glite_wms_wmproxy_load_monitor has a problem with lvm partitions Not implemented

This bug can only be tested on a machine mounting lvm partitions.

bug #64416: the proxycache purger needs to be made compatible with the latest gridsite releases Not implemented

Test description.

[root@devel11 proxycache]# ll
total 24
drwx------ 3 glite glite 4096 Apr 21 14:19 %2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DCNAF%2FCN%3DAlessandro%20Paolini
drwx------ 14 glite glite 4096 Apr 21 15:09 %2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DCNAF%2FCN%3DFabio%20Capannini
drwx------ 90 glite glite 4096 Apr 21 14:57 %2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DCNAF%2FCN%3DMarco%20Cecchi
drwx------ 6 glite glite 4096 Apr 20 22:26 %2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DCNAF%2FCN%3DMarco%20Cecchi%2FCN%3D57595619
drwx------ 116 glite glite 4096 Apr 20 17:02 %2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DPadova%2FCN%3DAlessio%20Gianelle
drwx------ 7 glite glite 4096 Apr 21 14:19 cache
[root@devel11 proxycache]# /usr/bin/glite-wms-wmproxy-purge-proxycache
Thu Apr 21 15:09:55 CEST 2011: processing file: ./%2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DCNAF%2FCN%3DAlessandro%20Paolini/paulaner/userproxy.pem
Thu Apr 21 15:09:55 CEST 2011: the proxy is still valid (40105 seconds left)
Thu Apr 21 15:09:55 CEST 2011: processing file: ./%2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DCNAF%2FCN%3DFabio%20Capannini/oABpQjeGPPvdLh41Tl3ybg/userproxy.pem
Thu Apr 21 15:09:56 CEST 2011: the proxy is still valid (34162 seconds left)
Thu Apr 21 15:09:56 CEST 2011: processing file: ./%2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DCNAF%2FCN%3DFabio%20Capannini/6xDauMvftsHomfwFTOHeEw/userproxy.pem
Thu Apr 21 15:09:56 CEST 2011: the proxy is still valid (34184 seconds left)
Thu Apr 21 15:09:56 CEST 2011: processing file: ./%2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DCNAF%2FCN%3DFabio%20Capannini/QDRm_Ynuz2IqbI8b_BDlng/userproxy.pem
Thu Apr 21 15:09:56 CEST 2011: the proxy is still valid (27605 seconds left)
Thu Apr 21 15:09:56 CEST 2011: processing file: ./tmp.LcpuIS6761
Thu Apr 21 15:09:56 CEST 2011: couldn't remove (not a proxy file)
Thu Apr 21 15:09:56 CEST 2011: processing file: ./%2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DCNAF%2FCN%3DMarco%20Cecchi/zURreRHoNkG9cmb4_ctH0Q/userproxy.pem
Thu Apr 21 15:09:56 CEST 2011: the proxy is still valid (21076 seconds left)
Thu Apr 21 15:09:56 CEST 2011: processing file: ./%2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DCNAF%2FCN%3DMarco%20Cecchi/HCIEIbuW9bg3tTp_PihPjg/userproxy.pem
Thu Apr 21 15:09:56 CEST 2011: the proxy is still valid (21113 seconds left)
Thu Apr 21 15:09:56 CEST 2011: processing file: ./%2FC%3DIT%2FO%3DINFN%2FOU%3DPersonal%20Certificate%2FL%3DCNAF%2FCN%3DMarco%20Cecchi/7dyB5AN8UiPW6x7v0XucDQ/userproxy.pem
Thu Apr 21 15:09:56 CEST 2011: the proxy is still valid (21074 seconds left)
<...>
etc. etc.

Considered successful if the expired certificates are removed.

bug #68944: Bug in ICE's start/stop script Not implemented

Test description.

[root@devel09 ~]# ps aux |grep ice
glite     2716  0.0  0.0 220996  2192 ?        Ss   09:36   0:00 /usr/bin/glite-wms-ice-safe --conf glite_wms.conf --daemon /tmp/glite-wms-ice-safe.pid
glite     2720  0.0  0.0  63856  1108 ?        S    09:36   0:00 sh -c /usr/bin/glite-wms-ice --conf glite_wms.conf /var/log/wms/ice_console.log 2>&1
glite     2721  0.0  0.1 483464  7404 ?        Sl   09:36   0:02 /usr/bin/glite-wms-ice --conf glite_wms.conf /var/log/wms/ice_console.log
root     31253  0.0  0.0  61184   788 pts/1    S+   14:29   0:00 grep ice
[root@devel09 ~]# /etc/init.d/glite-wms-ice restart
stopping ICE... ok
starting ICE... ok
[root@devel09 ~]# ps aux |grep ice
glite    31318  0.0  0.0 220996  2196 ?        Ss   14:30   0:00 /usr/bin/glite-wms-ice-safe --conf glite_wms.conf --daemon /tmp/glite-wms-ice-safe.pid
glite    31322  0.0  0.0  63856  1112 ?        S    14:30   0:00 sh -c /usr/bin/glite-wms-ice --conf glite_wms.conf /var/log/wms/ice_console.log 2>&1
glite    31323  1.0  0.1 483464  7352 ?        Sl   14:30   0:00 /usr/bin/glite-wms-ice --conf glite_wms.conf /var/log/wms/ice_console.log
root     31363  0.0  0.0  61180   784 pts/1    S+   14:30   0:00 grep ice

Considered successful if ice processes are restarted.

bug #70061: WMS hates collections with 192 nodes! Implemented

Test description.

[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf coll_192.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/T...

==========================================================================

Considered successful if the collection is submitted and executed correctly.

bug #70331: glite-wms-create-proxy "ambiguous redirect" Not implemented

Test description.

[root@cream-44 cron.d]# su - glite
[glite@cream-44 ~]$ /usr/sbin/glite-wms-create-proxy.sh /var/wms.proxy /var/log/wms/create_proxy.log
[glite@cream-44 ~]$

Considered successful if the command works with no error messages.

bug #70824: environment values in JDL cannot have spaces Implemented

Test description.

Submit this jdl:

[
Executable = "/bin/env";
Stdoutput = "env.out" ;
StdError = "env.err" ;
shallowretrycount = 2;
retrycount = 3;
OutputSandbox = { "env.out" ,"env.err" } ;
Environment = {
"ATHENA_RUN_EVENTS='[(152345L, 216721L)]'"
};

]

======================= glite-wms-job-status Success =====================
BOOKKEEPING INFORMATION:

Status info for the Job : https://devel09.cnaf.infn.it:9000/g...
Current Status: Done (Success)
Exit code: 0
Status Reason: Job terminated successfully
Destination: gridgate.ucc.ie:2119/jobmanager-pbs-sixhour
Submitted: Thu Apr 21 17:11:46 2011 CEST
==========================================================================

After output retrieval:

[ale@cream-12 UI]$ glite-wms-job-output https://devel09.cnaf.infn.it:9000/g...

Connecting to the service https://devel09.cnaf.infn.it:7443/g...

================================================================================

JOB GET OUTPUT OUTCOME

Output sandbox files for the job:
https://devel09.cnaf.infn.it:9000/g...
have been successfully retrieved and stored in the directory:
/tmp/jobOutput/ale_gvpbgkBN_KkLhly-uNjpxQ

================================================================================

Check the output file:

[ale@cream-12 UI]$ cat /tmp/jobOutput/ale_gvpbgkBN_KkLhly-uNjpxQ/env.out | grep ATHENA
ATHENA_RUN_EVENTS=[(152345L, 216721L)] 

Considered successful if the environment contains the variable with spaces.

bug #71863: JobWrapper tries to use "test -eq" for string comparison Not implemented

Test description.

Looking into a JW:

fatal_error() # 1 - reason, 2 - transfer OSB
{
if [ "x${LM_done_failed_reason}" != "x" ]; then
jw_echo "LM_log_done_begin" # parsed 'as is' by LM
jw_echo "${LM_done_failed_reason}"
jw_echo "LM_log_done_end" # parsed 'as is' by LM
fi
jw_echo "$1"
log_done_failed 1 "$1"
if [ "x$2" == "xOSB" ]; then
OSB_transfer
fi
doExit 1
} 

Considered successful if the comparison is done via ==

bug #73192: Submission failed due to a credential problem Implemented

Test description.

Simply try a submit:

[ale@cream-12 UI]$ glite-wms-job-submit -a -c etc/wmp_devel09.conf jdl/deepresub.jdl

Connecting to the service https://devel09.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel09.cnaf.infn.it:9000/L...

========================================================================== 

Considered successful if the job is correctly submitted.

bug #73699: Wrong retry count computation Implemented

Test description.

Submit this jdl:

[
requirements = (other.GlueCEStateStatus == "Production");
Rank = -2 * other.GlueCEStateWaitingJobs;
Executable = "/bin/ls";
prologue = "/bin/false";
shallowretrycount = -1;
RetryCount = 3;
usertags = [ exe = "deepresub" ];
requirements=regexp("8443/cream", other.GlueCEUniqueID);

]

The job should file after 3 resubmissions:

======================= glite-wms-job-status Success =====================
BOOKKEEPING INFORMATION:

Status info for the Job : https://devel09.cnaf.infn.it:9000/L...
Current Status: Aborted
Logged Reason(s):
- Prologue failed with error 1
- reason=1; Prologue failed with error 1
- Transfer to CREAM failed due to exception: Failed to create a delegation id for job https://devel09.cnaf.infn.it:9000/L...: reason is Received NULL fault; the error is due to another cause: FaultString=[] - FaultCode=[SOAP-ENV:Server.generalException] - FaultSubCode=[SOAP-ENV:Server.generalException] - FaultDetail=[<ns1:DelegationException><msg>User C=IT,O=INFN,OU=Personal Certificate,L=Padova,CN=Alessio Gianelle not authorized for operation {http://www.gridsite.org/namespaces/...</msg></ns1:DelegationException><ns2:exceptionName xmlns:ns2="http://xml.apache.org/axis/">org.glite.security.delegation.DelegationException</ns2:exceptionName><ns3:hostname xmlns:ns3="http://xml.apache.org/axis/">cluster50.knu.ac.kr</ns3:hostname>]
- Prologue failed with error 1
- Prologue failed with error 1
- Prologue failed with error 1
- reason=1; Prologue failed with error 1
Status Reason: hit job retry count (3)
Destination: infnlns-ce-01.ct.pi2s2.it:8443/cream-lsf-cert
Submitted: Thu Apr 21 16:47:33 2011 CEST

Infact:

[ale@cream-12 UI]$ glite-wms-job-logging-info --event Resubmission https://devel09.cnaf.infn.it:9000/L...

===================== glite-wms-job-logging-info Success =====================

LOGGING INFORMATION:

Printing info for the Job : https://devel09.cnaf.infn.it:9000/L...

---
Event: Resubmission
- Result = WILLRESUB
- Source = LogMonitor
- Timestamp = Thu Apr 21 16:49:44 2011 CEST
---
Event: Resubmission
- Result = WILLRESUB
- Source = WorkloadManager
- Timestamp = Thu Apr 21 16:49:45 2011 CEST
---
Event: Resubmission
- Result = WILLRESUB
- Source = LogMonitor
- Timestamp = Thu Apr 21 16:49:50 2011 CEST
---
Event: Resubmission
- Result = WILLRESUB
- Source = WorkloadManager
- Timestamp = Thu Apr 21 16:49:51 2011 CEST
---
Event: Resubmission
- Result = WILLRESUB
- Source = LogMonitor
- Timestamp = Thu Apr 21 16:52:00 2011 CEST
---
Event: Resubmission
- Result = WILLRESUB
- Source = WorkloadManager
- Timestamp = Thu Apr 21 16:52:01 2011 CEST
---
Event: Resubmission
- Result = WILLRESUB
- Source = LogMonitor
- Timestamp = Thu Apr 21 16:54:11 2011 CEST
==========================================================================

(Two events one from LM and the other from WM for every resubmission, plus the last try which failed because limit is reached.)

Considered successful if the job is resubmitted the right number of times (three in this case).

bug #73711: edg_wll_SetLoggingJobProxy with empty sequence code returns "no state in DB" Implemented

Test description.

This bug represents a development issue. It is simply verified by checking that submission for both a job and a collection returns the jobid. The wms must be configured in proxy mode only.

[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf job.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/B...

==========================================================================

[mcecchi@cert-19 ~]$ glite-wms-job-submit -a -c devel11.conf coll_10.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/T...

========================================================================== 

Considered successful if the jobids are correctly returned.

bug #73715: missing ReallyRunning event from LogMonitor Implemented

Test description.

[ale@cream-12 UI]$ glite-wms-job-logging-info -v 3 --event ReallyRunning https://devel09.cnaf.infn.it:9000/v...

===================== glite-wms-job-logging-info Success =====================

LOGGING INFORMATION:

Printing info for the Job : https://devel09.cnaf.infn.it:9000/v...

---
Event: ReallyRunning
- Arrived = Thu Apr 21 16:08:28 2011 CEST
- Host = grid003.fc.up.pt
- Level = SYSTEM
- Priority = synchronous
- Seqcode = UI=000000:NS=0000000004:WM=000005:BH=0000000000:JSS=000002:LM=000002:LRMS=000003:APP=000000:LBS=000000
- Source = LRMS
- Timestamp = Thu Apr 21 16:08:28 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
---
Event: ReallyRunning
- Arrived = Thu Apr 21 16:09:43 2011 CEST
- Host = devel09.cnaf.infn.it
- Level = SYSTEM
- Priority = synchronous
- Seqcode = UI=000000:NS=0000000004:WM=000005:BH=0000000000:JSS=000002:LM=000007:LRMS=000000:APP=000000:LBS=000000
- Source = LogMonitor
- Timestamp = Thu Apr 21 16:09:43 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy
- Wn seq = UI=000000:NS=0000000004:WM=000005:BH=0000000000:JSS=000002:LM=000002:LRMS=000000:APP=000000:LBS=000000
   Alessio Gianelle <ale>
Project Member

Considered successful if the reallyrunning event is effectively logged.

bug #74259: Previous matches information is not taken into account if direct submission is used Implemented

Test description.

Submit directly (i.e. use option -r ) a job which trigger a resubmission (e.g. prologue="/bin/false").

Then, when it fails, check the "UserTag" events:

[ale@cream-12 UI]$ glite-wms-job-logging-info -v 2 --event UserTag https://devel09.cnaf.infn.it:9000/Q...
===================== glite-wms-job-logging-info Success =====================

LOGGING INFORMATION:

Printing info for the Job : https://devel09.cnaf.infn.it:9000/Q...

---
Event: UserTag
- Arrived = Fri Apr 22 10:10:37 2011 CEST
- Host = devel09.cnaf.infn.it
- Name = exe
- Source = NetworkServer
- Src instance = https://devel09.cnaf.infn.it:7443/g...
- Timestamp = Fri Apr 22 10:10:37 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
- Value = deepresub
---
Event: UserTag
- Arrived = Fri Apr 22 10:10:39 2011 CEST
- Host = devel09.cnaf.infn.it
- Name = CEInfoHostName
- Source = WorkloadManager
- Src instance = 21624
- Timestamp = Fri Apr 22 10:10:39 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy/CN=proxy/CN=proxy
- Value = alice16.spbu.ru
---
Event: UserTag
- Arrived = Fri Apr 22 10:10:41 2011 CEST
- Host = devel09.cnaf.infn.it
- Name = CEInfoHostName
- Source = WorkloadManager
- Src instance = 21624
- Timestamp = Fri Apr 22 10:10:41 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy/CN=proxy/CN=proxy
- Value = alice16.spbu.ru
---
Event: UserTag
- Arrived = Fri Apr 22 10:10:43 2011 CEST
- Host = devel09.cnaf.infn.it
- Name = CEInfoHostName
- Source = WorkloadManager
- Src instance = 21624
- Timestamp = Fri Apr 22 10:10:43 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy/CN=proxy/CN=proxy
- Value = alice16.spbu.ru
---
Event: UserTag
- Arrived = Fri Apr 22 10:10:45 2011 CEST
- Host = devel09.cnaf.infn.it
- Name = CEInfoHostName
- Source = WorkloadManager
- Src instance = 21624
- Timestamp = Fri Apr 22 10:10:45 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy/CN=proxy/CN=proxy
- Value = alice16.spbu.ru
==========================================================================

Considered successful if the tag "CEInfoHostName" reports the same value (equal to the chosen CE) for all the events.

bug #74577: Wrong counter in ICE database is set at the job creation Not implemented

Test description.

we can check that the first time a job is submitted the counter should be set to 1:

[ale@cream-12 UI]$ glite-wms-job-submit -a -c etc/wmp_devel09.conf test.jdl

Connecting to the service https://devel09.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel09.cnaf.infn.it:9000/Z...

==========================================================================

and on the WMS:
[root@devel09 persist_dir]# sqlite3 ice.db "select counter from proxy where myproxyurl='myproxy.cern.ch';"
1 

Considered successful if the counter is set to 1.

bug #75223: wrong reason logged Implemented

Test description.

submit a jdl like this one:

executable="/bin/sleep";
arguments="6";
Stdoutput = "std.out";
OutputSandbox = {"std.out"};
requirements=regexp("8443/cream", other.GlueCEUniqueID);
ShallowRetryCount = 2;
prologue="/bin/false";

and look for the Done failed reason in the logging info:

[ale@cream-12 UI]$ glite-wms-job-logging-info -v 2 --event Done https://devel09.cnaf.infn.it:9000/T...

===================== glite-wms-job-logging-info Success =====================

LOGGING INFORMATION:

Printing info for the Job : https://devel09.cnaf.infn.it:9000/T...

---
Event: Done
- Arrived = Thu Apr 21 13:31:27 2011 CEST
- Exit code = 0
- Host = lcg1371.gridpp.rl.ac.uk
- Reason = Prologue failed with error 1
- Source = LRMS
- Status code = FAILED
- Timestamp = Thu Apr 21 13:31:27 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
---
Event: Done
- Arrived = Thu Apr 21 13:32:27 2011 CEST
- Exit code = 0
- Host = devel09.cnaf.infn.it
- Reason = Prologue failed with error 1
- Source = LogMonitor
- Status code = FAILED
- Timestamp = Thu Apr 21 13:32:27 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy 

Considered successful if the Reason contains only "Prologue failed with error 1".

bug #75368: ICE should log a DONE_FAILED to LB every time the job is going to be resubmitted Implemented

Test description.

Submit a jdl to a failure Cream CE (i.e. a CE where you know that for some reasons job will fail) and check the logging-info, before "Resubmission" event ICE should log a "Done Failed" with the failure reason.

Event: Done
- Arrived = Thu Apr 21 13:24:09 2011 CEST
- Exit code = 0
- Host = devel09.cnaf.infn.it
- Reason = Transfer to CREAM failed due to exception: Failed to create a delegation id for job https://devel09.cnaf.infn.it:9000/6...: reason is Received NULL fault; the error is due to another cause: FaultString=[] - FaultCode=[SOAP-ENV:Server.generalException] - FaultSubCode=[SOAP-ENV:Server.generalException] - FaultDetail=[<ns1:DelegationException><msg>User C=IT,O=INFN,OU=Personal Certificate,L=Padova,CN=Alessio Gianelle not authorized for operation {http://www.gridsite.org/namespaces/...</msg></ns1:DelegationException><ns2:exceptionName xmlns:ns2="http://xml.apache.org/axis/">org.glite.security.delegation.DelegationException</ns2:exceptionName><ns3:hostname xmlns:ns3="http://xml.apache.org/axis/">foam.grid.kiae.ru</ns3:hostname>]
- Source = LogMonitor
- Status code = FAILED
- Timestamp = Thu Apr 21 13:24:09 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy
---
Event: Resubmission
- Arrived = Thu Apr 21 13:24:09 2011 CEST
- Host = devel09.cnaf.infn.it
- Reason = Resubmitting because of exception Failed to create a delegation id for job https://devel09.cnaf.infn.it:9000/6...: reason is Received NULL fault; the error is due to another cause: FaultString=[] - FaultCode=[SOAP-ENV:Server.generalException] - FaultSubCode=[SOAP-ENV:Server.generalException] - FaultDetail=[<ns1:DelegationException><msg>User C=IT,O=INFN,OU=Personal Certificate,L=Padova,CN=Alessio Gianelle not authorized for operation {http://www.gridsite.org/namespaces/...</msg></ns1:DelegationException><ns2:exceptionName xmlns:ns2="http://xml.apache.org/axis/">org.glite.security.delegation.DelegationException</ns2:exceptionName><ns3:hostname xmlns:ns3="http://xml.apache.org/axis/">foam.grid.kiae.ru</ns3:hostname>] CEUrl https://foam.grid.kiae.ru:8443/ce-c...
- Result = WILLRESUB
- Source = LogMonitor
- Tag = unavailable
- Timestamp = Thu Apr 21 13:24:09 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy

bug #75402: Synchronization loss between real validity of proxy and exp. time saved in ICE's database Not implemented

Test description.

Rarely means that is not easy to reproduce, so we consider it hopefully fixed.

bug #77004: Wrong myproxyserver string processing in ICE Implemented

Test description.

Submit this jdl: executable="/bin/sleep"; arguments="6"; Stdoutput = "std.out"; OutputSandbox = {"std.out"}; requirements=regexp("8443/cream", other.GlueCEUniqueID); MyProxyServer = " ";

Then check the database's table "proxy" of ice in the WMS to see if the proxy is associated to a fake myproxy server:

[root@devel09 persist_dir]# sqlite3 ice.db "select * from proxy;" /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle-/dteam/Role=NULL/Capability=NULL||/var/ice/persist_dir/B23D0D7177A8B6234F1985493FA09FF41A4FA98C.proxy|1303415475|0

(the name of myproxy server should be in the second field)

bug #77055: "MyProxyServer: wrong type caught for attribute" for parametric jobs Implemented

Test description.

[ale@cream-12 UI]$ cat param.jdl
// Test parametric job.
// The standar Output and Error are parametrizied such as the
// enviroment variable: MYPATH_PARAM_.
[
JobType = "parametric";
Executable = "/bin/env";
Environment = {"MYPATH_PARAM_=$PATH:/bin:/usr/bin:$HOME"};
StdOutput = "echo_PARAM_.out";
StdError = "echo_PARAM_.err";
OutputSandbox = {"echo_PARAM_.out","echo_PARAM_.err"};
ParameterStart = 0;
ParameterStep = 2;
Parameters = 10;
myproxyserver="";
]
[ale@cream-12 UI]$ glite-wms-job-submit -a -c etc/wmp_cream-44.conf -o test.out param.jdl

Connecting to the service https://cream-44.pd.infn.it:7443/gl...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://cream-44.pd.infn.it:9000/H5...

The job identifier has been saved in the following file:
/home/ale/UI/test.out

========================================================================== 

Considered successful if the job is correctly submitted.

bug #77694: Resource BDII for WMS needs to be revisit Implemented

Test description.

[glite@cream-44 ~]$ cat /opt/glite/yaim/node-info.d/glite-wms | grep config_gip_only [glite@cream-44 ~]$

Considered successful if the function config_gip_only has been removed.

bug #77876: While purging DAGs/Collections the CLEAR event is only logged for the parent node Implemented

Test description.

Forcing a run of the cron purger you can then check the logging info of a "parent" node:

[ale@cream-12 UI]$ glite-wms-job-logging-info -v 2 https://cream-44.pd.infn.it:9000/vR...

===================== glite-wms-job-logging-info Success =====================

LOGGING INFORMATION:

Printing info for the Job : https://cream-44.pd.infn.it:9000/vR...

---
Event: RegJob
- Arrived = Wed Apr 20 14:01:25 2011 CEST
- Host = cream-44.pd.infn.it
- Jobtype = COLLECTION
- Ns = https://cream-44.pd.infn.it:7443/gl...
- Nsubjobs = 100
- Seed = WMPROXY
- Source = NetworkServer
- Src instance = https://cream-44.pd.infn.it:7443/gl...
- Timestamp = Wed Apr 20 14:01:25 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
---
Event: Accepted
- Arrived = Wed Apr 20 14:01:34 2011 CEST
- From = NetworkServer
- From host = cream-12.pd.infn.it
- Host = cream-44.pd.infn.it
- Source = NetworkServer
- Src instance = https://cream-44.pd.infn.it:7443/gl...
- Timestamp = Wed Apr 20 14:01:34 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
---
Event: EnQueued
- Arrived = Wed Apr 20 14:01:34 2011 CEST
- Host = cream-44.pd.infn.it
- Queue = /var/workload_manager/jobdir
- Result = START
- Source = NetworkServer
- Src instance = https://cream-44.pd.infn.it:7443/gl...
- Timestamp = Wed Apr 20 14:01:34 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
---
Event: EnQueued
- Arrived = Wed Apr 20 14:01:57 2011 CEST
- Host = cream-44.pd.infn.it
- Queue = /var/workload_manager/jobdir
- Result = OK
- Source = NetworkServer
- Src instance = https://cream-44.pd.infn.it:7443/gl...
- Timestamp = Wed Apr 20 14:01:57 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
---
Event: DeQueued
- Arrived = Wed Apr 20 14:05:03 2011 CEST
- Host = cream-44.pd.infn.it
- Queue = /var/workload_manager/jobdir
- Source = WorkloadManager
- Src instance = 24901
- Timestamp = Wed Apr 20 14:05:03 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy
---
Event: CollectionState
- Arrived = Wed Apr 20 14:05:11 2011 CEST
- Child = https://cream-44.pd.infn.it:9000/_y...
- Child event = Running
- Done code = 0
- Histogram = Submitted=0, Waiting=0, Ready=0, Scheduled=0, Running=1, Done=0, Cleared=0, Aborted=0, Cancelled=0, Unknown=99, Purged=0
- Host = cream-44.pd.infn.it
- Source = LBServer
- State = Running
- Timestamp = Wed Apr 20 14:05:11 2011 CEST
- User = /C=IT/O=INFN/OU=Host/L=Padova/CN=cream-44.pd.infn.it
---
Event: CollectionState
- Arrived = Wed Apr 20 14:29:59 2011 CEST
- Child = https://cream-44.pd.infn.it:9000/H3...
- Child event = Done
- Done code = 0
- Histogram = Submitted=0, Waiting=0, Ready=0, Scheduled=0, Running=0, Done=82, Cleared=0, Aborted=7, Cancelled=0, Unknown=11, Purged=0
- Host = cream-44.pd.infn.it
- Source = LBServer
- State = Waiting
- Timestamp = Wed Apr 20 14:29:59 2011 CEST
- User = LBProxy
---
Event: Clear
- Arrived = Thu Apr 21 10:13:20 2011 CEST
- Host = cream-44.pd.infn.it
- Reason = 2
- Source = NetworkServer
- Src instance = 20407
- Timestamp = Thu Apr 21 10:13:20 2011 CEST
- User = /C=IT/O=INFN/OU=Host/L=Padova/CN=cream-44.pd.infn.it/CN=431015228
==========================================================================

Or you can check the logging-info on a parent node after the output retrieval:

[ale@cream-12 UI]$ glite-wms-job-logging-info -v 2 https://cream-44.pd.infn.it:9000/tn...

===================== glite-wms-job-logging-info Success =====================

LOGGING INFORMATION:

Printing info for the Job : https://cream-44.pd.infn.it:9000/tn...

---
Event: RegJob
- Arrived = Thu Apr 21 10:18:34 2011 CEST
- Host = cream-44.pd.infn.it
- Jobtype = COLLECTION
- Ns = https://cream-44.pd.infn.it:7443/gl...
- Nsubjobs = 3
- Seed = WMPROXY
- Source = NetworkServer
- Src instance = https://cream-44.pd.infn.it:7443/gl...
- Timestamp = Thu Apr 21 10:18:34 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
---
Event: Accepted
- Arrived = Thu Apr 21 10:18:34 2011 CEST
- From = NetworkServer
- From host = cream-12.pd.infn.it
- Host = cream-44.pd.infn.it
- Source = NetworkServer
- Src instance = https://cream-44.pd.infn.it:7443/gl...
- Timestamp = Thu Apr 21 10:18:34 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
---
Event: EnQueued
- Arrived = Thu Apr 21 10:18:34 2011 CEST
- Host = cream-44.pd.infn.it
- Queue = /var/workload_manager/jobdir
- Result = START
- Source = NetworkServer
- Src instance = https://cream-44.pd.infn.it:7443/gl...
- Timestamp = Thu Apr 21 10:18:34 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
---
Event: EnQueued
- Arrived = Thu Apr 21 10:18:35 2011 CEST
- Host = cream-44.pd.infn.it
- Queue = /var/workload_manager/jobdir
- Result = OK
- Source = NetworkServer
- Src instance = https://cream-44.pd.infn.it:7443/gl...
- Timestamp = Thu Apr 21 10:18:35 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
---
Event: DeQueued
- Arrived = Thu Apr 21 10:18:36 2011 CEST
- Host = cream-44.pd.infn.it
- Queue = /var/workload_manager/jobdir
- Source = WorkloadManager
- Src instance = 15496
- Timestamp = Thu Apr 21 10:18:36 2011 CEST
- User = /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy
---
Event: CollectionState
- Arrived = Thu Apr 21 10:18:42 2011 CEST
- Child = https://cream-44.pd.infn.it:9000/d_...
- Child event = Running
- Done code = 0
- Histogram = Submitted=0, Waiting=0, Ready=0, Scheduled=0, Running=1, Done=0, Cleared=0, Aborted=0, Cancelled=0, Unknown=2, Purged=0
- Host = cream-44.pd.infn.it
- Source = LBServer
- State = Running
- Timestamp = Thu Apr 21 10:18:42 2011 CEST
- User = /C=IT/O=INFN/OU=Host/L=Padova/CN=cream-44.pd.infn.it
---
Event: CollectionState
- Arrived = Thu Apr 21 10:19:50 2011 CEST
- Child = https://cream-44.pd.infn.it:9000/zo...
- Child event = Done
- Done code = 0
- Histogram = Submitted=0, Waiting=0, Ready=0, Scheduled=0, Running=0, Done=2, Cleared=0, Aborted=0, Cancelled=0, Unknown=1, Purged=0
- Host = cream-44.pd.infn.it
- Source = LBServer
- State = Waiting
- Timestamp = Thu Apr 21 10:19:50 2011 CEST
- User = LBProxy
---
Event: CollectionState
- Arrived = Thu Apr 21 10:24:12 2011 CEST
- Child = https://cream-44.pd.infn.it:9000/as...
- Child event = Running
- Done code = 0
- Histogram = Submitted=0, Waiting=0, Ready=0, Scheduled=0, Running=1, Done=2, Cleared=0, Aborted=0, Cancelled=0, Unknown=0, Purged=0
- Host = cream-44.pd.infn.it
- Source = LBServer
- State = Running
- Timestamp = Thu Apr 21 10:24:12 2011 CEST
- User = /C=IT/O=INFN/OU=Host/L=Padova/CN=cream-44.pd.infn.it
---
Event: CollectionState
- Arrived = Thu Apr 21 10:28:04 2011 CEST
- Child = https://cream-44.pd.infn.it:9000/as...
- Child event = Done
- Done code = 0
- Histogram = Submitted=0, Waiting=0, Ready=0, Scheduled=0, Running=0, Done=3, Cleared=0, Aborted=0, Cancelled=0, Unknown=0, Purged=0
- Host = cream-44.pd.infn.it
- Source = LBServer
- State = Done
- Timestamp = Thu Apr 21 10:28:04 2011 CEST
- User = LBProxy
---
Event: CollectionState
- Arrived = Thu Apr 21 10:33:00 2011 CEST
- Child = https://cream-44.pd.infn.it:9000/zo...
- Child event = Clear
- Done code = 0
- Histogram = Submitted=0, Waiting=0, Ready=0, Scheduled=0, Running=0, Done=0, Cleared=3, Aborted=0, Cancelled=0, Unknown=0, Purged=0
- Host = cream-44.pd.infn.it
- Source = LBServer
- State = Cleared
- Timestamp = Thu Apr 21 10:33:00 2011 CEST
- User = LBProxy
---
Event: Clear
- Arrived = Thu Apr 21 10:33:00 2011 CEST
- Host = cream-44.pd.infn.it
- Reason = 1
- Source = NetworkServer
- Src instance = 21720
- Timestamp = Thu Apr 21 10:33:00 2011 CEST
- User = /C=IT/O=INFN/OU=Host/L=Padova/CN=cream-44.pd.infn.it/CN=431015228 

bug #78030: Alternative GLITE_WMS_LOG_DESTINATION in the jobwrapper Not implemented

Test description.

Stop the gatekeeper on the CE and check if the JW is able to log (looking for LRMS events into the logging-info).

bug #78047: LB Query timeout Not implemented

Test description.

Submit a lot of jobs setting EnableWmsFeedback = true; in the jdl to trigger a replanning and then looking into the wm log file:

20 Apr, 14:17:10 -I: [Info] operator()(/home/condor/execute/dir_24787/userdir/emi.wms.wms-manager/src/replanner.cpp:237): created replanning request for job https://cream-44.pd.infn.it:9000/JE... with token /var/SandboxDir/JE/https_3a_2f_2fcream-44.pd.infn.it_3a9000_2fJE5axeaAk2SU_5fG1ij5ljNg/token.txt_1

you should see that replanning works as expected.

[ale@cream-12 UI]$ glite-wms-job-status https://cream-44.pd.infn.it:9000/JE...

======================= glite-wms-job-status Success =====================
BOOKKEEPING INFORMATION:

Status info for the Job : https://cream-44.pd.infn.it:9000/JE...
Current Status: Done (Success)
Logged Reason(s):
- Cannot take token
- job completed
- Job Terminated Successfully
Exit code: 0
Status Reason: Job Terminated Successfully
Destination: ce202.cern.ch:8443/cream-lsf-grid_2nh_dteam
Submitted: Wed Apr 20 14:01:25 2011 CEST
Parent Job: https://cream-44.pd.infn.it:9000/vR...
==========================================================================

bug #78406: [ yaim-wms ] yaim should set IsmIiLDAPCEFilterExt according to the supported VO(s) Implemented

Test description.

[root@devel11 libexec]# grep IsmIiLDAPCEFilterExt /etc/glite-wms/glite_wms.conf
IsmIiLDAPCEFilterExt = "(|(GlueCEAccessControlBaseRule=VO:atlas)(GlueCEAccessControlBaseRule=VOMS:/atlas/)(GlueCEAccessControlBaseRule=VO:alice)(GlueCEAccessControlBaseRule=VOMS:/alice/)(GlueCEAccessControlBaseRule=VO:lhcb)(GlueCEAccessControlBaseRule=VOMS:/lhcb/)(GlueCEAccessControlBaseRule=VO:cms)(GlueCEAccessControlBaseRule=VOMS:/cms/)(GlueCEAccessControlBaseRule=VO:dteam)(GlueCEAccessControlBaseRule=VOMS:/dteam/)(GlueCEAccessControlBaseRule=VO:biomed)(GlueCEAccessControlBaseRule=VOMS:/biomed/)(GlueCEAccessControlBaseRule=VO:ops)(GlueCEAccessControlBaseRule=VOMS:/ops/)(GlueCEAccessControlBaseRule=VO:infngrid)(GlueCEAccessControlBaseRule=VOMS:/infngrid/))"; 

bug #78484: [ YAIM_WMS ] Multiple parameter configuration added in condor_config.local Implemented

Test description.

[root@cream-44 ~]# cat /opt/condor-c/local.cream-44/condor_config.local | grep GRID_MONITOR
GRID_MONITOR_HEARTBEAT_TIMEOUT = 300
GRID_MONITOR_RETRY_DURATION = 31536000
GRID_MONITOR = $(RELEASE_DIR)/libexec/glite/grid_monitor.sh
ENABLE_GRID_MONITOR = TRUE

[root@cream-44 ~]# /opt/glite/yaim/bin/yaim -c -s /usr/local/nfs/jra1_test/cream-44/site-info_wms.def -n WMS
WARNING:
WARNING: *****************************************************************************
WARNING: Your siteinfo directory is world readable. This is generally a bad idea
WARNING: as configuration files may contain passwords and other sensitive information.
WARNING: *****************************************************************************
WARNING:
[ ... ]
DEBUG: Skipping function: config_glite_initd_setenv because it is not defined
INFO: Executing function: config_glite_initd
INFO: Configuration Complete. [ OK ]
INFO: YAIM terminated succesfully.

[root@cream-44 ~]# cat /opt/condor-c/local.cream-44/condor_config.local | grep GRID_MONITOR
GRID_MONITOR_HEARTBEAT_TIMEOUT = 300
GRID_MONITOR_RETRY_DURATION = 31536000
GRID_MONITOR = $(RELEASE_DIR)/libexec/glite/grid_monitor.sh
ENABLE_GRID_MONITOR = TRUE

Considered successful if the line GRID_MONITOR is not added after configuring.

bug #79141: various bugs about parametric jobs Implemented

Test description.

[ale@cream-12 UI]$ cat 79141.jdl
[
JobType = "Parametric";
Parameters = 1;
ParameterStart = 0;
ParameterStep = 1;
Executable = "/bin/ls";
requirements = other.GlueCEStateStatus == "Production";
Arguments = "-la";
Stdoutput = "output.txt";
InputSandbox = {"Src_PARAM_/Makefile", "Test_PARAM_"};
OutputSandbox = {"output.txt"};
]
[ale@cream-12 UI]$ glite-wms-job-submit -a -c etc/wmp_devel11.conf 79141.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/5...

==========================================================================

[ale@cream-12 UI]$ glite-wms-job-status https://devel11.cnaf.infn.it:9000/5...

======================= glite-wms-job-status Success =====================
BOOKKEEPING INFORMATION:

Status info for the Job : https://devel11.cnaf.infn.it:9000/5...
Current Status: Waiting
Submitted: Wed Apr 20 09:37:09 2011 CEST
==========================================================================

- Nodes information for:
Status info for the Job : https://devel11.cnaf.infn.it:9000/S...
Current Status: Ready
Status Reason: unavailable
Destination: ce-enmr.chemie.uni-frankfurt.de:2119/jobmanager-lcgpbs-cert
Submitted: Wed Apr 20 09:37:09 2011 CEST
==========================================================================

[ale@cream-12 UI]$ glite-wms-job-status https://devel11.cnaf.infn.it:9000/5...

======================= glite-wms-job-status Success =====================
BOOKKEEPING INFORMATION:

Status info for the Job : https://devel11.cnaf.infn.it:9000/5...
Current Status: Done (Success)
Exit code: 0
Submitted: Wed Apr 20 09:37:09 2011 CEST
==========================================================================

- Nodes information for:
Status info for the Job : https://devel11.cnaf.infn.it:9000/S...
Current Status: Done (Success)
Logged Reason(s):
-
- Job terminated successfully
Exit code: 0
Status Reason: Job terminated successfully
Destination: ce-enmr.chemie.uni-frankfurt.de:2119/jobmanager-lcgpbs-cert
Submitted: Wed Apr 20 09:37:09 2011 CEST
==========================================================================

[ale@cream-12 UI]$ glite-wms-job-output https://devel11.cnaf.infn.it:9000/5...

Connecting to the service https://devel11.cnaf.infn.it:7443/g...

================================================================================

JOB GET OUTPUT OUTCOME

Output sandbox files for the DAG/Collection :
https://devel11.cnaf.infn.it:9000/5...
have been successfully retrieved and stored in the directory:
/tmp/jobOutput/ale_5x11U1aJFoKSDbpwu4-apw

================================================================================

[ale@cream-12 UI]$ cat /tmp/jobOutput/ale_5x11U1aJFoKSDbpwu4-apw/Node_0/output.txt
total 32
drwxr-xr-x 2 dteam035 dteam 4096 Apr 20 09:38 .
drwx------ 5 dteam035 dteam 4096 Apr 20 09:37 ..
-rw-r--r-- 1 dteam035 dteam 2150 Apr 20 09:37 .BrokerInfo
-rw-r--r-- 1 dteam035 dteam 118 Apr 20 09:38 https_3a_2f_2fdevel11.cnaf.infn.it_3a9000_2fShaETXWYzsKgV_5f0sd8o9uQ.output
-rw-r--r-- 1 dteam035 dteam 0 Apr 20 09:37 Makefile
-rw-r--r-- 1 dteam035 dteam 0 Apr 20 09:38 output.txt
-rw-r--r-- 1 dteam035 dteam 0 Apr 20 09:37 Test0
-rw------- 1 dteam035 dteam 103 Apr 20 09:38 tmp.fzuGl19574
-rw------- 1 dteam035 dteam 2 Apr 20 09:38 tmp.GxlYw19573
-rw------- 1 dteam035 dteam 103 Apr 20 09:38 tmp.mwAGt19459
-rw------- 1 dteam035 dteam 0 Apr 20 09:38 tmp.OfMof19572
-rw------- 1 dteam035 dteam 2 Apr 20 09:38 tmp.oizWr19458 

Considered successful if the parametric job is successfully executed and the ouptut correctly retrieved.

bug #81376: glite-wms-wmproxy.restart.cron MUST support graceful wmp restart Implemented

Simply verify the cron job glite-wms-wmproxy.restart.cron in the WMS: the command must use the graceful option:

0 0 * * * root . /usr/libexec/grid-env.sh ; HOSTNAME=cream-44.pd.infn.it ;  /etc/init.d/glite-wms-wmproxy graceful > /var/log/glite/glite-wms-wmproxy.restart.cron.log

bug #81568: some inconsistencies in locations for logs and configuration Implemented

There are two things to check:

  • The environment variable WMS_LOCATION_LOG should be /var/log/glite
All the log file of the WMS should be written in that directory

  • File /etc/wmproxy_logrotate.conf should not exists
The file should be put in $GLITE_WMS_CONFIG_DIR

bug #82687: emi WMS problems with ISB tar file handling Implemented

Submit a jdl like this one:

[ale@cream-12 UI]$ cat 82687.jdl 
[
Executable = "/bin/ls" ;
Arguments = "-lha " ;
Stdoutput = "ls.out" ;
InputSandbox = { "supercalifragilistichespiralidose.txt", "82687.jdl"} ;
OutputSandbox = { ".BrokerInfo", "ls.out"} ;
Retrycount = 2;
ShallowRetryCount = -1;
usertags = [ bug = "#82687" ];
VirtualOrganisation="dteam";
]

With a long file name as input file (and at least two files as input):

[ale@cream-12 UI]$ glite-wms-job-submit -a -c etc/wmp_cream-45.conf 82687.jdl

Connecting to the service https://cream-45.pd.infn.it:7443/glite_wms_wmproxy_server


====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://cream-45.pd.infn.it:9000/ehWYewLsPxn9Q3RYQz64gw

=========================================================================

When the job arrives to Done (Success):

[ale@cream-12 UI]$ glite-wms-job-status https://cream-45.pd.infn.it:9000/ehWYewLsPxn9Q3RYQz64gw


======================= glite-wms-job-status Success =====================
BOOKKEEPING INFORMATION:

Status info for the Job : https://cream-45.pd.infn.it:9000/ehWYewLsPxn9Q3RYQz64gw
Current Status:     Done (Success)
Logged Reason(s):
    - 
    - 
    - Job terminated successfully
Exit code:          0
Status Reason:      Job terminated successfully
Destination:        unime-ce-01.me.pi2s2.it:2119/jobmanager-lcglsf-cert
Submitted:          Mon Jul 18 16:47:53 2011 CEST
==========================================================================

Check in the output file of the job if all the input files are listed:

[ale@cream-12 UI]$ glite-wms-job-output https://cream-45.pd.infn.it:9000/ehWYewLsPxn9Q3RYQz64gw

Connecting to the service https://cream-45.pd.infn.it:7443/glite_wms_wmproxy_server


================================================================================

            JOB GET OUTPUT OUTCOME

Output sandbox files for the job:
https://cream-45.pd.infn.it:9000/ehWYewLsPxn9Q3RYQz64gw
have been successfully retrieved and stored in the directory:
/tmp/jobOutput/ale_ehWYewLsPxn9Q3RYQz64gw

================================================================================


[ale@cream-12 UI]$ cat /tmp/jobOutput/ale_ehWYewLsPxn9Q3RYQz64gw/ls.out 
total 32K
drwxr-xr-x 2 dteam033 dteam 4.0K Jul 18 16:53 .
drwx------ 5 dteam033 dteam 4.0K Jul 18 16:50 ..
-rw-r--r-- 1 dteam033 dteam  293 Jul 18 16:52 82687.jdl
-rw-r--r-- 1 dteam033 dteam 3.6K Jul 18 16:52 .BrokerInfo
-rw-r--r-- 1 dteam033 dteam  153 Jul 18 16:53 https_3a_2f_2fcream-45.pd.infn.it_3a9000_2fehWYewLsPxn9Q3RYQz64gw.output
-rw-r--r-- 1 dteam033 dteam    0 Jul 18 16:53 ls.out
-rw-r--r-- 1 dteam033 dteam   17 Jul 18 16:51 supercalifragilistichespiralidose.txt
-rw------- 1 dteam033 dteam    2 Jul 18 16:53 tmp.cgesDD8989
-rw------- 1 dteam033 dteam  103 Jul 18 16:53 tmp.ODAMIs8990
-rw------- 1 dteam033 dteam    0 Jul 18 16:53 tmp.taaebr8988

bug #82776: typo in WMS jobwrapper Implemented

Simply check the JobWrapper template on the WMS:

[root@cream-45 ~]# cat  /usr/share/glite-wms/jobwrapper.template.sh  | grep "Taken token ${__shallow_resubmission_token}"
      push_in_LM_done_reason "Taken token ${__shallow_resubmission_token}"

Check if a space is inserted beetween push_in_LM_done_reason and the following word "Taken

bug #82983: [yaim-wms] authorization problem in WMS EMI-1 Implemented

Simply check that in the file glite_wms_wmproxy.gacl file there are not entry with the word ROLE in upper case:

[root@devel09 ~]# grep ROLE /etc/glite-wms/glite_wms_wmproxy.gacl 
[root@devel09 ~]#

bug #83062: [yaim-wms] A different template should be used for glue2 publication Implemented

Query the resource BDII running on the CREAM-CE for Glue2:

ldapsearch -h  -x -p 2170 -b "o=glue"

Check if a GLUE2Service objectclass with GLUE2ServiceType: org.glite.wms.WMProxy is published. Check if a GLUE2Endpoint objectclass with GLUE2EndpointInterfaceName: org.glite.wms.WMProxy is published.

Run GlueValidor and check that no errors are reported

bug #84155: Internal proxy structure convertion error in ICE Not implemented

Test description.

Take a long jdl where MyProxyServer attribute is set and that require a cream CE as destination; like this one:

[ale@cream-12 UI]$ cat 84155.jdl 
[
Executable = "/bin/sleep" ;
InputSandbox = { "/home/ale/UI/data/pippo" };
Arguments="900";
Environment={"VARIABLE=test"};
Myproxyserver="myproxy.cnaf.infn.it";
requirements = RegExp("/cream-",other.GlueCEUniqueID);
]

Submit it with a shorter proxy:

[ale@cream-12 UI]$ voms-proxy-init -voms dteam  --valid 00:15
Your identity: /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
Creating temporary proxy ........................ Done
Contacting  voms.hellasgrid.gr:15004 [/C=GR/O=HellasGrid/OU=hellasgrid.gr/CN=voms.hellasgrid.gr] "dteam" Done
Creating proxy .................................................................................... Done

Your proxy is valid until Tue Jul 19 13:34:37 2011
[ale@cream-12 UI]$ glite-wms-job-submit -a -c etc/wmp_devel11.conf 84155.jdl

Connecting to the service https://devel11.cnaf.infn.it:7443/glite_wms_wmproxy_server


====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/AszPQkHVQRVNgOE9g25NpA

==========================================================================

Check that its proxy is valid for no more that 15 minutes:

[ale@cream-12 UI]$ glite-wms-job-info -p https://devel11.cnaf.infn.it:9000/AszPQkHVQRVNgOE9g25NpA

Connecting to the service https://devel11.cnaf.infn.it:7443/glite_wms_wmproxy_server


======================= glite-wms-job-info Success =======================

Your proxy delegated to the endpoint https://devel11.cnaf.infn.it:7443/glite_wms_wmproxy_server
 for the job https://devel11.cnaf.infn.it:9000/AszPQkHVQRVNgOE9g25NpA :
Subject     : /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy
Issuer      : /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy
Identity    : /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy
Type        : proxy
Strength    : 512
StartDate   : 19 July 2011 - 13:14:41
Expiration  : 19 July 2011 - 13:33:41
Timeleft    : 13 min 51 sec 
=== VO dteam extension information ===
VO          : dteam
Subject     : /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
Issuer      : /C=GR/O=HellasGrid/OU=hellasgrid.gr/CN=voms.hellasgrid.gr
URI         : voms.hellasgrid.gr:15004
Attribute   : /dteam/Role=NULL/Capability=NULL
Attribute   : /dteam/NGI_IT/Role=NULL/Capability=NULL
StartTime   : 19 July 2011 - 13:19:37
Expiration  : 19 July 2011 - 13:34:37
Timeleft    : 14 min 47 sec 

==========================================================================

Wait until it finishes successfully:

[ale@cream-12 ~]$ glite-wms-job-status https://devel11.cnaf.infn.it:9000/AszPQkHVQRVNgOE9g25NpA


======================= glite-wms-job-status Success =====================
BOOKKEEPING INFORMATION:

Status info for the Job : https://devel11.cnaf.infn.it:9000/AszPQkHVQRVNgOE9g25NpA
Current Status:     Done (Success)
Logged Reason(s):
    - job completed
    - Job Terminated Successfully
Exit code:          0
Status Reason:      Job Terminated Successfully
Destination:        grid-cr4.desy.de:8443/cream-pbs-operations
Submitted:          Tue Jul 19 13:19:42 2011 CEST
==========================================================================

We can also check directly the command on the WMS. First of all find the proxy associated with the job in the persist directory of ice:

[root@devel11 ~]# openssl x509 -noout -text -in /var/ice/persist_dir/2A9DAF04C398C21D6ADF7E884BC192ED95AF554C.betterproxy 
Certificate:
    Data:
        Version: 3 (0x2)
        Serial Number: 1311074381 (0x4e25684d)
        Signature Algorithm: md5WithRSAEncryption
        Issuer: C=IT, O=INFN, OU=Personal Certificate, L=Padova, CN=Alessio Gianelle, CN=proxy
        Validity
            Not Before: Jul 19 11:14:41 2011 GMT
            Not After : Jul 19 11:33:41 2011 GMT
        Subject: C=IT, O=INFN, OU=Personal Certificate, L=Padova, CN=Alessio Gianelle, CN=proxy, CN=proxy
        Subject Public Key Info:
            Public Key Algorithm: rsaEncryption
            RSA Public Key: (512 bit)
                Modulus (512 bit):
                    00:e3:ac:3f:29:d6:b2:51:8d:28:16:77:83:5a:c7:
                    6e:d8:fd:83:e2:23:8a:cd:be:c4:99:a1:84:8f:4f:
                    35:60:8b:3a:a7:30:e4:fd:ea:6a:cd:1a:c2:ea:ca:
                    30:94:73:c8:de:15:64:1f:7b:ca:94:40:20:cb:ee:
                    18:de:c7:3f:4d
                Exponent: 65537 (0x10001)
    Signature Algorithm: md5WithRSAEncryption
        69:dc:99:25:73:f5:c8:74:f8:5e:2f:8a:80:23:b0:d4:f5:1f:
        d8:44:a3:2b:e3:99:47:9c:88:17:d1:ba:8f:27:eb:d3:6e:79:
        e5:9f:11:46:59:5c:b2:ad:2d:c8:65:b3:dc:1f:f9:3f:6b:88:
        40:d1:6e:6b:5e:d0:15:1e:52:7a:7e:47:dd:81:51:47:09:0b:
        68:b2:81:d6:d7:41:43:55:b4:36:0f:9c:69:9d:f7:ec:63:8e:
        2f:8d:4b:db:2e:16:58:7a:25:47:2d:e6:c7:16:97:27:04:16:
        de:93:f8:3e:38:46:57:7d:96:b9:cd:b5:91:25:d0:4f:38:cb:
        ce:91

Then logged as user glite and execute this command: export X509_USER_CERT=/var/wms.proxy; export X509_USER_KEY=/var/wms.proxy; /usr/bin/glite-wms-ice-proxy-renew -s myproxy.cnaf.infn.it -p /var/ice/persist_dir/<user proxy>.betterproxy -o <output file> where /var/ice/persist_dir/<user proxy>.betterproxy is the proxy associate to the job and in <output file> you should find the renewed proxy:

[root@devel11 ~]# su - glite
[glite@devel11 ~]$ export X509_USER_CERT=/var/wms.proxy; export X509_USER_KEY=/var/wms.proxy; /usr/bin/glite-wms-ice-proxy-renew -s myproxy.cnaf.infn.it -p /var/ice/persist_dir/2A9DAF04C398C21D6ADF7E884BC192ED95AF554C.betterproxy -o /tmp/test.proxy
/var/ice/persist_dir/2A9DAF04C398C21D6ADF7E884BC192ED95AF554C.betterproxy.voms.6EPXDC

Check the validity of the new proxy:

[glite@devel11 ~]$ openssl x509 -noout  -dates -in /tmp/test.proxy
notBefore=Jul 19 11:18:12 2011 GMT
notAfter=Jul 19 21:23:18 2011 GMT

bug #85071: wmproxy authorization breaks with longer proxy chain Not implemented

Create a delega on a myproxy server:

[ale@cream-12 proxy]$ myproxy-init -d -s myproxy.cnaf.infn.it
Your identity: /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
Creating proxy .................................................... Done
Proxy Verify OK
Your proxy is valid until: Tue Sep  6 10:53:54 2011
Enter MyProxy pass phrase:
Verifying - Enter MyProxy pass phrase:
A proxy valid for 168 hours (7.0 days) for user /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle now exists on myproxy.cnaf.infn.it.

Then retrieve a new proxy from the same myproxy server:

[ale@cream-12 proxy]$  myproxy-logon -d -s myproxy.cnaf.infn.it --voms dteam
Enter MyProxy pass phrase:
Your identity: /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy/CN=proxy
Contacting  voms.hellasgrid.gr:15004 [/C=GR/O=HellasGrid/OU=hellasgrid.gr/CN=voms.hellasgrid.gr] "dteam" Done
Creating proxy ................................................................................. Done

Your proxy is valid until Tue Aug 30 21:54:17 2011
A credential has been received for user /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle in /tmp/x509up_u500.

Now you should have a proxy with a longer chain (see the number of "CN=proxy" token):

[ale@cream-12 proxy]$ voms-proxy-info -all
subject   : /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy/CN=proxy/CN=proxy
issuer    : /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy/CN=proxy
identity  : /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy/CN=proxy
type      : proxy
strength  : 2048 bits
path      : /tmp/x509up_u500
timeleft  : 10:59:49
key usage : Digital Signature, Key Encipherment, Data Encipherment
=== VO dteam extension information ===
VO        : dteam
subject   : /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle
issuer    : /C=GR/O=HellasGrid/OU=hellasgrid.gr/CN=voms.hellasgrid.gr
attribute : /dteam/Role=NULL/Capability=NULL
attribute : /dteam/NGI_IT/Role=NULL/Capability=NULL
timeleft  : 11:58:48
uri       : voms.hellasgrid.gr:15004

Then the submission to WMS using this proxy should not failed.

[ale@cream-12 proxy]$ glite-wms-job-submit -a -c ~/UI/etc/wmp_devel11.conf ~/UI/test.jdl 

Connecting to the service https://devel11.cnaf.infn.it:7443/glite_wms_wmproxy_server


====================== glite-wms-job-submit Success ======================

The job has been successfully submitted to the WMProxy
Your job identifier is:

https://devel11.cnaf.infn.it:9000/aog8s2q5Lnh0-pAjyaAGAg

==========================================================================

bug #85327: [yaim-wms] glite-wms-wmproxy.restart.cron missing HOSTNAME environment variable Implemented

Check two things:

  • The cron job glite-wms-wmproxy.restart.cron should contains the setting of HOSTNAME before the graceful restarting of wmproxy service

  • If you issue a service gLite restart you should not see, after wmproxy starting, this warning:
    [warn] PassEnv variable HOSTNAME was undefined
    

bug #86485: ICE doesn't send the iceId to CREAM Not implemented

To see if the iceid arrives to CREAM, you have to set CREAM in debug mode: set log4j.rootLogger=debug in /etc/glite-ce-cream/log4j.properties, then submit a job to this CREAM through the WMS, in the log file of CREAM you should find an entry like this one (grep string "selectToRetrieveJobStatusQuery"):

05 Oct 2011 13:35:46,963 DEBUG org.glite.ce.cream.jobmanagement.db.table.JobStatusTable (JobStatusTable.java:419) - (TP-Processor25) selectToRetrieveJobStatusQuery = select job_status.type AS type,job_status.exitCode AS exitCode,job_status.failureReason AS failureReason,job_status.description AS description,job_status.time_stamp AS time_stamp,job_status.jobId AS jobId,job_status.id AS id,job.gridJobId AS gridJobId,job.workerNode AS workerNode from job_status, job where job_status.jobId = job.id and job_status.type IN ( '10', '5', '8', '7', '6', '4', '3') and job.userId = '_C_IT_O_INFN_OU_Personal_Certificate_L_Padova_CN_Alessio_Gianelle_dteam_Role_NULL_Capability_NULL' and job.iceId = 'C_IT_O_INFN_OU_Host_L_Padova_CN_cream-44.pd.infn.it' and job_status.id >= 0 and time_stamp >= '2011-10-05 11:59:42.0' ORDER BY job_status.id limit 500

Where the attribute job.iceId is set tipically with the WMS DN. You can also check the creamdb on CREAM using this query: select iceId from job where id="<CREAMID>";

mysql> select iceId from job where id="CREAM089157286";
+-----------------------------------------------------+
| iceId                                               |
+-----------------------------------------------------+
| C_IT_O_INFN_OU_Host_L_Padova_CN_cream-44.pd.infn.it | 
+-----------------------------------------------------+
1 row in set (0.00 sec)

On the WMS you can also check in the ice log file the number of the REGISTERED using this query: grep -c "CREAM Returned CREAM-JOBID" $WMS_LOCATION_LOG/ice.log; take in mind that you have to compare the same time interval (i.e. check the range of time covered by the ice.log file and then use option -f and -t of queryStats)

bug #86490: EMI WMS stops accepting jobs after 31999th submission for the same DN (on ext3) Implemented

There is a cron job glite-wms-wmproxy-purge-proxycache.cron which usually run every 6 hours to cleaning the $WMS_LOCATION_VAR/proxycache. Inside this directory there should not empty subdirectories older than 6 hours, or expired proxies from more than 6 hours.

bug #86682: [yaim-wms] yaim changes ownership of /var Implemented

SImply check the ownership of the system directories: /var and /var/log, they should be root.root

bug #86769: EMI WMS Update 6: WMS 'hangs' after several hours of operation Not implemented

This bug is not reproducible. You can try to stress the WMS submitting for several hours jobs which require Data management. This could be an example of jdl:

   
###########################################
#      JDL with Data Requirements         #
###########################################

Executable = "/bin/ls";
Arguments = "-la";
StdOutput = "std.out";
StdError = "std.err";
Prologue = "prologue.sh";

Environment={"LFC_HOST=lfc.gridpp.rl.ac.uk", "LFC_HOME=lfc.gridpp.rl.ac.uk:/grid/t2k.org"};
InputSandbox = {"prologue.sh"};
OutputSandbox = {"std.out", "std.err","prologue.out"};

requirements = true;

DataRequirements = {
[
DataCatalogType = "DLI";
DataCatalog ="http://lfc.gridpp.rl.ac.uk:8085/"; 
InputData = { "lfn:/grid/t2k.org/nd280/raw/ND280/ND280/00005000_00005999/nd280_00005000_0002.daq.mid.gz" };
]
};

DataAccessProtocol = "gsiftp";

bug #86956: Match-matching of requests with access to catalogues can cause the WM to crash Not implemented

Like the previous one: BUG #86769, it is not easy to reproduce. You can be comfortable if the stress test describe in the previous bug does not cause a WM crash.

bug #87259: WMS purger errors in jobcontroller/logmonitor Not implemented

This bug reports two problems.

  • Purging of the JC local directories
Submit a normal job (NOT a collection), when it finishes check the log file of the LogMonitor: ${WMS_LOCATION_LOG}/logmonitor_events.log; it should reports the successfully remove of 3 files under directory SubmitFileDir/<two_digit_prefix> ( Condor.<JOBID>.submit, ClassAd.<JOBID> and JobWrapper.<JOBID>.sh ) and the directory OutputFileDir/<two_digit_prefix>/<JOBID> :
07 Oct, 16:01:04 -I- JobFilePurger::do_purge(...): Removing job directory: /var/jobcontrol/condorio/bc/https_3a_2f_2fcream-44.pd.infn.it_3a9000_2fbcpFQ-3g3c_5flDMYJqPvjOQ
07 Oct, 16:01:04 -U- JobFilePurger::do_purge(...): Removed 1 files.
07 Oct, 16:01:04 -I- JobFilePurger::do_purge(...): Removing submit file: /var/jobcontrol/submit/bc/Condor.https_3a_2f_2fcream-44.pd.infn.it_3a9000_2fbcpFQ-3g3c_5flDMYJqPvjOQ.submit
07 Oct, 16:01:04 -U- JobFilePurger::do_purge(...): Removed...
07 Oct, 16:01:04 -I- JobFilePurger::do_purge(...): Removing classad file: /var/jobcontrol/submit/bc/ClassAd.https_3a_2f_2fcream-44.pd.infn.it_3a9000_2fbcpFQ-3g3c_5flDMYJqPvjOQ
07 Oct, 16:01:04 -U- JobFilePurger::do_purge(...): Removed...
07 Oct, 16:01:04 -I- JobFilePurger::do_purge(...): Removing wrapper file: /var/jobcontrol/submit/bc/JobWrapper.https_3a_2f_2fcream-44.pd.infn.it_3a9000_2fbcpFQ-3g3c_5flDMYJqPvjOQ.sh
07 Oct, 16:01:04 -U- JobFilePurger::do_purge(...): Removed...

You have also to check that these files have been correctly removed.

  • The SBD purger fails in trapping EIDRM. To reproduce this bug you have to follow these steps:
    • Submit a job to the WMS
    • On the WMS create a copy of its SBD (pay attention to preserve files ownership)
      [root@cream-44 ~]# cd /var/SandboxDir/1u/
      [root@cream-44 1u]# ls
      https_3a_2f_2fcream-44.pd.infn.it_3a9000_2f1un8CjRsC1c-P0R4H-KYjQ
      [root@cream-44 1u]# cp -r -p https_3a_2f_2fcream-44.pd.infn.it_3a9000_2f1un8CjRsC1c-P0R4H-KYjQ/ https_3a_2f_2fcream-44.pd.infn.it_3a9000_2f1un8CjRsC1c-P0R4H-KYjQ.old
    • Cancel the job or when it finishes retrieve output files.
    • On the WMS restore its SBD
      [root@cream-44 ~]# cd /var/SandboxDir/1u/
      [root@cream-44 1u]# ls
      https_3a_2f_2fcream-44.pd.infn.it_3a9000_2f1un8CjRsC1c-P0R4H-KYjQ.old
      [root@cream-44 1u]# mv https_3a_2f_2fcream-44.pd.infn.it_3a9000_2f1un8CjRsC1c-P0R4H-KYjQ.old/ https_3a_2f_2fcream-44.pd.infn.it_3a9000_2f1un8CjRsC1c-P0R4H-KYjQ
      [root@cream-44 1u]# ls
      https_3a_2f_2fcream-44.pd.infn.it_3a9000_2f1un8CjRsC1c-P0R4H-KYjQ
    • On the WMS create a file with the jobid
      [root@cream-44 ~]# echo "https://cream-44.pd.infn.it:9000/1un8CjRsC1c-P0R4H-KYjQ" > jobid.txt
    • On the WMS use LB command to purge job (first from LBProxy then from LBServer)
      [root@cream-44 ~]# /usr/bin/glite-lb-purge -m cream-44.pd.infn.it -x -j jobid.txt 
      The jobs were not dumped.
      [root@cream-44 ~]# /usr/bin/glite-lb-purge -m cream-44.pd.infn.it -j jobid.txt 
      The jobs were not dumped.
    • Now if you try a status you should find this error message: Identifier removed (i.e. error EIDRM):
      [ale@cream-12 UI]$ glite-wms-job-status https://cream-44.pd.infn.it:9000/1un8CjRsC1c-P0R4H-KYjQ
      
      **** Error: API_NATIVE_ERROR ****  
      Error while calling the "Job:getStatus" native api 
      Unable to retrieve the status for: https://cream-44.pd.infn.it:9000/1un8CjRsC1c-P0R4H-KYjQ
      glite.lb.Exception: edg_wll_JobStatus: Identifier removed: matching job already purged
         at glite::lb::Job::status[./src/Job.cpp:87]
    • On the WMS we can now try to purge the SBD, as user glite invoke the command glite-wms-purgeStorage.sh
      [glite@cream-44 ~]$ /usr/sbin/glite-wms-purgeStorage.sh -p /var/SandboxDir/1u
      07 Oct, 17:44:46 -E: [Error] query_job_status(/home/condor/execute/dir_2715/userdir/emi.wms.wms-purger/src/purger.cpp:134): https://cream-44.pd.infn.it:9000/1un8CjRsC1c-P0R4H-KYjQ: edg_wll_JobStat [43] Identifier removed(matching job already purged)
      07 Oct, 17:44:46 -I: [Info] operator()(/home/condor/execute/dir_2715/userdir/emi.wms.wms-purger/src/purger.cpp:306): https://cream-44.pd.infn.it:9000/1un8CjRsC1c-P0R4H-KYjQ: forced removal, unknown/removed L&B job
As you can see the same error message is reported by the purger: Identifier removed (i.e. error 43) but then the SBD is forced removed.
[root@cream-44 ~]# ls /var/SandboxDir/1u/
[root@cream-44 ~]# 

Bug #87994 yaim-wms doesn't fill the attributes get_acbr and get_owner in the conf file for Glue2 Not implemented

Check that the attributes get_acbr and get_owner in /etc/glite/info/service/glite-info-glue2-wmproxy.conf are define with all the supported VOs

[root@cream-44 ~]# grep get_acbr /etc/glite/info/service/glite-info-glue2-wmproxy.conf
get_acbr = echo -e "VO:alice\nVO:atlas\nVO:cms\nVO:lhcb\nVO:dteam\nVO:testers.eu-emi.eu"
[root@cream-44 ~]# grep get_owner /etc/glite/info/service/glite-info-glue2-wmproxy.conf
get_owner = echo -e "alice\natlas\ncms\nlhcb\ndteam\ntesters.eu-emi.eu"

bug #88072: EMI-1 WMS cannot submit to ARC CEs Not implemented

Submit to an "arc" ce and check the submission cycle (submit->get-output)

bug #88450: Ice configuration section doesn't use environment variable Implemented

Check in the glite_wmc.conf file if all the paths of the log files are defined using envoronment variables:

 [root@wms ~]# cat /etc/glite-wms/glite_wms.conf | grep -i " logfile"
    logfile   =   "${WMS_LOCATION_LOG}/ice.log";
    LogFile  =  "${WMS_LOCATION_LOG}/wmproxy.log";
    LogFile  =  "${WMS_LOCATION_LOG}/networkserver_events.log";
    LogFile   =  "${WMS_LOCATION_LOG}/workload_manager_events.log";
    LogFile  =  "${WMS_LOCATION_LOG}/logmonitor_events.log";
    LogFile  =  "${WMS_LOCATION_LOG}/jobcontoller_events.log";

bug #88569: WMS: LCMAPS should log on file only and not on syslog Not implemented

Check that no messages about "lcmaps" is logged by the wmproxy on syslog during a submission, and that messages are instead saved in a dedicated log file: /var/log/glite/lcmaps.log

bug #88558: WM creates submit requests with old sequence code upon a replan Implemented

Trigger a replan then check the logged events; in the jdl logged with the enqueud event after the replanning the LB_sequence_code should be the same of the one used to logged the event:

Event: Resubmission
- Arrived                    =    Wed Nov 23 19:16:16 2011 CET
- Host                       =    devel19.cnaf.infn.it
- Level                      =    SYSTEM
- Priority                   =    synchronous
- Reason                     =    token still exists
- Result                     =    SHALLOW
- Seqcode                    =    UI=000000:NS=0000000004:WM=000007:BH=0000000000:JSS=000002:LM=000000:LRMS=000000:APP=000000:LBS=000000
- Source                     =    WorkloadManager
- Src instance               =    13159
- Tag                        =    job will be replanned
- Timestamp                  =    Wed Nov 23 19:16:16 2011 CET
- User                       =    /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy/CN=proxy/CN=proxy

[...]

Event: EnQueued
- Arrived                    =    Wed Nov 23 19:16:16 2011 CET
- Host                       =    devel19.cnaf.infn.it
- Level                      =    SYSTEM
- Priority                   =    synchronous
- Queue                      =    /var/workload_manager/jobdir
- Result                     =    OK
- Seqcode                    =    UI=000000:NS=0000000004:WM=000009:BH=0000000000:JSS=000002:LM=000000:LRMS=000000:APP=000000:LBS=000000
- Source                     =    WorkloadManager
- Src instance               =    13159
- Timestamp                  =    Wed Nov 23 19:16:16 2011 CET
- User                       =    /C=IT/O=INFN/OU=Personal Certificate/L=Padova/CN=Alessio Gianelle/CN=proxy/CN=proxy/CN=proxy/CN=proxy
- Job            =   
      
       [
        version = "1.0.0"; 
        arguments = 
         [
          ad = 
           [
            NodeName = "Node_jdl_953"; 
            FuzzyRank = true; 
            CEInfoHostName = "cream-31.pd.infn.it"; 
            OutputSandboxPath = "/var/SandboxDir/ji/https_3a_2f_2fdevel19.cnaf.infn.it_3a9000_2fjiHL26jdUVDO7Bl6BJlhEQ/output"; 
            StdOutput = "out.log"; 
            GlobusResourceContactString = "cream-31.pd.infn.it:8443/cream-pbs"; 
            QueueName = "creamtest2"; 
            ShallowRetryCount = 3; 
            RetryCount = 2; 
            Executable = "/bin/sleep"; 
            Type = "job"; 
            LB_sequence_code = "UI=000000:NS=0000000004:WM=000009:BH=0000000000:JSS=000002:LM=000000:LRMS=000000:APP=000000:LBS=000000"; 
  • new_tests.odt: Current list of regression tests implemented in the wms test suite, with verification comments to be sent back as feedback to Aristotelis

issue #WMS-104: Directory /var/condor/spool needs to be created when configuring WMS node Not implemented

To see the test method please go to the issue and look at the comments.

issue #WMS-97: UMD Verification: EMI-3 WMS 3.5.0 Problem with Argus and WMS integration Not implemented

To see the test method please go to the issue and look at the comments.

issue #WMS-94: Three major bugs in EMI2 glite-wms-wmproxy 3.4.1-0.sl6 Not implemented

To see the test method please go to the issue and look at the comments.

issue #WMS-93: glite-wms-check-daemon wrongly checks ICE status Not implemented

To see the test method please go to the issue and look at the comments.

issue #WMS-92: check the policies that keep ISBfiles in the WMS after sandbox creation Not implemented

To see the test method please go to the issue and look at the comments.

issue #WMS-87: glite_wms_wmproxy_load_monitor needlessly installed with setuid privileges Not implemented

To see the test method please go to the issue and look at the comments.

issue #WMS-85: Use yaim WMS_PEPC_RESOURCEID to set the argus resource id Not implemented

To see the test method please go to the issue and look at the comments.

issue #WMS-91: WMS to limit ISB size as a whole, not only per file Not implemented

To see the test method please go to the issue and look at the comments.

issue #WMS-90: EMI-2 WMS UI glite-wms-job-output creates files without checking VO membership first, also changing status Not implemented

To see the test method please go to the issue and look at the comments.
Topic attachments
I Attachment Action Size Date Who Comment
Unknown file formatodt new_tests.odt manage 23.9 K 2012-03-09 - 16:02 FabioCapannini Current list of regression tests implemented in the wms test suite, with verification comments to be sent back as feedback to Aristotelis
Edit | Attach | PDF | History: r32 < r31 < r30 < r29 < r28 | Backlinks | Raw View | More topic actions
Topic revision: r32 - 2013-07-16 - AlviseDorigo
 
This site is powered by the TWiki collaboration platformCopyright © 2008-2019 by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding TWiki? Send feedback