Workbrain Scalability Assessment Test Completion Report · Workbrain Scalability Assessment Test...
-
Upload
nguyenthien -
Category
Documents
-
view
222 -
download
0
Transcript of Workbrain Scalability Assessment Test Completion Report · Workbrain Scalability Assessment Test...
Workbrain Scalability Assessment
Test Completion Report
Reference Number: CD-805·0005
Stream:
Version: 2.0
Approvers
. •!.-·. -Title
Brett Matthews SDA Emwum<e>Director
.z}. Paul Hickey Program 42 Program
/1 L-------~----~D_irn_ct_or __________ ~~~~~~~--~+c7-~~ -:J~:rmc:.s &~tJ,J PI)JJ
Shared SeNice SoftJtions
Commercial-in-Confidence
Program LJ2
Approvals
PREPAREO BY:
APPROVED BY:
Amir Mahmoudi and John Sinclair Name Role
Name Role
Commercial-in-Confidence
CD-805-0005 Test Completion Report
V2.0
Date
Date
Page 2 of 13
Program 42
Document Information
Docum~nt Source
CD-805-0005 Test Completion Report
V2.0
The source of this document is named Program 42 Work:brain Scalability Assessment Tes.t
Completion Report il2.0.doc.
Author : Amir Mahmoudi and John Sinclair
Reference Number: CD-805-0005
Revision History ve~ion Revision d~te sun#n~ry of ch~·~ges Priqi~-r.y Author ~)J.rij~er ., .. 0.1 28/04/08 Initial version Amir Mahmoudi
0.2 30/04/08 Review and update John Sinclair
0.3 06/05/08 Updates based on lntf~mal review John Sinclair
1.0 15/05/08 Updated based on comments from the Amir Mahmoudi client and minor update to section 4,3.2
1.1 23/05/08 Updates the recommendations section Amir Mahmoudi
2.0 25/07/08 Final version PDO
Reviewer List I N,.ame ·
Distribution
~ ~ame . _.;-·.: I Title
Related D.ocuments Test• ~ele~~-~ Titi~· of OOcli~~rttlLocati~~
.. ···. v~rslon Typ~· .. p~te.
. .. - 5.:. '-~ . . ' ~.· .'!5-· ··- . .
Project 26/03/08 \\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS- Prime 0.1 Execution Contractor- IBM\23. Worl<brain Scalability: Assessmeni\Pian\Proiecf Execution Plan Plan\Workbrain Scalabilit~ Assessment PEP v0.1 .doc
Test Plan 23/04/08 \\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS - Prime 1.1 Contractor- IBM\23. Workbrain Scalability: Assessment\Deliverables\Test Plan\Workbrain Scalability Assessment Test Plan ver 1.1 final.doc
1 07/02/08 \\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS - Prime 0.2
Cotnmercial•in-Confidence Page 3 of 13
Program42
Test Release. Type I) ate ······
2a 06/03/08
2b 28/03/08
2c 07/04/08
3-.online 18/04/08 to 24/04/08
3-batch
CD-805-0005 Test Completion Report
V2.0
· · Title of Oqiu-Mentll ocation Versioil ... ,._ _ _ • • 1 • , , _.:
Contractor- IBM\23. Workbrain Scalability ASsessment\Pian\Test1\Results\P42 Scalability Testing Test1 Re11ort V0.2.doc ·
\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS- Prime Contractor - IBM\23. Workbrain Scalability AssessmentiPian\Test 2a1Results\Workbrain POC rostering Test 2 summa!Y of find ings. doc
\INW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS - Prime Contractor - IBM\23. Workbrain Scalabilitll Assessment\Pian\Test 2a1Results1Workbrain POC rosteri!HJ Test 2 summar:y of [!roblems encountered. doc
\INW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS- Prime Contractor- IBM\23.Workbrain Scalabilitll Assessment\Pian\Test 2a1Results\Workbrain Scalabililll - Test 2 first rostering test- Anal:tsis v0.1 .doc
\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS - Prime Contractor- IBM\23. Workbrain Scalabililll Assessment\Pian\Test 2a1Results\Workbrain Scalabili!ll - Test 2a - Anal~is vb.2:doc
\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS - Prime Contractor- IBM\23. Workbrain Scalabililll AssessmentiPian\Test 2b\Results\Workbrain POC rostering Test 2b Executive Summa[Y.doc
\\NW-CLUSTER-01 CTDATA SERVERICTDATA\DATA\Cn 17 - SSS - Prime Contractor ~ IBM\23. Workbrain Scalability Assessment\Pian\Test 2c\Results\Microsoft Word -Test2c 500VU Run1 3Server 03042008 Re(!ort.odf
\\NW-CU..iSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS- Prime Contractor - IBM\23. Workbtain Scalabilitll Assessment\Pian\Test 2c\Resuits\Microsoft Word -Test2c 1000VU Run1 3Server 03042008 Re(!ort.Qdf
\\NW-CLUSTER-01 CTDATA S!;RVERICTDATA\DATA\Cn17- SSS - Prime Contractor - IBM\23. Workbrain Scalabilitll Assessment\Pian\Test 2c\Results\Microsoft Word -Test2c 1500VU Run2 3Server NS RR 04042008 ReQort.Qdf
1\NV\i-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS- Prime Contractor" IBM\23. Workbrain Scalabililll Assessment\Pian\Test 3\Results\Test3 500VU Run1 1Server 17042008 Re[!ort.!!df
\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS- Prime Contractor- IBM\23. Workbrain Scalability Assessment\Pian\Test 3\Results\Test3 500VU Run1 .2Server 18042Q08 Re11ort.doc
\INW-CUJSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS - Prime Contractor- IBM\23. Woil<brain Scalabii[!Y Assessment\Piari\Test 3\Results\Test3 1000VU Run1 1Server 17042008 Rellorl.(!df
\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS- Prime Contractor- IBM\23. Workbrain Scalability Assessment\Pian\Test 3\Results1Test3 1000VU Run1 2 Server 18042008 ReQort.doc
\\NW-CLUSTER-{)1 CTDATA SERVER\CTDATA\DATA\Cn 17- SSS - Prime Contractor - IBM\23. Workbrain Scalabilitll Assessment\Pian\Test 3\Resulls\Test3 1200VU Run1 1Server 18042008 Rellort.doc
\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn1 7 - SSS - Prime Contractor - IBM\23: Workbrain Scalabilitll Assessment\Pian\Test 3\Resulls\Test3 1500VU Run1 1Server 18042008 Re(!orl.doc
\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS - Prime Contractor - IBM\23. Workbrain Scalability AssessmentiPlan\Test 3\ResuHs\Test3 1500VU Run1 3Server 22042008 Re12ort.!1df
\\NW-CLUSTER-01 CTDATA SERVER\~TDATA\DATA\Cn17- SSS- Prime Contractor - IBM\23. Workbrain Scalabililll AssessmentiPian\Test 3\Results\Test3 2000VU Run1 3Server 22042008 Re12ori.Qdf
\\NW-Cl USTER-01 CTDATA SERVERICTDATA\DATA\Cn 17 - SSS - Prime Contractor - IBM\23. Workbrain Scalabililll AssessmentiPlan\Test 3\Results\Test3 2500VU Run1 3Server 22042008 Re12ort.!ldf
1\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS- Prime Contractor- IBM\23. Woil<brain Scalabilitl! Assessment\Pian\Test 3\Results\Test3 3000VU .Run1 3Server 2204200S ReQort·.!ldf
\\NW-CLUSTER-{)1 CTDATA SERVER\CTDATA\DATA\Gn17 - SSS - Prime 0.2 Contractor - IBM\23. Workbrain Scalability Assessment\Pian\Test 3\Results\P42
Commercial-in-Confide.nce Page 4 of 13
Program 42 CD-805-0005
Test Completion Report V2.0
T.~~t Tifp~ .
Rel.e.as~ oaiij"·
Version
·. , · ;,: -..rl .•
Scalability Testing Test 3 Batch Report V0.2.doc
Definitions
Term Definition
CPU Central Processing Unit.
End-state This term refers to the end of the SSI, SSS design, build and implementation program; when all Queensland Government agencies anQ Shared Service Providers have .gone live with the new SSS b,usiness solutions.
HTTP Hypertext Transfer Protocol.
J2EE Java 2 Platform, Enterprise Edition.
LDAP Lightweight Directory Access ProtocoL
POC Proof-of-Concept
SDA Solution Design Authority
SSI Shared Service Initiative,
SSL Secure Socket Layer.
sss Shared Service Solutions.
TPH Transactions per hauL
TPS Transactions per second.
Vuser Virtual User, HP/Mercury LoadRunner user that emulates the. business transaction of a normal user.
WoG Whole-Of-Government
CoinmerciaJ-in-Confidence P9ge 5 of 13
PrQgram 42
Table of Contents
CD-805-0005 Test Completion Report
V2.0
1. Introducti.on .................................... , ............................................................................................ 7 .
1.1 Purpos.e ...... ........ ..... ......................................................... ..... ................. ..... ....... .. ................... 7
1.2 Audience· ... ..... .. .... .... ... .... ................... .. .. ...... ..................... .... ............ .. .... .. ....... .. ........ .......... . 7
2. o·escription .................................................................................................................................... 7
2.1 Ovetview .. .... .......... .. .... .. .. .......... .. .. .. .. .................. .. .. .. ...................... .................. .. ................. 7
3. Test Scope ................... , ................................................................................................................. 8
4. Test Results ............................... , .................................................................................................. 8
4.1 Test 1 .. ............... ......... ............................ .......... ....... ; .. .......... ; ....... ... ...................... ..... ..... ..... 8
4.2 Test 2 ............ .. ... ................. .. , ... .............. ... ..... ..... .... ... ... .. ....... .. ....... .... ... ....... .... .... ... ........ .... 9
4.3 Test 3 .. ..... .. .. .... ......... ..... ...... .. .. .. ... ... ... ... ........... .. ... ....... .. ...... .... .... ... ....... .... .. .... .. ... ... .. .... .. .. 11
4.4 Any variance to originai plan .. .... ..................................... .. .. .. .. ............ , .... .. .. ..... .......... .. 13
4.5 Observations ......... ................. ............. .. ...... .. ......... ...... ...................................... ... ........ : ... . 13·
5. Conclusions ................................................................................................................................ 13
5.1 Exit criteria .... ..... ... ..................... ..... ................. .. ................................ .. .. ....... ........ .. .. , .. .. , .. .'13
5.2 Recommendations ...... ..... ............. ...... ; ... ........ .. .. .. .. ............ ................. ... ...... ... .... ............. 13
·Commercial-in-Confidence Page 6 of 13
Program 42
1 ~ Introduction
1.1 Purpose The purpose of this document is to:
~ Provide a summary of the Work brain Scalability Assessment testing effort;
CD-805-0005 Test Completion Report
V2.0
~~ Enabl~ management to m,ake an informed decision regarding whether Workbrain is the right solution for award interpretation, or if it is not, to identify the associated risks;
• Ensure exit criteria are met, or if they are not, that risk is accepted.
This document does not cover the detailed testing results. Rather it provides a high level summary of the results. For further detail, please refer to the individual testing results.
1.2 Audience The target audience includes all those who have had a role to play both during test preparation and test execution as well as those who will be involved in making the go/no go decision. This will include:
e Corp Tech SDA and SSS Technology Solutions • IBM • Program I Project Managers
2. Description
2.1 Overview The original SSS solution design for non rostering Agencies relied solely .on the SAP application for award interpretation processing. Based on actual payroll processing results at the Department of Housing, and e)(trapolating these to include the remainder of non rostering agencies, it became evident that the SAP application would be unable to process the award interpretation function within an acceptable timefram_e to me~t SSS business requirements. In IBM's Invitation to Offer, it was proposed to repiC~ce the SAP centric approach to award interpretation with the Workbrain application to alleviate processing time constraints and to realise the additional following benefits: · • Economies of scale through developing only one set of ri.!les that will be utilised across all Agencies.
The Workbrain architecture allows pay rules to be de~igned so they are fully configurable across multiple awards and across multiple Agencies. The award configurations have a large number of similarities allowing a small set of rule components to be built that can then simply be reconfigured to fit any number of awards. Currently ~wards i?)nd their associated pay ru les will be developed in Workbrain as part of the Phase One roll out to Health. Therefore, huge efficiencies can be gained through re-using these existing rules and configurations for other Agencies, eliminating the costly and time consuming development phase iri SAP; .
e Implementing all awards in Workbrain provides a single system of rec:ord for their configuration, therefore providing significant efficiencies for maintenance, EBA updates, training, etc;
• The Workbrain rule and leave accrual engine is extremely efficient in providing considerable time savings, particularly when all awards will already have been interpreted when SAP payroll runs are made, vastly improving the time window required for this;
o The system ar<;::hiteclure of tne Workbrain rule engine is designed to allow simple and efficient integration with custom components. This allows for a very short development cycle where a.ll custoinisations are written in Java and fully extendable and re-usable; ·
" Workbrain provides a large number of standard pay ru les within the core system, considerably reducing the size of custom development required;
G Changes to award agreements can be made quickly and easily via the pay rule editor in Worl<brain, eliminating the need for further custom developmemt and associated costs;
• Configuration of pay rules in Workbrain is simple relative to SAP, with all configurations fully exportable via XML.
Commercial-in-Confidence Pa~e 7 of 13
Program42 CD-805-0005 Test Completion Report
V2.0
In principle, SSS accepted IBM's proposal to utilise the Workbrain application for both rostering and non rostering agency award interpretation. Before proceeding with this approach, SSS has requested that IBM perform a series of tests that must evidence Workbrain's ability to scale, providing comfort that b~siness requirements can be satisfied. This scalability test requirement is documented as a deiiverable in SOW 5 -Priority Core HR & Finance Development. On test completion, a Workbrain Scalability Assessment Test Completion Report (contracted deliverable, this document) will be presented to SSS, outlining the tests performed, results obtained and interpreted findings. A go/no-go d.ecision on the proposed solution will be made by SSS after review and consideration of this document.
3. Test Scope
The scope of this testing exercise is to assess the scalability capability of the Workbraih epplication. As such any applica.tion functional testing, performance testing, stress and volume testing, and any tests to confirm production sizing is out of scope of this activity. ··
There are three types of testing involved in this assessment: • Non-rostering agency processing (Test 1) will execute award interpretation in batch utilising an
imported employee ·file as input. This test will be executed on interim sized batch and production sized database servers;
• Rostering agency processing (Test 2) will te$t concurrent user scalability by submitting timesheets through the onlihe Workbrain application fr9nf end and subsequently perform the Award interpretation. This test will be conducted on a production sized database server and on interim hardware (approximately - 50% productiqn) for the application servers;
• Combined rostering and non-rostering (Test 3) scenarios on production sized hardware using a range of transactions per hour and a range of concurrent users. Results will be assessed against anticipated volumes. The rostering and non-rostering tests will not overlap in time, i.e. The rostering and non-rostering tests will be executed at different times.
Important Note: Initially Test 3 included a file tr<!nsfer portion via SAP XI. A deCision has been made to exclude the same from the scope of Test 3, A CR has been raised by Corp Tech in order to remove this from the scope of the Test 3. Below is a representation of the testing types:
Workbraln
RcstDringeinpiO'JI!is
1'1-- ~~·-~ 1::- A•""
/..-;:st 3 inc/ud~';,\ .... . .. : combination or Test 1 J
\ and 2tes1 cases / .. '.:; ...... _____ ......
tnterpretallon
I Stllodule Vl~rk , ___, H PayroH Export I Delails Import
'---· . . t .. . . L------l - +-.! _· _ _...,
Non-rt;~oi~iing t-d.: [F.'~ ~o:;teling and
iimp\_0'1~' It~; CSV CSV ~~ ~'j ":;~~~~~~!)
~~~ . c.__ __ ~I,·_J ______ ~x, _______ ,+f_' ~ Le{]end - t
i~ csv csv 'IT~ .. ~ . ~~~t t R;:;::,~~~.·; · [~ _ _ t::::.~~=...::--·:::....j ____ s_·A_P ____ l::..:·.}::..::~=--4-~
Ta t 2
4. Test Results
4.1 Test 1
The results of Test 1 demonstrate the ability of the Workbrain batch processing solution to scale horizontally and vertically. In addition, they indicate very fast processing speeds for two of the three
Commercial-in-Confidence Page 8 of 13
!Program 42 CD-805-0005 Te~t Completion Report
V2.0
tests- full recalculation and creating the export fife . Both show increased performance proportionate to changes in hori.?;ontal and. vertic~! configuration.
The results ofthe schedule work detail import test, were inconsistent and required further database tuning and test configuration before developing a final assessment on either processing speeds or th~ solution's ability to scale to meet d.ata import requirements. The test results to date. demonstrate the ability to scale vertically. They als.o indicate fast processing times which, when extrapolated, would very easily fall within the window for pay processing. However, an issue was surfaced with horizontal scaling (i:e., addition of job schedulers). The database timed out in an irregular manner when running 40 or more import process threads simultaneously, even though processing completed within required windows.
Schedule Work Detail Import File
Full Recalculation Process
Export File
4.2 Test 2
4.2.1 Test 2a
112 minutes
48 minutes
12 minutes
• 0
• • 0 .. 0
• 0
• • • • • • • •
Batch Servers: 1 Job schedulers: 4 Tasks per schedl,ller: 10 Total Tasks: 40 Periop: 7 Days (Q5May- 12May)
Batch servers: 2 Jcib schedulers: 2 Tasks pEir scheduler: 2 Threads per task: 4 Total threads: 16
Batch servers: 2 Job schedulers: 8 Tasks per scheduler: 1 Threads per iask: 1 Total threads: 8
8
Tests were performed to determine the maximum number of users that could be successfully run on 1 server, 2 servers and 3 servers. Below matrix shows the results witl11, 2 and 3 servers:
1
2
3
. ·: 4.:~: . ·:~JL .· ~'~·~· Sucri~'k~
100
200
300
Commercial-in-Confidence
300
450
· ~:~ ... ~:·· • I · ~ ~!; ·:·
600
• " .b~~·.·
<.'dj~~· . ;..-·
Page 9 of 13
Program 42
~ C1J Ill
~ .... c:: ~ :I (.) c 0 0
3000 2700 2400 2100 1800 1500 1200 900 600 300
0
Test2a
2
Servers
3
CD-805-0005 Test Completion Report
V2.0
It was determined that Workbrain was installed using default settings for WebSphere and that a WebSpi:Jere Specialist should be engaged to tune WepSphere. It was also determined that the Load Runner VUsers were executing at the rate of 32 timesheets submitted per VUser per hour. It was agreed that a more realistic scenario was that each of the concurrent users was executing at the. rate of 12 timesheets per VUser per hour, and this definition of concurrent users was used for subsequent tests, and VUsers were calibrated to this rate.
4.2.2 Test 2b
Tests were performed to determine the maximum number of concurrent users that could be successfully run on 1 server, 2 servers and 3 servers. Below matrix shoWs the results with 1, 2 and 3 servers:
·-· ")' • · _ '1 '0 1 •• tl 0 o o' 0 ',\ t~> ,• ~?~. !. :.Ll~ ~ ·- .... ~ ':r:. ;..{' G. '
·': Wci.r~Brai~rr :-· >- .: ~·-· · , ··:q.oncur~nt 0$~-rs, ;~,~~ ".',->: '):.:. 1 ·:~ s':. . ~~ ~ . ~ · ,.,..--~· . · . ~ ;· · .c;t. ~ ~~ ~~ ~. \ c._-n·. ~~ ~fit ~· . :A!·.~ ~;: ~~ers ,~:.>. -c:-._.}succ'e~~f . •: , Boj7fj,~tli~lf~.f. ( · ~- F~!~,;: , . :. ·
1 380 400
2 500 630 675
760 775 800
Test2b
2 3
Se~vers
Commercial-in-Confidence Page 10 of 13
Progr(tm 42 CD-805-0005 Test Completion Report
V2. 0
It was determined that the cause of the problem was the load balancing method used by the W eb Server Plug-in. During this test; the "Weighted Round Robin Equal Weighf' method was used. The method was changed to Random and subsequently Test 2c was executed.
4.2.3 Test2c
After completion and certification of test 2b, further tuning was done to the interim landscape. Capacity was do!Jbled by employing an alternative load balancing algorithm on the weq server plug-in. Where "Weighted Round Robin Equal Weighf' had been employed previously, the use of "Random" doubled the performance. Tests were performed to determine the maximum number of users that could be successfully run on 1 server, 2 servers and 3 servers. Below matrix shows the re.sults with 1. 2 and 3 servers:
Test2c
3000 2700 ·--·-":·'
'· 2400
112 2100 C1l. (/)
~ 1800 +> .. . ,· c 1500 : ·~:-
~ ::2 1200 u c 900 0 0 600
300 0
2 3
L Servers
--· -- - - - - - -
With this configuration, the interim environment showed scaling as additional servers were added, more concurrent users could be handled, and the configuration could successfully handle the 1500 concurrent users that it was sized for.
4.3 Test 3
4.3.1 Test 3 online
The online portion of Test 3 was e~ecuted against the Workbrain environment on the Production infrastructure. Note that SSL is fully configured from Client all the way up to the WebSphere Application Servers, due to securitY requirements in the Production infrastructure.
Tests were performed to qetermine the maximum number of users that could be successfully ·run on 1 server, 2 servers and 3 servers. Below matrix shows the results with 1, 2 and 3 servers:
Commercial~in-Confidence Pa~e 11 of 13
Program42
rl Q) en :J ... c: ~ :I u c:
I 0 0
i
L
;3000
2900
2000
1500
1000
500
0
----Test3
2
Servers
3000
3
CD-805-0005 Te~t Completion Report
V2.0
This test was executed using 3, 000 concurrent virtual w~ers on three Production Workbrain servers. 3,000 concurrent users represent a total user base of 90,000 and i.s considered to be the end state figure for rostered employees. This test was successful.
4.3.2 Test 3 batch
The results of Test 3 demonstrate the ability of the Workbraih batch processing solution to meet the exit criteria on production~sized hardware.
Full Recalculation Process
Export File . 28minvtes
-' ~ ,; .. .
!- ,1 \ . · . ·' .·
Commercial-in-Confidence · 1
Two sequential jobs of • Batch Servers: 1 0 Schedulers per Server: 4
" Ti!sks per Scheduler: 2 e Total Tasks: a
• Batch Servers:. 2 0 Schedulers per Server: 4 0 Tasks per Scheduler: 2 • Threads per Task:,a • Total Threads: 128 • • Batch servers: 2 • Schedulers per Server: 2 • Tasks per Scheduler: 1 .. Threads per Ta:>k 1 • Total Tasks: 4 • 4
. . . ~ ·· ; . .
Paqe 12 of 13
Progra~42
4.4 Any variancE:> to original plan
CD-805-0005 Test Completion Report
V2.0
The original plan for n~st 3 included a file transfer portion to SAP via XI. Corp Tech raised a Change Request in order to remove this from th'e scope of Test 3, hence Test 3 execution included the Online Award processing and Batch Awarq processing. The B~tch Award processing included an import to Workbrain, processing by Workbrain and export from Workbrain, without transferring files to and ffom Workbrain from SAP (via XI).
4.5 Observations
Ouring Test 2c, an additional approach was also undertaken. The web servers were eliminated from the solution and Citrix Netscaler performed all load balancing. All traffic was routed directly from Citrix Netscaler to the Workbrain application servers. The three load balancing algorithms were:
• Simple round robin; • Fastest response time; • Least connections,
The first algorithm provided performance comparable to (if not marginally better than) the performance when using the web servers. The remaining algorithms did not exhibit good performance.
During Test 3, it was found that the Oracle SQL execution plan for Payroll Export was not optimal, and the registry parameter was changed from FiRST_ROWS to ALL_ROWS.
During Te$t 3, it was found that the Oracle SQL execution plan was not optimal, snd that Oracle was choosing the Wrong index for the query. The workaround used to was to drop the IDX_EMPSKD_WORKDT index. A Workbrain Support Hub Incident was raised to add a config1.1rable registry parameter for this tuning (Incident number 1978603).
5. Conclusions
5.1 Exit criteria
All exit criteria, as defined in the Test Plan, have been reached.
Online
Batch
3000 concurrent users on 3 Servers -Application scales linearly -
Work Schedule Import less than 4 hoursFull Awarq ~ecalculalion less than 1 hour Payroll Export less than 30 minutes -
reached reached
reached reached reached
5.2 Recommendations
Recommendation 1: Based on tt1e results of the scalability assessment tests, it is recommended that a "Go" decision be made to adopt Workbrain as the whole-of-government solution for awards interpretatiqn. The tests did hot identify any unacceptable risks. The Workbrain scalability assessment has been sucoessfuliy completed.
Recommendation 2: It is re.commended that further architectural analysis and testing be undertaken to determine if the Web Server functiOnality can b~ handJed by the NetScaler device.
Recommendation 3: It is recommended that detailed Non-Functional Requirements be gathered aild architectural work products devetop~d for Program 42. ·
This is the last page of this ,document.
· CommerCial-in-Confidence PaQe 13 of 13
Deliverable Acceptance Sheet -A ·separate Acceptance Sheet is to be completed for each deliverable. Upon completion, please submit to Vendor Management
~ -· - -- -
D~liverable Descript_ion
l)eliver.i~te IQ ...
CD 805 0005
Deliverable N;~me Workbrain Scalability Assessment Test Report
~ow N1,un~~r SOW5
Qeliverabi.e L-ea(l Tom Lambert
Purpose Submission of Workbrain Scalability Assessment for SOW 5
Date of Deliverable .. .-,
26/05108 I Date d!:l~··t9. IJ3M J 30/06/08
Deljverable Cqmposition Word document .
Dep~pdent Deliverables N/A
Prere.guisite Deli·.J~~ables N/A
Linked. to Payment Milestone . __ [gj Yes 0 No . .-,, Mileiito~e Date: -3bY1 June ZOOS
Acceptance Details
breat~ii; -
· Oatei · ro· ''' ' .... ,., ···t .. ::. ... ,:_ ::
; . •; ... , . ~i· ••
~Acc'eP.ta~ce Lea& lillike Robin~on · - P.e~t~i -...... '
!_:;::_: .·., • .. ·n.--· .. . : .. ~ .-•---!_!( _ • . ·-:-:' !
ACCEPTED (refer to 'Comments' Section) NOT ACCEPTED (refer to 'Comments' Section)
181 D ;:Accell~ii~~ Lead .s·i~~ature: ' '· D~t~;·:~p/o.S/2008 ·_.
··:: :_·'·; ,, I
· !- •·' - e
. -~ . ; ~.~:.\ .. ... .. !--~ '
Acceptance Board
Acceptance Panel Area of Acceptance Date (list Board Members) (e.g. Technical, Quality, Audit, Resourcing etc)
Mike Robinson HR Lead 30/05/2008 - ---
-
-
-
~-
-
:: :
C:\Documents and Settings\rzgott\l ocal Settings\Temp\notesE259CD\CD_B05...:.0005_0eliverable Acceptance Sheet.doc Page 1 of2
,.,: ·-
.. .. . • .·. \ .
. ;;;:: . ,-.· . .:.:/., .
- -
--
- - --
C~mments
Comments /Issues I Defects (add notes or attach document. If not accepted, please provide list of defects and issues). - -
- --··
- -
- -
r-- · -
f-- - -- -
f-----.- . -
- - -
- - · ___:____ -
-
t---- -
r- -
r--
- -·
-
-·
t---
C:\Documerits and Settings\rzgolt\Local Settings\Temp\notesE259CD\CD_B05_0005_Deliverable Aci:eptarice Sheet.doc Page 2 ·or2
-
-
·· -
- -
-
-
-
-
-
--
-
- -