INFO 631 Prof . Glenn Booker

Post on 25-Feb-2016

30 views 0 download

Tags:

description

INFO 631 Prof . Glenn Booker. Week 2 – Reliability Models and Customer Satisfaction. Reliability Models. Reliability Models are used here to treat software as though we expect predictable performance from using it on a regular basis - PowerPoint PPT Presentation

Transcript of INFO 631 Prof . Glenn Booker

www.ischool.drexel.edu

INFO 631 Prof. Glenn Booker

Week 2 – Reliability Models and Customer Satisfaction

1INFO631 Week 2

www.ischool.drexel.eduINFO631 Week 2 2

Reliability Models

• Reliability Models are used here to treat software as though we expect predictable performance from using it on a regular basis

• Hence this assumes we’re dealing with fairly stable requirements, and a well controlled environment

www.ischool.drexel.eduINFO631 Week 2 3

Why use Reliability Models?

• Determine objective quality of the product• Use for planning resources needed to fix

problems

www.ischool.drexel.eduINFO631 Week 2 4

Independent Variable in Reliability Growth Models

• Typical scope of measurement (X axis) includes one of these:– Calendar time (days of testing)– Cumulative testing effort (hours of testing)– Computer execution time (e.g. number of

CPU hours)

www.ischool.drexel.eduINFO631 Week 2 5

Key Dependent Variables for Reliability Growth Models

• Typical dependent variables (Y) include:– Number of defects found per life cycle phase,

or total number ever– Cumulative number of failures over time– Failure rate over time– Time between failures

www.ischool.drexel.eduINFO631 Week 2 6

Terminology• Reliability – probability that system

functions without failure for a specified time or number of natural units in a specified environment– “Natural unit” is related to an output of a

system• Per run of a program• Per hours of CPU execution• Per transaction (sale, shipment)

www.ischool.drexel.eduINFO631 Week 2 7

Terminology

• Availability – probability at any given time that a system functions satisfactorily in a given environment

• Failure intensity – the number of failures per natural or time unit

www.ischool.drexel.eduINFO631 Week 2 8

Software Reliability Modeling

• Requires characterizing and applying– The required major development

characteristics or goals• Reliability• Availability• Delivery date• Life-cycle cost (development,

maintenance, training, etc.)

www.ischool.drexel.eduINFO631 Week 2 9

Software Reliability Modeling– The expected relative use of the software’s

functions (i.e. its operational profile)• Focus resources on functions in proportion to their

use and criticality

www.ischool.drexel.eduINFO631 Week 2 10

Operational Profile

• Operational profile - a complete set of operations with their probabilities of occurrence– Operation = a major system logical task of

short duration which returns control to the system when complete, a.k.a. a scenario

www.ischool.drexel.eduINFO631 Week 2 11

Types of Reliability Models

• Static and Dynamic• Static

– Uses other product characteristics (size, complexity, etc.) to estimate number of defects

– Good for module-level estimates (detailed)– See discussions of size and complexity measures

www.ischool.drexel.eduINFO631 Week 2 12

Types of Reliability Models

• Dynamic– Based on statistical distribution; uses current defect

pattern to estimate future reliability– Good for product-level estimates (large scale)– Includes the Rayleigh and Exponential models

www.ischool.drexel.eduINFO631 Week 2 13

Dynamic Reliability Models• Model entire development process

– Rayleigh model; • Model back-end formal testing

(after coding)– Exponential model

• Both are a function of time or life cycle phase, and are part of the Weibull family of distributions

“back-end” here refers to the later phases of the life cycle

www.ischool.drexel.eduINFO631 Week 2 14

Define

• PDF = Probability Density Function, is the number of defects which will be found per life cycle phase

• CDF = Cumulative Density Function, is the total number of defects which will be found, as a function of life cycle phase

www.ischool.drexel.eduINFO631 Week 2 15

Weibull Model

Let: m = curve family, c = shape parameter, t = time

Then: PDF = (m/t)*(t/c)^m*exp(-(t/c)^m) CDF = 1 - exp(-(t/c)^m) What can this look like? Lots of things!First, fix c=1.5 and look at various ‘m’ values

www.ischool.drexel.eduINFO631 Week 2 16

Weibull for M=0.5 and 1, c=1.5

TIME

5.50

5.20

4.90

4.60

4.30

4.00

3.70

3.40

3.10

2.80

2.50

2.20

1.90

1.60

1.30

1.00

.70

.40

.10

PD

F

1.2

1.0

.8

.6

.4

.2

0.0

PDF for M=0.5

PDF for M=1M=0.5

M=1

www.ischool.drexel.eduINFO631 Week 2 17

M=1.5

M=2

Weibull for M=1.5 and 2, c=1.5

TIME

5.50

5.20

4.90

4.60

4.30

4.00

3.70

3.40

3.10

2.80

2.50

2.20

1.90

1.60

1.30

1.00

.70

.40

.10

PDF

.7

.6

.5

.4

.3

.2

.1

0.0

PDF for M=1.5

PDF for M=2

Notice the Y axis

range is changing

www.ischool.drexel.eduINFO631 Week 2 18

Weibull for M=3 and 5, c=1.5

TIME

5.50

5.20

4.90

4.60

4.30

4.00

3.70

3.40

3.10

2.80

2.50

2.20

1.90

1.60

1.30

1.00

.70

.40

.10

PD

F

1.4

1.2

1.0

.8

.6

.4

.2

0.0

PDF for M=3

PDF for M=5

M=3

M=5

www.ischool.drexel.eduINFO631 Week 2 19

M=7

M=9

Weibull for M=7 and 9, c=1.5

TIME

5.50

5.20

4.90

4.60

4.30

4.00

3.70

3.40

3.10

2.80

2.50

2.20

1.90

1.60

1.30

1.00

.70

.40

.10

PDF

2.5

2.0

1.5

1.0

.5

0.0

PDF for M=7

PDF for M=9

For large ‘m’ values, Weibull

looks like a normal

distribution centered on ‘c’

www.ischool.drexel.eduINFO631 Week 2 20

Rayleigh Model

• The history of defect discovery across the life cycle phases often looks like the Rayleigh probability distribution

• Rayleigh model is a formal parametric model, used to produce estimates of the future defect count

www.ischool.drexel.eduINFO631 Week 2 21

Rayleigh Model

• Rayleigh model and defect origin/found analyses deal with the defect pattern of the entire software development process

• Is a good tool, since it can provide sound estimates of defect discovery from fairly early in the life cycle

www.ischool.drexel.eduINFO631 Week 2 22

Rayleigh Model

Let: m = 2, c = scale parameter, t = time Then:PDF = (2/t)*(t/c)^2*exp(-(t/c)^2)

CDF = Cumulative defect arrival patternCDF = 1 - exp(-(t/c)^2)

www.ischool.drexel.eduINFO631 Week 2 23

Rayleigh Model Assumptions

1. Defect rate during development is correlated with defect rate after release.

2. If defects are discovered and removed earlier in development, fewer will remain in later stages.

In short, “Do it right the first time.”

www.ischool.drexel.eduINFO631 Week 2 24

Rayleigh Model

• The value of ‘c’ determines when the curve peaks– tmax = c/(2) is the peak

• Area up to tmax is where 39.35% of all defects will be found (ideally)

• Now look at influence of ‘c’ value on curve shape

www.ischool.drexel.eduINFO631 Week 2 25

Weibull for M=2, c=1, 1.5, and 2

TIME

5.50

5.20

4.90

4.60

4.30

4.00

3.70

3.40

3.10

2.80

2.50

2.20

1.90

1.60

1.30

1.00

.70

.40

.10

PDF

1.0

.8

.6

.4

.2

0.0

PDF for M=2, c=1

PDF for M=2

PDF for M=2, c=2

, c=1.5

www.ischool.drexel.eduINFO631 Week 2 26

Rayleigh Model Implementation

• Various tools can model the Rayleigh curve– PASW/SPSS (using Regression Module)– SAS– SLIM (by Quantitative Software Management)– STEER (by IBM)

www.ischool.drexel.eduINFO631 Week 2 27

Rayleigh Model Reliability• Statistical reliability relates to confidence interval

of the estimate, which is in turn related to sample size

• Small sample size (only 6 data points per project) means low statistical reliability, often underestimating actual later reliability

• Improve this by using other models and comparing results

www.ischool.drexel.eduINFO631 Week 2 28

PTR Submodel

• A variation on the Rayleigh model can be used for predicting defects which will be found during integration of new software into a system– PTR is Program Trouble Report or Problem

Tracking Report, a common mechanism for defect tracking

• Follows the same idea as Rayleigh

www.ischool.drexel.eduINFO631 Week 2 29

Reliability Growth Models - Exponential Model

• Exponential model is the basic reliability growth model - i.e. reliability will tend to increase over time

• Other reliability models include: Time Between Failure Models and Fault Count Models

www.ischool.drexel.eduINFO631 Week 2 30

Exponential Model• Reliability growth models are based on data

from the formal testing phase– After the software has been completely

integrated (compiled & built)– When the software is being tested with test

cases chosen randomly to approximate an operational (real-world usage) profile

– Testing is customer oriented

www.ischool.drexel.eduINFO631 Week 2 31

Exponential Model

• Rationale is that defect arrival during testing is a good indicator of the reliability of the product when used by customers

• During this testing phase, failures occur, defects are fixed, software becomes more stable, and reliability grows over time

www.ischool.drexel.eduINFO631 Week 2 32

Exponential Model

• Is a Weibull distribution with m = 1• Let: c = scale parameter,

t = time, l = 1/cCDF = 1 - exp(-t/c) = 1 - exp(-lt)PDF = (1/c)*exp(-t/c) = l*exp(-lt)• l is the error detection rate or hazard rate• This form also works for light bulb failures, computer

electrical failures, etc.

www.ischool.drexel.eduINFO631 Week 2 33

Typical Time Between Failure Model Assumptions

• There are N unknown software faults at the start of testing

• Failures occur randomly• All faults contribute equally to failure• Fix time is negligibly small• Fix is perfect for each fault

www.ischool.drexel.eduINFO631 Week 2 34

Time Between Failure Models

• Jelinski-Moranda (J-M) Model– Assumes random failures, perfect zero time fixes, all

faults equally bad• Littlewood Models

– Like J-M model, but assumes bigger faults arefound first

• Goel-Okumoto Imperfect Debugging Model– Like J-M model, but with bad fixes possible

www.ischool.drexel.eduINFO631 Week 2 35

Fault Count Model Assumptions

• Testing intervals are independent of each other

• Testing during intervals is reasonably homogeneous

• Number of defects detected is independent of each other

www.ischool.drexel.eduINFO631 Week 2 36

Fault Count Models• Goel-Okumoto Nonhomogeneous Poisson

Process Model (NHPP)– # of failures in a time period, exponential

failure rate (i.e. the exponential model!)• Musa-Okumoto Logarithmic Poisson

Execution Time Model– Like NHPP, but later fixes have less effect

on reliability

www.ischool.drexel.eduINFO631 Week 2 37

Cumulative Defects versus Cumulative Test Hours

Goel Okumoto model:

m(t) = a*(1 - e-b*t)

l(t) = m’(t) = a*b* e-b*t

where: m(t) = expected number of failures observed at time t l(t) = failure density a = expected total number of defects b = constant

www.ischool.drexel.eduINFO631 Week 2 38

Fault Count Models

• The Delayed S and Inflection S Models– Delayed S: Recognizes time between failure

detection and fix– Inflection S: As failures are detected,

they reveal more failures

www.ischool.drexel.eduINFO631 Week 2 39

Mean Time to Failure (MTTF)

• Mean Time to Failure is the average amount of time using the product between failures

• MTTF = (total run time) / (number of failures)

www.ischool.drexel.eduINFO631 Week 2 40

Software Reliability Modeling:Time Between Failures

• Time between failures is expected to increase, as failures occur and faults are fixed

Execution Time Line0

Failure

www.ischool.drexel.eduINFO631 Week 2 41

Software Reliability Modeling: Time Between Failures

Reliability, R(t) - probability of failure free operation for a specified period of time

Time Since Last Failure (t)

1.0

Rel

iabi

lity

www.ischool.drexel.eduINFO631 Week 2 42

WARNING

• Reliability models can be wildly inaccurate, particularly if based on little and/or irrelevant data (e.g. from other industries, or using bad assumptions)

• Validate estimates with other models and common sense

www.ischool.drexel.eduINFO631 Week 2 43

Reliability Modeling

1. Examine data on a scatter diagram. Look for trends and level of detail.

2. Select model(s) to fit the data.3. Estimate the parameters of each model.4. Obtain fitted model using those parameters.5. Check goodness-of-fit and reasonableness

of models.6. Make predictions using fitted models.

www.ischool.drexel.eduINFO631 Week 2 44

Test Compression Factor

• Defect detection during testing is different from that by customer usage, hence the defect rates may change.

• Result is that fewer defects are found just after product release

• Or, testing is better at finding defects than customer usage

www.ischool.drexel.eduINFO631 Week 2 45

Test Compression Factor

• Hence for maintenance, use reliability models ONLY for defect number or rate, and look for field defect rate patterns to be different from those found during development (number of defects found drops after release, due to less effective customer “testing”)

www.ischool.drexel.eduINFO631 Week 2 46

Customer Satisfaction

www.ischool.drexel.eduINFO631 Week 2 47

Customer Satisfaction

• Customer evaluation of software is the most critical “test”

• Want to understand what their priorities are, in order to obtain and keep their business

www.ischool.drexel.eduINFO631 Week 2 48

Total Quality Management

• Expanded from just product quality to maintaining a long term customer relationship

• 5x cheaper to keep an existing customer than find a new one

• Unhappy customers tell 7-20 people, versus happy customers tell only 3-5 people

www.ischool.drexel.eduINFO631 Week 2 49

Customer Satisfaction Surveys

• Customer call-back after x days• Customer complaints• Direct customer visits• Customer user groups• Conferences

www.ischool.drexel.eduINFO631 Week 2 50

Customer Satisfaction Surveys

• Want representative sample of all customers

• Three main methods– In person interviews

Can note detailed reactionsMay introduce interviewer biasExpensive

www.ischool.drexel.eduINFO631 Week 2 51

Customer Satisfaction Surveys– Telephone interviews

Can still be very validCheaper than in person interviewsLack of interactionLimited audience

– Mail questionnairesHow representative?Low response rateVery cheap

www.ischool.drexel.eduINFO631 Week 2 52

Sampling Methods

• Often can’t survey entire user population• Four methods

– Simple random sampleMust be truly random, not just convenient

– Systematic samplingUse every nth customer from a list

www.ischool.drexel.eduINFO631 Week 2 53

Stratified Sampling

– Group customers into categories (strata); get simple random samples from each category (stratum). Can be very efficient method.

– Can weigh each stratum equally (proportional s.s.) or unequally (disproportional s.s.)

– For unequal, make fraction ~ standard deviation of stratum, and ~ 1/ square root (cost of sampling). F ~ s/sqrt(cost)where “sqrt” is “square root”

“~” means “is proportional to”

www.ischool.drexel.eduINFO631 Week 2 54

Cluster Sampling

• Divide population into (geographic) clusters, then do simple random samples within each selected cluster– Try for representative clusters– Not as efficient as simple random sampling,

but cheaper– Typically used for in-person interviews

www.ischool.drexel.eduINFO631 Week 2 55

Bias

• Look out for sample bias!• E.g. basing a national voting survey on a

Web-based poll

www.ischool.drexel.eduINFO631 Week 2 56

Sample Size

• How big is enough?• Depends on:

– Confidence level (80 - 99%, to get Z)– Margin of error (B = 3 - 5%)

• For simple random sample, also need – Estimated satisfaction level (p), and – Total population size (N = total number

of customers)

www.ischool.drexel.eduINFO631 Week 2 57

What’s ‘Z’?

• ‘Z’ is the critical Z value for a two-sided test of means

• Here we are striving for a sample whose mean customer satisfaction is close enough to the population’s mean – where “close enough” is defined by the Z value

www.ischool.drexel.eduINFO631 Week 2 58

What Confidence Level?

• The results are always subject to the desired confidence level – since we are never perfectly sure of our results– For analysis of medical test results, typically

insist on 99% confidence– Otherwise 95% is commonly used– Software tests may use as low as 80%

www.ischool.drexel.eduINFO631 Week 2 59

Critical Z values

Confidence Level 2-sided critical Z

80% 1.28

90% 1.645

95% 1.96

99% 2.57

www.ischool.drexel.eduINFO631 Week 2 60

Sample Size

• Sample size is given by n = [N*Z^2*p*(1-p)]/[N*B^2 + Z^2*p*(1-p)]

• Note that the sample size depends heavily on the answer we want to obtain, the actual level of customer satisfaction (p)!

www.ischool.drexel.eduINFO631 Week 2 61

Sample Size

• If we choose– 80% confidence level, then Z = 1.28– 5% margin of error, then B = 0.05– and expect 90% satisfaction, then p = 0.90

• n = (N*1.28^2*0.9*0.1)/ (N*0.05^2 + 1.28^2*0.9*0.1)

• n = 0.1475*N/(0.0025*N + 0.1475)Notice for B and p that percents are converted to decimals!

www.ischool.drexel.eduINFO631 Week 2 62

Sample SizeGiven:Z 1.28p 0.9B 0.05

Hence:Z^2 1.6384p(1-p) 0.09B^2 0.0025

Find:N n

10 8.55035520 14.9355850 27.06052

100 37.09996200 45.54935500 52.75873

1000 55.6972410000 58.63655

100000 58.947631000000 58.97892Infinity 58.9824

<- Sampling isn’t very helpful for small populations!

www.ischool.drexel.eduINFO631 Week 2 63

Sample Size

• If don’t know customer satisfaction value ‘p’, use 0.5 as worst-case estimate

• Once the real value of ‘p’ is known, solve for the actual value of B (margin of error)

• Key challenge is finding a truly representative sample

www.ischool.drexel.eduINFO631 Week 2 64

Analysis of Customer Satisfaction Data

• Use five point scale (very satisfied, sat., neutral, dissat., very dissat.)

• May convert to numeric scale; 1=very dissatisfied, 2=dissatisfied, etc.

• Typically use 95% confidence level (Z=1.96), but 80% may be okay to show hint of trend

www.ischool.drexel.eduINFO631 Week 2 65

Presentation of Customer Satisfaction Data

• Make running plot of % satisfied vs time, with +/- margin of error (B)

• Some like to plot percent dissatisfied instead

• May want to break satisfaction into detailed categories, and track each of them separately

www.ischool.drexel.eduINFO631 Week 2 66

Other Satisfaction Notes

• Key issues raised by customers may not be most needed areas of development (e.g. documentation vs reliability)

• Can examine correlation of specific satisfaction attributes to overall satisfaction; is bad X really an indicator of dissatisfied customers?

• Use regression analysis to answer this

www.ischool.drexel.eduINFO631 Week 2 67

CUPRIMDA (per IBM)

• Capability (functionality)• Usability• Performance• Reliability• Installability• Maintainability• Documentation• Availability

Can measure customer satisfaction for each of

these areas, plus overall satisfaction

www.ischool.drexel.eduINFO631 Week 2 68

Multiple Regression

• We have had models with one variable related to another, e.g.Schedule = a*(Effort)^b

• Linear and logarithmic regression can also be done with many variables, like: Overall Satisfaction = a + b*(Usability Sat.) + c*(Performance Sat.) + d*(Reliability Sat.) and so on

www.ischool.drexel.eduINFO631 Week 2 69

Multiple Regression

• This results in estimates of constants a, b, etc.– A linear regression is often better

for real-valued data– Logistical regression is often better for data

which may only have two values (Yes/No, T/F)• Sometimes both are tried to see which

gives the best results

www.ischool.drexel.eduINFO631 Week 2 70

Now What?

• Plot each factor’s regression coefficient (a, b, …) vs. the customer satisfaction level (%) for that factor; then on this plot:

• Determine priorities for improving customer satisfaction from top to bottom (then left to right, if there are equal coefficients)

www.ischool.drexel.eduINFO631 Week 2 71

Non-product Satisfaction

• Many other areas can affect customer satisfaction– Technical solutions - product factors, and

technologies used– Support & Service - availability, knowledge– Marketing - point of contact, information– Administration - invoicing, warranty– Delivery - speed, follow-through– Company image - stability, trustworthiness

www.ischool.drexel.eduINFO631 Week 2 72

Next Steps

• Measure and monitor your and competitors’ customer satisfaction– In order to compete, your satisfaction level

must be better than your competition’s• Analyze what aspects are most critical to

customer satisfaction• Determine the root cause of shortcomings• Set quantitative targets, both overall and for

specific aspects• Prepare & implement a plan to do the above