Predictive Modelling
-
Upload
rajiv-advani -
Category
Documents
-
view
173 -
download
0
Transcript of Predictive Modelling
What is Predictive Modeling
DATA
CurrentAnd Past
FUTURISTIC DATA
PREDICT ANALYTICS
IDENTIFY TRENDS
RECOGNIZE PATTERNS
APPLY TECHNIQUES
Predictive Modeling
• Predictive modelling (aka machine learning)(aka pattern recognition)(...) aims to generate the most accurate estimates of some quantity or event.
• As these models are not generally meant to be descriptive and are usually not well–suited for inference.
Predictive Modeling• Statistical Technique - Predictive modeling is a
process used in predictive analytics to create a statistical model of future behaviour.
• Mathematical Technique - Predictive analytics is the area of data mining concerned with forecasting probabilities and trends.
DATA + TECHNIQUE = MODEL
How to build a Predictive Model• Assemble the set of input fields into a dataset
• Example: Age, Gender, Zip Code, Number of Items Purchased, Number of Items Returned
• This is a vector in a multi-dimensional space as multiple features are being used to describe the customer
Independent (to be determined)
Eg: Number of customers who buy watches as per gender
Dependent (they are measured or observed)
Eg: Gender
Types of Variables
Independent
Influences the dependent variable
Manipulated by the researcher
Dependent
Affected by changes in the dependent variable
Not manipulated by the researcher
Difference between variables
ControlControlled by the researcher keeping the values constant in both the groups eg: Price of items
ModeratingStudied along with other variables eg: Items returned
InterveningCan neither be controlled nor studied. Its effect is inferred from the results. Eg: behavior
Other types - Extraneous Variables
How to build a predictive model Steps
1. Gather data2. Answer questions3. Design the structure well4. Variable Generation5. Exploratory Data Analysis6. Variable Transformation7. Partitioning model set for model build
Algorithms1. Time Series2. Regression3. Association4. Clustering5. Decision Trees6. Outlier Detection7. Neural Network8. Ensemble Models9. Factor Analysis10. Naive Bayes11. Support Vector Machines12. Uplift13. Survival Analysis
Forecasting MethodsMethods
Qualitative
Casual Time Series
Smoothing
Quantitative
What is Time Series1. Review historical data over time
2. Understand the pattern of past behaviour
3. Better predict the future
Set of evenly spaced numerical data - Obtained by observing response variables at regular time periods
Forecast based only on past values- Assumes that factors influencing past, present and future will continue
Example: Year 2010 2011 2012 2013 2014Sales 78.7 93.2 93.1 89.7 63.5
Components of Time Series
TREND CYCLICAL
SEASONAL IRREGULAR
Components of Time Series - Patterns
Smoothing Methods1. Moving Averages
2. Weighted Moving Averages
3. Centered Moving Averages
4. Exponential Smoothing
Smoothing Methods – Moving AveragesMoving Average = ∑ (most recent n data values)
nTime Response Moving Total (n = 3) Moving Average (n=3)
2011 4 NA NA2012 6 NA NA2013 5 NA NA2014 3 15 5.002015 7 14 4.672016 NA 15 5.00
Smoothing Methods – Weighted Moving Averages
WMA= ∑ (Weight for period n) (Value in period n)∑Weights
Month Sales Weights MA WMAJan 10.00 1.00 Feb 12.00 2.00 Mar 16.00 3.00 Apr 12.67 13.67
Smoothing Methods – Centered Moving Averages
5 10
6 1310+13+11: 3= 11.33
7 11
Smoothing Methods – Exponential Smoothing Single Exponential Smoothing– Similar to single MA
Double (Holt’s) Exponential Smoothing– Similar to double MA– Estimates trend
Triple (Winter’s) Exponential Smoothing– Estimates trend and seasonality
Smoothing Methods – Single Exponential Formula Single exponential smoothing model
Ft+1 = αyt + (1 – α) Ft Ft+1= forecast value for period t + 1
yt = actual value for period t
Ft = forecast value for period tα = alpha (smoothing constant)
Smoothing Methods – Single Exponential Example
Suppose α = 0.2
QtrSalesAct Forecast from Prior Period Forecast for Next Period
1 23 NA 23Here Ft= yt since no prior information exists
2 40 23 (.2)(40)+(.8)(23)=26.4
3 25 26.4 (.2)(25)+(.8)(26.4)=26.296
Ft+1 = αyt + (1 – α) Ft
Regression Algorithms1. Linear Regression
2. Exponential Regression
3. Geometric Regression
4. Logarithmic Regression
5. Multiple Linear Regression
Regression Algorithms - LinearA linear regression line has an equation of the form Y = a + bX,
• X is the explanatory variable
• Y is the dependent variable.
• The slope of the line is b
• a is the intercept (the value of y when x = 0)
• a and b are regression coefficients
Regression Algorithms – Linear ExampleX Y Y' Y-Y' (Y-Y')2
1.00 1.00 1.21 0.21 0.042.00 2.00 1.64 0.37 0.133.00 1.30 2.06 0.76 0.584.00 3.75 2.49 1.27 1.605.00 2.25 2.91 0.66 0.44
MX MY sX sY r3 2.06 1.581 1.072 0.627
Regression Algorithms - ExponentialAn exponential regression produces an exponential curve that best fits a single set of data points.
Formula :(smoothing constant) X (previous act demand) + (1- smoothing constant) X (previous forecast)
1. Suppose you have been asked to generate a demand forecast for a product for year 2012 using an exponential smoothing method. The forecast demand in 2011 was 910. The actual demand in 2011 was 850. Using this data and a smoothing constant of 0.3, which of the following is the demand forecast for year 2012?
The answer would be F = (1-0.3)(910)+0.3(850) = 892
2. Use exponential smoothing to forecast this period’s demand if = 0.2, previous actual demand was 30, and previous forecast was 35.
The answer would be F = (1-0.2)(35)+0.2(30) = 34
Regression Algorithms - GeometricSequence of numbers in which each term is a fixed multiple of the previous term.
Formula: {a, ar, ar2, ar3, ... }
where:a is the first term, and
r is the factor between the terms (called the "common ratio")
Example:
2 4 8 16 32 64 128 256 ...
The sequence has a factor of 2 between each number
Each term (except the first term) is found by multiplying the previous term by 2.
222222224888
Regression Algorithms - LogarthmicIn statistics, logistic regression, or logit regression, or logit mode is a regression model where the dependent variable (DV) iscategorical.
Example: Grain size
Spiders(mm)0.245 absent0.247 absent0.285 present0.299 present0.327 present0.347 present0.356 absent0.36 present
0.363 absent0.364 present
Regression Algorithms – Multiple LinearA regression with two or more explanatory variables is called a multiple regression
222222224888Formula: Y = b 0 + b 1 * 1 + b 2 * 2 + .... + b k * X k + e
Y is the dependent variable (response)
X 1 , X 2 ,.. .,X k are the independent variables (predictors)
e is random error
b 0 , b 1 , b 2 , .... b k are known as the regression coefficients – to be estimated
Regression Algorithms – Multiple Linear Example
Association Algorithms
• If/ then statements
1. Apriori Example
Transactions Items bought
T1 item1, item2, item3
T2 Item1, item2
T3 Item1, item5
T4 Item1, item2, item5
Association Algorithms - ExampleTransaction ID Items Bought Mango Onion Nintendo Keychains Eggs Yo-Yo Doll Apple Umbrella Corn Ice cream
T1 Yes Yes Yes Yes Yes Yes T2 Yes Yes Yes Yes Yes Yes T3 Yes Yes Yes Yes T4 Yes Yes Yes Yes Yes T5 Yes Yes Yes Yes Yes
Items Bought Item # transactions Pairs Pairs # transactions Item # trans OKE 3
{M,O,N,K,E,Y} M 3 MO MO 1 KEY 2{D, O, N, K, E, Y } O 3 MK MK 3 STEP6{M, A, K, E} N 2 ME ME 2{M, U, C, K, Y } K 5 MY MY 2{C, O, O, K, I, E} E 4 OK OK 3
U 3 OE OE 3STEP1 D 1 OY OY 2
U 1 KE KE 4A 1 KY KY 3C 2 EY EY 2I 1
STEP2 STEP3 STEP4STEP5
Clustering Algorithms - Definition• Finding a structure in a collection of unlabeled data.
• The process of organizing objects into groups whose members are similar in some way.
• Collection of objects which are “similar” between them and are “dissimilar” to the objects belonging to other clusters.
Clustering Algorithms - Example
Clustering Algorithms - Classification • Exclusive Clustering
• Overlapping Clustering
• Hierarchical Clustering
• Probabilistic Clustering
Clustering Algorithms – Most Used • K-means
• Fuzzy C-means
• Hierarchical clustering
• Mixture of Gaussians
Clustering Algorithms – K Means Example • The distance between two points is defined as D (P1, P2) = | x1 – x2 | + | y1 – y2|
Table 1 C1 = (2,2) C2 = (1,14) C3= (4,3) ClusterPoints Coordinates D (P,C1) D (P,C2) D (P,C3)
P1 (2,2) 0 13 3 C1P2 (1,14) 13 0 14 C2P3 (10,7) 13 16 10 C3P4 (1,11) 10 3 11 C2P5 (3,4) 3 12 2 C3P6 (11,8) 15 16 12 C3P7 (4,3) 3 14 0 C3P8 (12,0) 17 16 14 C3
C1 = (2/1, 2/1) = (2,2)C2 = (2/2, 14+11/2) = (1,12.5)C3 = ((10 + 11 + 3 + 4 + 12/5), (7 + 4 + 8 + 3 + 9)/5) = (8,6.2)
Clustering Algorithms – Fuzzy C Means • Allows degrees of membership to a cluster
• 1. Choose a number c of clusters to be found (user input).
• 2. Initialize the cluster centers randomly by selecting n data points
• 3. Assign each data point to the cluster center that is closest to it
• 4. Compute new cluster centers as the mean vectors of the assigned data points. (Intuitively: center of gravity if each data point has unit weight.)
• Repeat 3 and 4 until clusters centers do not change anymore.
Clustering Algorithms – Hierarchical ClusteringBOS NY DC MIA CHI SEA SF LA DEN
BOS/NY/DC MIA CHI SEA SF LA DEN BOS/NY/
DC/MIA SEA SF/LA DEN
BOS 0 206 429 1504 963 2976 3095 2979 1949 BOS/NY/DC 0 1075 671 2684 2799 2631 1616
NY 206 0 233 1308 802 2815 2934 2786 1771 MIA 1075 0 1329 3273 3053 2687 2037 CHI
DC 429 233 0 1075 671 2684 2799 2631 1616 CHI 671 1329 0 2013 2142 2054 996 BOS/NY/DC/CHI
0 1075 2013 2054 996
MIA 1504 1308 1075 0 1329 3273 3053 2687 2037 SEA 2684 3273 2013 0 808 1131 1307 MIA 1075 0 3273 2687 2037
CHI 963 802 671 1329 0 2013 2142 2054 996 SF 2799 3053 2142 808 0 379 1235 SEA 2013 3273 0 808 1307
SEA 2976 2815 2684 3273 2013 0 808 1131 1307 LA 2631 2687 2054 1131 379 0 1059 SF/LA 2054 2687 808 0 1059
SF 3095 2934 2799 3053 2142 808 0 379 1235 DEN 1616 2037 996 1307 1235 1059 0 DEN 996 2037 1307 1059 0
LA 2979 2786 2631 2687 2054 1131 379 0 1059After merging DC with BOS-NY: (3) After merging CHI with BOS/NY/DC: (5)
DEN 1949 1771 1616 2037 996 1307 1235 1059 0
After merging BOS with NY: (2) BOS/ MIA CHI SEA SF/
LADEN BOS/NY/
DC/CHIMIA SF/LA/SEA DEN
BOS/NY DC MIA CHI SEA SF LA DEN BOS/NY/DC/CHI
0 1075 2013 996
BOS/NY 0 223 1308 802 2815 2934 2786 1771 NY/DC MIA 1075 0 2687 2037
DC 223 0 1075 671 2684 2799 2631 1616 BOS/NY/DC 0 1075 671 2684 2631 1616 SF/LA/SEA 2054 2687 0 1059
MIA 1308 1075 0 1329 3273 3053 2687 2037 MIA 1075 0 1329 3273 2687 2037 DEN 996 2037 1059 0
CHI 802 671 1329 0 2013 2142 2054 996 CHI 671 1329 0 2013 2054 996After merging SEA with SF/LA: (6)
SEA 2815 2684 3273 2013 0 808 1131 1307 SEA 2684 3273 2013 0 808 1307 BOS/NY/DC/CHI/DEN
MIA SF/LA/SEA BOS/NY/DC/CHI/DEN/SF/LA/SEA
MIA
SF 2934 2799 3053 2142 808 0 379 1235 SF/LA 2631 2687 2054 808 0 1059 BOS/NY/DC/CHI/DEN
0 1075 1059 BOS/NY/DC/CHI/DEN/SF/LA/SEA
0 1075
LA 2786 2631 2687 2054 1131 379 0 1059 DEN 1616 2037 996 1307 1059 0 MIA 1075 0 2687 MIA 1075 0
DEN 1771 1616 2037 996 1307 1235 1059 0After merging SF with LA: (4)
SF/LA/SEA 1059 2687 0
After merging DEN with BOS/NY/DC/CHI: (7)After merging SF/LA/SEA with BOS/NY/DC/CHI/DEN: (8)
Clustering Algorithms – Probabilistic Clustering• Gaussian mixture models (GMM) are often
used for data clustering.
• A probabilistic model that assumes all the data points are generated from a mixture of a finite number of Gaussian distributions with unknown parameters
Decision Tree Algorithms• A decision tree is a structure that divides a
large heterogeneous data set into a series of small homogenous subsets by applying rules.
• It is a tool to extract useful information from the modeling data
All Data
Designer Watches > 5000
Wallets > 1000
Jewellery > 10000
Bags > 5000
MalesAge > 20
FemalesAge > 20
Outlier Detection AlgorithmsAn outlier is an observation that lies an abnormal distance from other values in a random sample from a population.The data set of N = 90 ordered observations as shown below is examined for outliers:
30, 171, 184, 201, 212, 250, 265, 270, 272, 289, 305, 306, 322, 322, 336, 346, 351, 370, 390, 404, 409, 411, 436, 437, 439, 441, 444, 448, 451, 453, 470, 480, 482, 487, 494, 495, 499, 503, 514, 521, 522, 527, 548, 550, 559, 560, 570, 572, 574, 578, 585, 592, 592, 607, 616, 618, 621, 629, 637, 638, 640, 656, 668, 707, 709, 719, 737, 739, 752, 758, 766, 792, 792, 794, 802, 818, 830, 832, 843, 858, 860, 869, 918, 925, 953, 991, 1000, 1005, 1068, 1441
The computations are as follows:• Median = (n+1)/2 largest data point = the average of the 45th and 46th ordered points = (559 + 560)/2 = 559.5• Lower quartile = .25(N+1)th ordered point = 22.75th ordered point = 411 + .75(436-411) = 429.75• Upper quartile = .75(N+1)th ordered point = 68.25th ordered point = 739 +.25(752-739) = 742.25• Interquartile range = 742.25 - 429.75 = 312.5
• Lower inner fence = 429.75 - 1.5 (312.5) = -39.0• Upper inner fence = 742.25 + 1.5 (312.5) = 1211.0• Lower outer fence = 429.75 - 3.0 (312.5) = -507.75• Upper outer fence = 742.25 + 3.0 (312.5) = 1679.75
Neural Networks• a “connectionist” computational system
• a field of Artificial Intelligence (AI)
• Kohonen self-organising networks
• Hopfield Nets
• BumpTree
Ensemble Models • Monte Carlo Analysis
Task Time Estimatemonths
Minmonth
Most Likelymonth
Maxmonths
1 5 4 5 72 4 3 4 63 5 4 5 6
14 11 14 19
Time Months # of Times out of 500 Percentage of Total (rounded)12 1 013 31 614 171 3415 394 7916 482 9617 499 10018 500 100
Factor Analysis• Data reduction tool • Removes redundancy or duplication from a set of
correlated variables • Represents correlated variables with a smaller set of
“derived” variables. • Factors are formed that are relatively independent of
one another. • Two types of “variables”: – latent variables: factors –
observed variables
Naive Bayes Theorem
Naive Bayes Theorem ExampleIn Orange County, 51% of the adults are males. (It doesn't take too much advanced mathematics to deduce that the other 49% are females.) One adult is randomly selected for a survey involving credit card usage.
a. Find the prior probability that the selected person is a male.
b. It is later learned that the selected survey subject was smoking a cigar. Also, 9.5% of males smoke cigars, whereas 1.7% of females smoke cigars (based on data from the Substance Abuse and Mental Health Services Administration). Use this additional information to find the probability that the selected subject is a male
Naive Bayes Theorem SolutionM = MaleC = Cigar SmokerF = FemaleN = Non Smoker
P (M) = 0.51 as 51% are smokersP (F) = 0.49 as 49% are femalesP (C/M) = 0.095 because 9.5% of males smoke cigarsP (C/F) = 0.017 because 1.7% of females smoke cigars
So P (M/C) = 0.51 . 0.095_____________________0.51 . 0.095 + 0.49 . 0.017
= 0.853
Support Vector machines
Uplift Modelling
• How is it related to Individual’s behaviour?
• When can we use it as a solution?
• Predict change in behaviour
P T (Y | X1, . . . , Xm) − P C (Y | X1, . . . , Xm)
Survival AnalysisChristiaan Huygens' 1669 curve showing how many out of 100 people survive until 86 years.From: Howard Wainer STATISTICAL GRAPHICS: Mapping the Pathways of Science. Annual Review of Psychology. Vol. 52: 305-335.
Examples to be solvedBaye’s Theorems1. A company purchases raw material from 2 suppliers A1 and A2. 65% material comes from A1 and the rest from A2. According to inspection reports, 98% material supplied by A1 is good and 2% is bad. The material is selected at random and was tried on machine for processing. The machine failed because the material selected was bad or defective. What is the probability that it was supplied by A1?
2. The chance that Dr. Joshi will diagnose the disease correctly is 60%. The chance that the patient will die by his treatment after correct diagnosis is 40% otherwise 65%. A patient treated by the doctor has died. What is the probability that the patient will was diagnosed correctly?
3. A consultancy firm has appointed three advisors A,B and C. They have advised 500 customers in a week. A has advised 200, B has advised 180 and C has advised 120. Advisor A being reported popular, 90% of the customers benefit from his advice. Corresponding figures for B and C are 80% and 75%. After a week a customer was selected at random and was found he was not benefitted. What is the probability he was advised by B?
Answers Bayes Theorem1. A company purchases raw material from 2 suppliers A1 and A2. 65% material comes from
A1 and the rest from A2. According to inspection reports, 98% material supplied by A1 is good and 2% is bad. Corresponding figures for supplier A2 and A1 are 95% and 5%. The material is selected at random and was tried on machine for processing. The machine failed because the material selected was bad or defective. What is the probability that it was supplied by A1?
Substitute these values in the formula for Bayes Theorem
P(A1) = 0.65 can have two outcome P (G/A1) i.e. 0.98 are good and P(B/A1) i.e. 0.02 are bad
P (G/A1) = P (A1) X P(G/A1) = 0.65 X 0.98 = 0.6370
P (B/A1) = P (A1) X P(B/A1) = 0.65 X 0.02 = 0.013
P(A2) = 0.35 can have two outcome P (G/A2) i.e. 0.95 are good and P(B/A1) i.e. 0.05 are bad
P (G/A2) = P (A2) X P(G/A2) = 0.35 X 0.95 = 0.3325
P (B/A2) = P (A2) X P(B/A2) = 0.05 X 0.35 = 0.0175
Examples to be solvedProbability – Survival Analysis
The probability that a 30 year old man will survive is 99% and insurance company offers to sell such a man a Rs 10,000 1 year term insurance policy at a yearly premium of Rs 110. What is the company’s expected gain?
Let x be the companies expected gain
X1 = Rs 110 corresponding probability (0.99) P1 man will survive
X2 = Rs 10000 + Rs 110 corresponding probability (0.01) P2
= - 9890
∑ p1 x1 = p 1 x x1 + p2 x2 = 0.01 x 110 + 0.01 (- 9890)
= 108.9 – 98.9 = 10