Learning Framework
Learning Framework
https://register.epo.org/espacenet/regviewer?AP=21834552&CY=EP&LG=en&DB=REG
https://patentscope.wipo.int/search/en
/detail.jsf?docId=WO2022003733&_fid
=US393708628
https://ppubs.uspto.gov/pubwebapp/external.html?q=(US-
20230083762-A1).did.&db=US-PGPUB
Unique Objectives : Why we need this Model / Framework
• Interpretable & flexibility over hypothesis space ( open to hand picked rule )
• Missing Value Imputation & Feature Engineering Tool (Band/NLP/Sequence/ DL – Adversarial Forest)
• Robust to sensitive data & High stability
• Multivariate interaction Extraction & stable operating conditions
• Combined Feature selection (Interaction – Resp index )& combined impact/ stability analysis
• Single model for Multi phase data - process cycle optimization (Applicable only for same sensor across multiple phase)
• Controllability vs optimization - optimal operating condition (with Both Side Definition - Min & Max)
• Multi criteria optimization - combined response optimization (multiple Y / Response)
• Anomaly tracking & detection - band/rule fluctuations ( RCA )
• Systematic tracking of anomaly - sensitive var Subset
• Auto EDA – Key Interaction highlight w.r.t Response
Topic Reference :
Response surface method
Information Theory
Bandit Algorithm for Supervised Learning
Tree, Bagging & Boosting Algorithm
Active Learning
Supervised Learning Module : Bandit Control Tree / Forest
Concepts :
Alternative to supervised learning model : Rule engine {similar algo to extract rule like Decision tree }
Raw Input Data
Bucket mapping from Similarity Pair : Do binning/ bucket ( quartile / vector discretization ) - each
run for all continuous variable for bucket mapping. Number of bin will be based on data driven technique.
like :
Based on Y var1 has : 5 bin, var2 has : 10 bin, var3 has : 15 bin
advantage : rule will include all variable and give a fixed cutoff
we can rank them based on %_match, and tot pair - higher is better
Using These
Aggregate trees for
/ Group by generating
scores from out
of bag sample
=A
Filter band tree based on information content & For each data point /obs
Apply split details (bucket
predictive power – IV, AUC, R2 /band definition) to map
specific band of a band tree
Take scaled score for ensemble model – weighted table to generate
score responses score
Basic eda - uva, Correlation & BVA analysis, features cluster Model Building Steps :
1) extracted band / rules from training - boundary details for Data dimension : 10M X 200 Input [Full Batch – 2000]
splitting criteria for all eligible vars Response : Unit Count, Yield, Quality Class
(bucket / band definition)
Split Data into 70% Train [FB-1400]– 30% Test [FB-
2) apply these rule & get response count / avg... From validation 600] [15% ITV, 15% OTV]
sample
For 7M Data points – 5M for Generating tree, 2M for
3) column join training band tree table & validation band tree OOB Validation
table – Anomaly Tracking & sensitivity analysis
Anomaly tracking detection as band / rule fluctuation Algorithm Sequence :
4) band score : measure residual, cov, Cross Entropy - relative, IV 1] Split & Bucket generation from data
conditional expectation for both 2] Data Transformation on Direct Input & Transformed Feature
training & validation response (oob validation & measurement) 3] Band Tree Table formation after applying different split & Vars Combination
4] Band Tree Table ensemble Prediction – Regression / Classification / Scoring
5) get band tree table score - psi, R2, Mae, CE, IV, sq err
5] Band Tree Table ensemble Prediction – Regularization
6) filter/Prune/Prioritize band & band tree table (based on score/ 6] Tune Hyperparameter of Learning Model
normalized ) 7] Model Diagnostics & Validation
Once Model is Ready & Running we can use it for
Anomaly Tracking & Band update [ Same (Split, vars & Bucket) combination
=> Band Tree Training vs New comparison ]
Hyperparameter of Learning Model
Hyperparameters of algo: Model Diagnostics :
Split range - 10 - 100 ( continues vars with specific granularity) Summary Table : List of Tree Table with all scoring & contribution,
[ Higher Split range - high variance model ] – stability issue in long run residual diagnostics details for weighting -ensemble
[ lower split range – low variance model ]
Residual Diagnostics – Confusion Matrix
AUC – ROC, R2, MSE, MAE … etc.
Split / bucket size – based on cardinality & few other( SNR, SD) factor from input vars (unique
pt. / tot obs)
Score quality checking, Rank order checking
Var Subset index - sqrt(tot vars), n/2…..etc Stability & Sensitivity Analysis – Deviation in previous vs Current
score & critical parameters
Split criteria –
binning / discretization : ChiMerge, MDLP, CAIM, khiops, Adaptive Further diagnostics – IML XAI ( Optional )
Quantizer, Mutual Information based : PMI, MIC,
change pt, cov / density / contour based…….etc Feature Importance( Forest level ) & Combined interaction
Importance
( Tree Level )
Band tree Table Weight score – IV,AUC, R2 ….etc
A) join training band tree & New band tree – Anomaly Tracking &
sensitivity analysis
Anomaly tracking detection as band / rule fluctuation
Hurricane
Irma Created
Some New Rule might appear in data & few old rule might come a New Island
Along the
insignificant due to fluctuation Georgia Coast
New Data with Band Tree A1
X1 x2 x4 x5 y1 y2
3 5 a 7 96 2
3 5 a 1 92 0
1 9 b 2 50 1
1 9 b 1 86 1
3 7 c 4 46 0
Soumyajit Das
[email protected]