Applied StatisticsforNeuroscientists Part IIa:...
Transcript of Applied StatisticsforNeuroscientists Part IIa:...
![Page 1: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/1.jpg)
Applied Statistics for NeuroscientistsPart IIa: Machine Learning
Dr. Seyed-Ahmad Ahmadi
04.04.201716.11.2017
![Page 2: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/2.jpg)
Outline – Machine Learning
• “Difference” between statistics and machine learning• Modeling the problem with (un-)supervised learning• Typical Machine Learning pipeline• Decision functions• Important supervised learning algorithms:
– kNN / SVM / Decision trees à Random Forests• Cross-validation• Concrete example from my research
– Anatomy locatlization in images with RF regression• Outlook:
– Clustering– Manifold learning
• Questions?
![Page 3: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/3.jpg)
Statistics vs. machine learning
According to Larry Wassermann(professor for dept.s for statistics and Machine Learning at Carnegie Mellon University)
“The short answer is: None. They are both concerned with the same question: how do we learn from data?”
• Statistics emphasizes formal statistical inference (confidence intervals, hypothesis tests, optimal estimators) in low dimensionalproblems.– Survival analysis, spatial analysis, multiple testing, minimax theory, deconvolution, semiparametric inference, bootstrapping, time series
• Machine Learning emphasizes high dimensional prediction problems.– Online learning, (semi-)supervised learning, manifold learning, active learning, boosting
![Page 4: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/4.jpg)
Mindmap of ML algorithms
Adapted from: https://machinelearningmastery.com
![Page 5: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/5.jpg)
ML categories: Modeling your problem
Classification
Regression
Clustering
Manifold learning
Supervised Learning Unsupervised Learning
Semi-supervised / Reinforcement LearningAdapted from: https://machinelearningmastery.com
![Page 6: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/6.jpg)
Supervised Learning: Classification and regression
y categorical: classification
y continuous: regression
Adapted from: Andrew Ng, CS 229, fall 2012, Stanford university
![Page 7: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/7.jpg)
A typical ML pipeline
Healthy cells
Cancer cells
Image normalizationDe-noising
Cell localizationROI cropping
...
Feature extraction:Data acquisition:
„Cell roundness“
„Mitosis grade“
Features: x ∈ X ⊂ Rn
Classes: c ∈C∪ 0,1[ ]
Pre-processing: Machine learning:
Inference forpreviously unseendata points
???
c = f x( )
![Page 8: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/8.jpg)
Decision functions – „hand-made“
„Cell roundness“
„Mitosis grade“
Rule-based algorithm / Decision tree• If mitosis > 0.5: cancer
– If roundness >0.6: healthy
• Else – If 0.4 < roundness < 0.5: cancer
0 1
1
![Page 9: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/9.jpg)
Decision trees
P. Geurts et al., Supervised learning with decision tree-based methods in computational and systems biology, Mol. BioSyst. 5 (2009) 1593–1605
![Page 10: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/10.jpg)
Decision functions• A ML algorithm tries toautomatically find an optimal decision boundary to separate classes
• The decision boundary is thethreshold of a decision function
• A ML algorithm fits the decisionfunction to the data in order to find the optimal boundary
• Devision boundary (on the right) in 3D: http://biostatmatt.com/HTF-Figure-5.11b/
• NB: The decision boundary is:– Maximally separating– Highly non-linear
![Page 11: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/11.jpg)
Different classifiers and their decision boundaries
Adapted from: http://scikit-learn.org/stable/auto_examples/classification/plot_classifier_comparison.html
Instance based Kernel based Rule based Ensemble methods
![Page 12: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/12.jpg)
Machine learning helps with high-dimensional, complexproblems
http://stackoverflow.com/questions/10266195/plot-svm-in-3-dimension
Decision function visualized for3-dimensional feature space
Classification of 32 x 32 cell images: 1024 dimensional feature space
e.g. Javidi, Image Recognition and Classification: Algorithms, Systems, and Applications, p. 462, CRC press, 2002
Segmentation of 128 x 128 x 64 MRI volumes:1.048.576 dimensional feature space
Milletari et al., V-Net: FCNNs for VolumetricMedical Image Segmentation, 3DV, 2016
![Page 13: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/13.jpg)
Classification vs. Regression
Classification Regression
???
x1
x2
X (Rn)
Y (Rm)
c = f x( ) y = f x( )
![Page 14: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/14.jpg)
Challenges
• Modeling the problem• Choosing the best learner• Hyper-parameter tuning• Optimization!!!• Overfitting
https://upload.wikimedia.org/wikipedia/commons/1/19/Overfitting.svg
![Page 15: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/15.jpg)
Overfitting
P. Geurts et al., Supervised learning with decision tree-based methods in computational and systems biology, Mol. BioSyst. 5 (2009) 1593–1605
![Page 16: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/16.jpg)
K-Nearest Neighbors (kNN): “Intuitive idea”
• “A point’s class should be determined by its nearest neighbors”.
„Cell roundness“
„Mitosis grade“
?
?
![Page 17: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/17.jpg)
K-Nearest Neighbors (kNN): Training
• kNN does not require an explicit “training”• “Training” usually means partitioning the space to make NN searches faster and computationally more efficient
https://en.wikipedia.org/w/index.php?title=File:KDTree-animation.gif
![Page 18: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/18.jpg)
K-Nearest Neighbors (kNN): Testing (regression)
• Classification: majority vote of classes from neighbors• Regression: average the targets from neighbors
http://scikit-learn.org/stable/auto_examples/neighbors/plot_regression.html
1D àà 1D
2D àà 3D
![Page 19: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/19.jpg)
K-Nearest Neighbors (kNN): Parametrization
• Classification (regression) depends on parameter k
k=1 k=5 k=10
![Page 20: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/20.jpg)
Support Vector Machines: “Intuitive idea”• “A few carefully selected key points at the boundary between two training
classes should be enough to determine an unknown point’s class.”Careful selection à “optimally separating hyperplane”
Key points à “support vectors”
Image: https://en.wikipedia.org/wiki/Support_vector_machine#/media/File:Svm_separating_hyperplanes_(SVG).svgFor a really beautiful derivation of SVM, watch MIT AI course 6.034 (2010) on https://youtu.be/_PwhiWxHK8o
𝐻𝐻 𝑥 → > 0 → 𝑐𝑐𝑐𝑐𝑐𝑐𝑐𝑐𝑐𝑐 1< 0 → 𝑐𝑐𝑐𝑐𝑐𝑐𝑐𝑐𝑐𝑐 0
Classification at test time:
Finding parameters for H is based on mathematical constraints and involves an iterative optimization given all training data to find
support vectors.
![Page 21: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/21.jpg)
Support Vector Machines: Training with soft-margin• Soft-margin: Given non-separability, how do we still find the separating plane?
• C: complexity parameter which determines how much “slack” is given to support vectors (especially mis-classified ones)
• C large: find a hard separation plane• C small: find a soft separation plane, while allowing many mis-classifications (useful when we do not “trust” the distribution of our training data)
© Yaroslav Butalov, http://yaroslavvb.com/upload/save/so-libsvm.png
![Page 22: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/22.jpg)
Support Vector Machine: Non-linear classification?
• No linear hyperplane (i.e. no straight line in 2D) will be able to separate these two classes…
![Page 23: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/23.jpg)
Support Vector Machine: Non-linear classification with the “Kernel trick”
© udiprod, Youtube (upload Feb 5, 2007;; last access Mar 30, 2017)https://www.youtube.com/watch?v=3liCbRZPrZA
![Page 24: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/24.jpg)
Support Vector Machines: Parametrization
LinearSVM
Non-linearSVM(Gaussian kernel)
![Page 25: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/25.jpg)
Decision trees and Random Forests: Principle
• Intuitive idea behind decision trees:“The class of an unknown point can be determined by a series of rules/decisions”
P. Geurts et al., Supervised learning with decision tree-based methods in computational and systems biology,
Mol. BioSyst. 5 (2009) 1593–1605
• Problem: Single decision trees overfit quickly (i.e. generalize badly)
• Intuitive idea behind Random Forests:“Improve classification by training several decision trees on different aspects of our training data, and combine them into one common ensemble – a ‘forest’.”
• “Where one tree fails, the others do well.”
![Page 26: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/26.jpg)
Random Decision Forests: Principle
P. Geurts et al., Supervised learning with decision tree-based methods in computational and systems biology, Mol. BioSyst. 5 (2009) 1593–1605
Randomizabletraining aspects: • Training data• Subset of dimensions at each split
• Random split decisions(extremely randomized trees)
Randomization provides excellent
regularization!
![Page 27: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/27.jpg)
Random Decision ForestsRandomizable training aspects: • Training data• Subset of dimensions at each split
• Random split decisions(extremely randomized trees)
Randomization is a very effective means to achieve
regularization!
![Page 28: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/28.jpg)
Random Decision Forests: Parametrization
N=1
• Classification depends e.g. on number of trees N
N=3 N=10
![Page 29: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/29.jpg)
Which classifier/regressor to choose?
kNN SVM Random ForestRequired dimensionality
Small Medium Very high
Required dataset size
Medium(acceleration due to parcellation e.g. with kdtrees)
Small(training on large datasets can be painfully slow)
Very large
Efficiency Very slow to trainon >lab-size datasets
Highly efficient, parallelizable
![Page 30: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/30.jpg)
Many more choices…
Adapted from: http://scikit-learn.org/stable/auto_examples/classification/plot_classifier_comparison.html
![Page 31: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/31.jpg)
Statistical robustness via cross-validation
Dataset splittingFold 1Fold 2Fold 3Fold 4
Train Test
• Achieve robustness in machine learning validation via cross-validation or
• Splitting into training-/validation-/test-dataset• Give confidence intervals for classification accuracy
• Give a p-value via label permutation tests
![Page 32: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/32.jpg)
Statistical robustness via validation and test set
Dataset splittingSingle fold Training set Validation set Test set
• Useful when training of folds takes very long– Complex models, e.g. deep neural networks– Extremely large datasets, e.g. ImageNet in Computer Vision
The validation set is used • to observe generalizability of the model to unknown data
• to tune the parameters of the network• to observe convergence behaviourof the optimization
Image from: CS321nhttp://cs231n.github.io/neural-networks-3/
![Page 33: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/33.jpg)
Machine learning in medical imaging: Anatomy localization with random forest classification
Criminisi et al., Decision Forests with Long-Range Spatial Context for Organ Localization in CT Volumes, MICCAI Workshop on Probabilistic Models for Medical Image Analysis (MICCAI-PMMIA), Sept. 2009
![Page 34: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/34.jpg)
Machine learning in medical imaging: Anatomy localization with random forest classification
Criminisi et al., Decision Forests with Long-Range Spatial Context for Organ Localization in CT Volumes, MICCAI Workshop on Probabilistic Models for Medical Image Analysis (MICCAI-PMMIA), Sept. 2009
![Page 35: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/35.jpg)
Machine learning in medical imaging: Anatomy localization with random forest classification
Criminisi et al., Decision Forests with Long-Range Spatial Context for Organ Localization in CT Volumes, MICCAI Workshop on Probabilistic Models for Medical Image Analysis (MICCAI-PMMIA), Sept. 2009
![Page 36: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/36.jpg)
Machine learning in medical imaging: Anatomy localization with random forest regression
Criminisi et al., Regression Forests for Efficient Anatomy Detection and Localization in CT Studies, International MICCAI Workshop on Medical Computer Vision (MICCAI-MCV), LNCS vol. 6533, pp. 106-117, 2010
![Page 37: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/37.jpg)
Machine learning in medical imaging: Anatomy localization with random forest regression
Criminisi et al., Regression Forests for Efficient Anatomy Detection and Localization in CT Studies, International MICCAI Workshop on Medical Computer Vision (MICCAI-MCV), LNCS vol. 6533, pp. 106-117, 2010
![Page 38: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/38.jpg)
Machine learning in medical imaging: Anatomy localization with random forest regression
Criminisi et al., Regression Forests for Efficient Anatomy Detection and Localization in CT Studies, International MICCAI Workshop on Medical Computer Vision (MICCAI-MCV), LNCS vol. 6533, pp. 106-117, 2010
![Page 39: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/39.jpg)
Machine learning in medical imaging: Anatomy localization with random forest regression
Criminisi et al., Regression Forests for Efficient Anatomy Detection and Localization in CT Studies, International MICCAI Workshop on Medical Computer Vision (MICCAI-MCV), LNCS vol. 6533, pp. 106-117, 2010
![Page 40: Applied StatisticsforNeuroscientists Part IIa: MachineLearningsglasauer.userweb.mwn.de/Ahmadi_17_ASN_01_MachineLearning_w… · Applied StatisticsforNeuroscientists Part IIa: MachineLearning](https://reader030.fdocuments.in/reader030/viewer/2022040112/5ec4d39922367157d34e4759/html5/thumbnails/40.jpg)
German Center for Vertigo and Balance Disorders