L4-ML Introduction To Machine Learning Algorithms 2024-01
L4-ML Introduction To Machine Learning Algorithms 2024-01
5 6
Session 1
7 8
Churn Prediction
Model
9 10
1
21/06/24
O ct O ct § High Risk
Low Risk
2015 2015
Nov
C v re rh fd2 0 1 6
Nov
C v re rh fd2 0 1 6 §
High Risk
V d s y h CO V d s y h C v re rh fd
d fg h V2d0
c trh fd
v re
s y1h5 Nov
d fg h Vd syh
§
dJdugnd
2 0 1 7 dV
dCfgvhre rh fd2 0 1 6
dd
dJdugnd
g sdy h C v re rh fd 2 0 1 7 d d g d
d fg h § Very High Risk
CRM System § Churn Prediction A
C v re rh fd d fg h
Vd syh dJ2du0
p
gn1
r
Vd syh
d 8 d fg h
C v re rh fd A p r
Vd syh 2018 § Very Low Risk
Data about your customer § Upselling Likelihood dOfgcht 2C0v1 re7rh fd
Fdedbg d d fg h C v re rh fd
Feb § Medium Risk
d2d0g1d5 NCVo vdre
vs yrh
h fd A p r ddgd Vd syh
§ … Vd syh 2018
C v re rh fd
d fg h
ddgd
d fg h
Feb
ddgd Vd syh
Model Model
2019
d fg h
C v re rh fd
ddgd
Vd syh
d fg h
ddgd
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 11 Material of KNIME AG used under CC BY 4.0 12
11 12
Recommendation
IF è
Model Model
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 13 Material of KNIME AG used under CC BY 4.0 14
13 14
Suspicious Transaction
Transactions
§ Trx 1
§ Trx 2
§ Trx 3
§ Trx 4
§ Trx 5
§ Trx 6
§ …
Model
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 15 Material of KNIME AG used under CC BY 4.0 16
15 16
2
21/06/24
Anomaly Detection
Predicting mechanical failure as late as possible but before it happens
A 1 -SV 3 [0 , 1 0 0 ]
Hz
3 1 A u g u st
2007
Basic Concepts in Data Science
Predictive B reakin g
A 1 -SV 3 [5 0 0 , 6 0 0 ]
Hz
Maintenance p o in t
Ju ly 2 1 , 2 0 0 8
Training Set
via REST
Only some Spectral Time Series shows the break down
These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 17
17 19
§ The model learns / is trained during the learning / training phase to produce the right
§ Class answer y (a.k.a., label)
§ Input features § Label
§ Input attributes 𝑦 § Target
§ Independent variables § Output feature/attribute § That is why machine learning J
§ Dependent variable
§ Many different algorithms for three ways of learning:
Model § Supervised
§ Unsupervised
Model parameters § Semi-supervised
𝑦 = 𝑓( 𝜷, 𝑿 ) with 𝜷 = [𝛽 1, 𝛽 2, … , 𝛽 𝑚]
20 21
§ 𝑿 = (𝑥1, 𝑥2) and 𝑦 = {𝑦𝑒𝑙𝑙𝑜𝑤, 𝑔𝑟𝑎𝑦} § 𝑿 = (𝑥1, 𝑥2) and 𝑦 = {𝑙𝑎𝑏𝑒𝑙 1, … , 𝑙𝑎𝑏𝑒𝑙 𝑛} or 𝑦 ∈ ℝ
§ A training set with many examples of (𝑿, 𝑦) § A training set with many examples of (𝑿, 𝑦)
§ The model learns on the examples of the training set to produce the right value of y for § The model learns on the examples of the training set to produce the right value of 𝑦 for
an input vector 𝑿 an input vector 𝑿
x2
𝑿 model 𝑦 Classification Numerical Predictions (Regression)
Age y = {yellow, gray} y = temperature
Money Sunny vs. Cloudy y = {churn, no churn} y = number of visitors
Temperature Healthy vs. Sick y = {increase, unchanged, decrease} y = number of kW
Speed Churn vs. Remain
Number of taxi Increase vs. y = {blonde, gray, brown, red, black} y = price
... Deacrease y = {job 1, job 2, ... , job n} y = number of hours
...
x1
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 22 Material of KNIME AG used under CC BY 4.0 23
22 23
3
21/06/24
x1
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 24 Material of KNIME AG used under CC BY 4.0 26
24 26
x2
X
Age
Money
Temperature
Speed
Number of taxi
...
https://en.wikipedia.org/wiki/Cross_Industry_Standard_
x1 Process_for_Data_Mining
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 27 Material of KNIME AG used under CC BY 4.0 28
27 28
The Data Science Life Cycle KNIME Software for the entire Data Science Life Cycle
Blend & Transform Validate & Deploy
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 29 Material of KNIME AG used under CC BY 4.0 30
29 30
4
21/06/24
sunny 3 30 no yes
sunny 3 25 no no
rain 12 15 no yes
overcast 15 2 yes no
rain 16 25 no yes
sunny 14 18 no yes
sunny 9 20 no yes
overcast 14 5 yes no
sunny 1 7 yes no
rain 4 25 no no
rain 14 24 no yes
sunny 11 20 no yes
sunny 2 18 no no
overcast 8 22 no yes
overcast 13 24 no yes
These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 32
31 32
How can we Train a Decision Tree with KNIME Analytics Platform Goal: A Decision Tree
Outlook W ind Temp Storage Sailing Option 1
sunny 3 30 yes yes
sunny 3 25 yes no
overcast 15 2 no no
overcast 14 5 no no
sunny 1 7 no no
rain 4 25 yes no
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 33 Material of KNIME AG used under CC BY 4.0 34
33 34
Possible Split Criterion: Gain Ratio Possible Split Criterion: Gain Ratio
Based on entropy = measure for information / uncertainty
Split criterion:
𝐺𝑎𝑖𝑛 = 𝐸𝑛𝑡𝑟𝑜𝑝𝑦<=>?@= − 𝐸𝑛𝑡𝑟𝑜𝑝𝑦A>B=@
𝐸𝑛𝑡𝑟𝑜𝑝𝑦 𝑝 = − ∑'$%& 𝑝$ log ( 𝑝$ for 𝑝 ∈ ℚ'
, *
𝐺𝑎𝑖𝑛 = 𝐸𝑛𝑡𝑟𝑜𝑝𝑦<=>?@= − )+ 𝐸𝑛𝑡𝑟𝑜𝑝𝑦) − )+ 𝐸𝑛𝑡𝑟𝑜𝑝𝑦(
𝐸𝑛𝑡𝑟𝑜𝑝𝑦89*+-9
7 4
= 𝐸𝑛𝑡𝑟𝑜𝑝𝑦 35 , 35
𝐸𝑛𝑡𝑟𝑜𝑝𝑦 𝑝 = − *⁄)+ log( *⁄)+ + ,⁄)+ log( ,⁄)+ 𝐸𝑛𝑡𝑟𝑜𝑝𝑦 𝑝 = − )+⁄ log )+⁄
)+ ( )+ + &⁄)+ log( &⁄)+ Solution: Gain Ratio
= 0,995 =0
!"#$ ,$(-+&.! " # $ % " /∑)&' ( 1& ,$(-+&.&
𝐺𝑎𝑖𝑛𝑅𝑎𝑡𝑖𝑜 = = : 3 6 :
%&'#()$*+ / ∑)&' ( 1& '+2* 1& 𝐸𝑛𝑡𝑟𝑜𝑝𝑦3 = 𝐸𝑛𝑡𝑟𝑜𝑝𝑦 , 𝐸𝑛𝑡𝑟𝑜𝑝𝑦6 = 𝐸𝑛𝑡𝑟𝑜𝑝𝑦 ,
4 4 7 7
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 35 Material of KNIME AG used under CC BY 4.0 36 𝑤3 = 4⁄35 𝑤6 = 7⁄35
35 36
5
21/06/24
Possible Split Criterion: Gini Index What happens for numerical Input Features?
Subset for each value? – NO
Gini index is based on Gini impurity:
𝑝3 = 7⁄35 Solution: Binary splits
𝐺𝑖𝑛𝑖(𝑝) = 1 − ∑ ' (
$%) 𝑝$ for 𝑝 ∈ ℚ '
𝑝6 = 4⁄
35
7( 6(
𝐺𝑖𝑛𝑖 𝑝 = 1 − ( − (
13 13
Split criterion:
𝐺𝑖𝑛𝑖C'D=E = ∑ '
$%) 𝑤$ 𝐺𝑖𝑛𝑖$
, * 𝑥 = 1.2 𝑥=<
3.4𝑥 𝑥 = 1.7 𝑥 = 3.6
≥𝑥 𝑥 = 4.9
𝐺𝑖𝑛𝑖C'D=E = 𝐺𝑖𝑛𝑖) + 𝐺𝑖𝑛𝑖( 𝑥 = 9.2 𝑥=2 𝑥 = 12.6 𝑥 = 7.4 𝑥=8 𝑥 = 2.3
)+ )+
𝐺𝑖𝑛𝑖3Next :⁄ 3⁄
= 𝐺𝑖𝑛𝑖splitting
4, 4 feature: 𝐺𝑖𝑛𝑖6 = 𝐺𝑖𝑛𝑖 6⁄7, :⁄7
Feature with lowest 𝐺𝑖𝑛𝑖C'D=E 𝑤3 = 4⁄35 𝑤6 = 7⁄35
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 37 Material of KNIME AG used under CC BY 4.0 38
37 38
30 temp temp
≥ 10 < 10 ≥ 25 < 25
25
temp wind
20
≥ 22 < 22 ≥6 <6
15
temp
≥ 26 < 26
10
Model memorizes the
5
wind Model overlooks Model captures training set rather then
≥6 <6 underlying patterns correlations in the
1 2 3 4 5 6 7 𝑤𝑖𝑛𝑑
finding underlying
in the training set training set patterns
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 39 Material of KNIME AG used under CC BY 4.0 40
39 40
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 41 Material of KNIME AG used under CC BY 4.0 42
41 42
6
21/06/24
Pruning - Minimum Description Length Pruning (MDL) Applying the Model – What are the Outputs?
Definition: Description length = #bits(tree) + #bits(misclassified samples)
wind wind
Many misclassified
Example 1
samples in tree 1
temp
12 0 12 0 è DL(Tree 1) > DL(Tree 2)
6 7 è Select Tree 2
sample in tree 1
temp
12 0 1 13 12 0 è DL(Tree 1) < DL(Tree 2)
è Select Tree 1
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 43 Material of KNIME AG used under CC BY 4.0 44
43 44
oving a branch with a small improvement in performance
smaller tree is preferred
rain 3 5 no no
What happens with outlook = overcast?
sunny 9 20 yes yes Evaluation of Classification Models
sunny 1 7 no no
rain 4 25 yes no
sunny 2 18 yes no
45 46
§ Downsides:
§ Only considers the performance in general and not for the different classes
§ Therefore, not informative when the class distribution is unbalanced
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 47 Material of KNIME AG used under CC BY 4.0 48
47 48
7
21/06/24
49 50
51 52
Course exercises
§ All exercises are available here:
Session 1 exercises
53 54
8
21/06/24
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 55 Material of KNIME AG used under CC BY 4.0 56
55 56
You can download the training workflows from the KNIME Community Hub at this link
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 57 Material of KNIME AG used under CC BY 4.0 58
57 58
59 60
9
21/06/24
Regression Overview
§ Goal: Explain how target attribute depends on descripitive attributes
§ Target attribute è Response variable
§ Descriptive attribute(s) è Regressor variable(s)
61
61 62
Regression
Predict numeric outcomes on existing data (supervised)
Applications
§ Forecasting
§ Quantitative Analysis
Methods
§ Linear
Linear Regression Algorithm
§ Polynomial
§ Regression Trees
§ Partial Least Squares
63 64
Intercept 𝑎 x
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 65 Material of KNIME AG used under CC BY 4.0 66
65 66
10
21/06/24
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
x
Material of KNIME AG used under CC BY 4.0 67 Material of KNIME AG used under CC BY 4.0 68
67 68
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 69 Material of KNIME AG used under CC BY 4.0 70
69 70
§ Negative:
§ Many dependencies are non-linear § Simple regression: one input feature à regression curve
(Can be generalized)
§ Multiple regression: several input features à regression hypersurface
§ Model is global and cannot adapt well to locally different data distributions
But: Locally weighted regression, CART § Residuals: differences between observed and predicted values (errors)
Use the residuals to measure the model fit
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 71 Material of KNIME AG used under CC BY 4.0 72
71 72
11
21/06/24
$
Mean signed difference 1 Only informative about the direction
g 𝑦# − 𝑓 𝑥# of the error
𝑛
#=3
$
Mean absolute percentage error 1 |𝑦# − 𝑓(𝑥# )| Requires non-zero target column
(MAPE) g values
𝑛 |𝑦# |
#=3
74 75
MAE (Mean Absolute Error) vs. RMSE (Root Mean Squared Error) R-squared vs. RMSE
MAE RMSE R-squared RMSE
Easy to interpret – mean absolute error Cannot be directly interpreted as the average error Relative measure: Absolute measure:
Proportion of variability explained by the model How much deviation at each point
All errors are equally weighted Larger errors are weighted more Range: Usually between 0 and 1. Same scale as the target
0 = no variability explained
Generally smaller than RMSE Ideal when large deviations need to be avoided 1 = all variability explained
Example: Example:
Actual values = [2,4,5,8], MAE RMSE Actual values = [2,4,5,8], R-sq RMSE
Case 1: Predicted Values = [4, 6, 8, 10] Case 1 2.25 2.29 Case 1: Predicted Values = [3, 4, 5, 6] Case 1 0.96 1.12
Case 2: Predicted Values = [4, 6, 8, 14] Case 2: Predicted Values = [3, 3, 7, 7]
Case 2 3.25 3.64 Case 2 0.65 1.32
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 76 Material of KNIME AG used under CC BY 4.0 77
76 77
Numeric Scorer
§ Similar to scorer node, but for nodes with numeric predictions
Regression Tree
78 79
12
21/06/24
x s x
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 80 Material of KNIME AG used under CC BY 4.0 81
80 81
Y N segment Y N
Y N
𝐶+ = 33.9 𝐶, = 26.4
s x s x
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 82 Material of KNIME AG used under CC BY 4.0 83
82 83
x
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 84 Material of KNIME AG used under CC BY 4.0 85
84 85
13
21/06/24
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 86 Material of KNIME AG used under CC BY 4.0 87
86 87
2 2
4
7 … 7
1
88 89
…
1 4 1
5 2 5 7 7 6
2 9 6 7 2 8 9 3 3 9 5 7
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 90 Material of KNIME AG used under CC BY 4.0 91
90 91
14
21/06/24
… …
1 4 1 1 4 1
5 2 2 7 7 6 5 2 2 7 7 6 Build tree
2 9 6 7 6 8 9 3 3 9 5 7 2 9 6 7 6 8 9 3 3 9 5 7
P1 P2 … Pn P1 P2 … Pn
y 1OOB y 2OOB 5 2
2 9 6 7
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 92 Material of KNIME AG used under CC BY 4.0 93
92 93
…
1 4 1
5 2 5 7 7 6
2 9 6 7 2 8 9 3 3 9 5 7
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 94 Material of KNIME AG used under CC BY 4.0 95
94 95
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 96 Material of KNIME AG used under CC BY 4.0 97
96 97
15
21/06/24
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 98 Material of KNIME AG used under CC BY 4.0 99
98 99
100 101
Why Shouldn’t we Always use the Decision Tree? Decision Boundary of a Logistic Regression
? ?
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 102 Material of KNIME AG used under CC BY 4.0 103
102 103
16
21/06/24
Linear Regression vs. Logistic Regression Let’s find out how Binary Logistic Regression works!
§ Idea: Train a function, which gives us the probability for each class (0 and 1) based on
Linear Regression Logistic Regression the input features
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 104 Material of KNIME AG used under CC BY 4.0 105
104 105
Let’s Find Out How Binary Logistic Regression Works! More General: Binary Logistic Regression
§ Model:
1
𝑃 𝑦 = 1 = 𝑓 𝑥) , 𝑥( ; 𝛽& , 𝛽) , 𝛽( ≔ 𝜋 = 𝑃(𝑦 = 1) =
)
1 + 𝑒 ^(l]mlnE nmloE o) )mpqr(^s)
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 106 Material of KNIME AG used under CC BY 4.0 107
106 107
How can we Find the Best Coefficients β? Max Likelihood and Log Likelihood Functions
Maximize the product of the probabilities è Likelihood function § Maximize the Likelihood function 𝐿 𝜷; 𝒚, 𝑿
} }
yt z{y t }
𝐿 𝛽; 𝑦, 𝑋 = 5 𝑃(𝑦 = 𝑦x ) = 5 𝜋x 1 − 𝜋x yt z{y t
x|z x|z max 𝐿 𝛽; 𝑦, 𝑋 = max 5 𝜋x 1 − 𝜋x
€ € x|z
𝜋x 𝑖𝑓 𝑦x = 1 •
𝑃 𝑦 = 𝑦x = )1 − 𝜋
x 𝑖𝑓 𝑦x = 0 Remember: max 𝐿𝐿(𝜷; 𝒚, 𝑿) = max > 𝑦x ln 𝜋x + 1 − 𝑦x ln 1 − 𝜋x
€ € x|z
𝜋$ = P 𝑦 = 1
yt z{y t 𝑢& = 1 for 𝑢 ∈ ℝ
= 𝜋x 1 − 𝜋x 𝑢) = 𝑢 for 𝑢 ∈ ℝ
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 108 Material of KNIME AG used under CC BY 4.0 109
108 109
17
21/06/24
Optimal 𝛽T
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 110 Material of KNIME AG used under CC BY 4.0 111
110 111
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 112 Material of KNIME AG used under CC BY 4.0 113
112 113
𝑙 𝛽; T 𝑦, 𝑋 + v ||𝛽||
T 𝑦, 𝑋 ≔ −𝐿𝐿 𝛽; T ((
(
(
§ 𝐿) regularizations = Coefficients are Laplace distributed with 𝜎 = v
• 𝑦, 𝑋 ≔ −𝐿𝐿 𝛽;
𝑙 𝛽; • 𝑦, 𝑋 + 𝜆||𝛽||
• )
&
=> The smaller 𝜎, the smaller the coefficients 𝛽
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 114 Material of KNIME AG used under CC BY 4.0 115
114 115
18
21/06/24
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 116 Material of KNIME AG used under CC BY 4.0 117
116 117
Session 2 exercises
118 119
Exercises
§ Regression Exercises:
§ Goal: Predicting the house price
§ 01_Linear_Regression_exercise
§ 02_Regression_Tree_exercise
§ Classification Exercises: [L4-ML] Introduction to Machine
§ Goal: Predicting the house condition (high /low) Learning Algorithms
§ 03_Random_Forest_exercise (with optional exercise to build a parameter optimization loop)
§ 04_Logistic_Regression_exercise
Session 3
120 121
19
21/06/24
122 123
𝑜)(
∑ 𝑓 𝒐 = 𝑓 𝑊E( 𝒙
6
𝑊3,3 5
6 𝑊3,3
𝑥) 𝑊3,6 𝑦 = 𝑓(𝑊y+ 𝒐)
Artificial Neural Networks
6
𝑜((
Artificial Neuron (Perceptron) (Multilayer Perceptron, MLP) 𝑊6,3
∑ 𝑓 5
𝑊3,6 ∑ 𝑓 𝑦
6
𝑊6,6
𝑦 = 𝑓(𝑥3 𝑤3 + 𝑥 6 𝑤 6 + 𝑏) 𝑥)
𝑥) 𝑏 𝑥( 6
𝑊5,3 5 fully connected
𝑤3 𝑥( 𝑊3,5
𝑏 = 𝑤< y 6
𝑊5,6 𝑜+( feed forward
∑f( )σ y $
𝑤6 𝑦 = 𝑓(g 𝑥 # 𝑤 # )
𝑥+ ∑ 𝑓
𝑥( #=<
𝑥U
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 124 Material of KNIME AG used under CC BY 4.0 125
124 125
𝑥) 𝑦 = 𝑓(𝑊y+ 𝒐)
𝑜((
∑ 𝑓 ∑ 𝑓 𝑦
𝑥( f( ) = activation function
𝑜+( 1 𝑒 (}~ − 1
𝑓 𝑎 = 𝑓 𝑎 = 𝑓 𝑎 = 𝑚𝑎𝑥 0, ℎ𝑎
∑ 𝑓 1 + 𝑒 ^}~ 𝑒 (}~ + 1
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 126 Material of KNIME AG used under CC BY 4.0 127
126 127
20
21/06/24
Example: Passing the KNIME L1-Certification Example: Passing the KNIME L1-Certification
Input Hidden Output
Layer Layer Layer
-0.566
Workflow builds
Passed certification ∑ 𝑓)
-0.044 2.275
Didn’t pass certification 𝑥) -1.41
-0.608
∑ 𝑓( 𝑦D
-1.431
-0.513
𝑥( -1.298
1.0733
∑ 𝑓)
128 129
Example: Passing the KNIME L1-Certification Example: Passing the KNIME L1-Certification
Input Hidden Output
Layer Layer Layer
This prediction looks wrong.
-0.566 Why does the network
predict this low probability?
Workflow builds
Passed certification ∑ 𝑓)
-0.888
Normalize -0.044 2.275
Didn’t pass certification 170 𝑥)
0.567
-1.41
-0.608
New sample ∑ 𝑓(
0.013 𝑦D
0.013
Normalize -1.431
-0.513
𝑥)= minutes attended = 170 8 𝑥(
0.8
-1.298
1.0733
130 131
Training a Neural Network = Finding Good Weights Learning Rule from Gradient Descent
Input Hidden Output § Adjust the weight for the next step by the adjustment term ∆𝒘(𝑡)
Layer Layer Layer
𝒘 𝑡 + 1 = 𝒘 𝑡 + 𝜂 ∆𝒘(𝑡)
∑ 𝑓)
170 𝑥)
Predicted 𝑦f = 0.013
∑ 𝑓( 𝑦D True y = 1
∆𝒘(𝑡)
Loss function
8 𝑥( Weight in
∑ 𝑓) current step 𝑡
Optimal
Binary cross entropy Weight
solution
ℒ = −(𝑦 log 𝑦f + (1 − 𝑦) log(1 − 𝑦))
f Updated weight
adjustment based
in next step 𝑡 + 1
𝐽 𝑊 = + ℒ (𝑦# 𝑥A, 𝑥B, 𝑊 , 𝑦) on gradient
Weight 𝒘
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 132 Material of KNIME AG used under CC BY 4.0 133
132 133
21
21/06/24
Forward pass
𝑤) 𝑤(
𝑥 𝑧 𝑦D 𝐽(𝑦,
D 𝑦; 𝑤) , 𝑤( )
𝐽(𝑤) , 𝑤( )
Backward pass
𝑤) 𝑤(
𝑥 𝑧 𝑦D 𝐽(𝑦,
D 𝑦; 𝑤) , 𝑤( )
𝑤(
ۥ
𝑤)
€‚O 𝜕𝐽 𝜕𝐽 𝜕𝑦D 𝜕𝐽 𝜕𝐽 𝜕𝑦D 𝜕𝑧
∇ • 𝐽 𝑥, 𝑊 = €• = ∗ = ∗ ∗
€‚P 𝜕𝑤( 𝜕𝑦D 𝜕𝑤( 𝜕𝑤) 𝜕𝑦D 𝜕𝑧 𝜕𝑤)
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 134 Material of KNIME AG used under CC BY 4.0 135
134 135
Example: Passing the KNIME L1 Certification Loss Landscape of a Real Neural Network
Input Hidden Output
Layer Layer Layer § A good choice for the step size 𝜂, aka
learning rate, is important
-0.106
𝑊 ← 𝑊 − 𝜂 ∇ • 𝐽(𝑥, 𝑊)
∑ 𝑓)
0.396
Normalize -0.554 0.931
170 𝑥)
0.567
-2.603
0.117
§ Many different optimizers with
∑ 𝑓(
0.967 𝑦D
0.967 adaptive learning rates are available
Normalize 0.146
1.309 § Adam, Adadelta, Adagrad, ect
8 𝑥(
0.8
-1.554
-3.096
∑ 𝑓)
-0.929
§ Other important settings
§ Batch size, aka number of samples for one
𝑓) = 𝑡𝑎𝑛ℎ 𝑓( = 𝑠𝑖𝑔𝑚𝑜𝑖𝑑 update
Input features: Output: § Number of epochs, aka how often each
𝑥)= minutes attended 𝑦 = Probability that a person passed sample is used
Source: https://www.cs.umd.edu/~tomg/projects/landscapes/
𝑥(= workflows build 𝑦 ≥ 0.5 ⇒ 𝑃𝑎𝑠𝑠𝑒𝑑 and 𝑦 < 0.5 ⇒ 𝐹𝑎𝑖𝑙𝑒𝑑
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 136 Material of KNIME AG used under CC BY 4.0 137
136 137
Sigmoid
Softmax
-Computationally
-Different learning
Linear
Hinge
ReLU
MAE
Tanh
Tanh
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 138 Material of KNIME AG used under CC BY 4.0 139
138 139
22
21/06/24
Codeless Deep Learning with KNIME Analytics Platform Codeless Deep Learning with KNIME Analytics Platform
§ Simple option for feed forward neural networks with activation function sigmoid
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 140 Material of KNIME AG used under CC BY 4.0 141
140 141
142 143
What are Recurrent Neural Networks? Why do we need RNNs for Sequential Data?
§ Recurrent Neural Network (RNN) are a family of neural networks used for processing of § Goal: Translation network from German to English
sequential data ∑σ
“Ich mag Schokolade” ∑σ
§ RNNs are used for all sorts of tasks:
§ Language modeling / Text generation => “I like chocolate” ∑σ
§ Text classification 𝑥 ∑σ ∑σ 𝑦
§ Neural machine translation § One option: Use feed forward network to ∑σ
§ Image captioning translate word by word ∑σ
§ Speech to text ∑σ
§ Numerical time series data, e.g. sensor data
§ But what happens with this question?
Input x Output y
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 144 Material of KNIME AG used under CC BY 4.0 145
144 145
23
21/06/24
Input x Output y
Ich I
mag like
Schokolade chocolate
Image Source: Christopher Olah, https://colah.github.io/posts/2015-08-Understanding-LSTMs/
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 146 Material of KNIME AG used under CC BY 4.0 147
146 147
From Feed Forward to Recurrent Neural Networks From Feed Forward to Recurrent Neural Networks
𝑦z 𝑦ž 𝑦Ÿ 𝑦
𝑾𝟑𝒚 𝑾𝟑𝒚 𝑾𝟑𝒚 𝑾𝟑𝒚
𝑾𝒙𝟐 𝑾𝒚𝟑
∑ σ ∑ σ ∑ σ ∑ σ ∑ σ ∑ σ
𝑥)
𝑾𝟐𝒙 𝑾𝟑𝒚
∑ σ ∑ σ 𝑦 𝑥 ∑ σ 𝑦 ∑ σ ∑ σ ∑ σ ∑ σ
𝑥(
∑ σ ∑ σ ∑ σ ∑ σ ∑ σ ∑ σ
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 148 Material of KNIME AG used under CC BY 4.0 149
148 149
150 151
24
21/06/24
LSTM = Long Short Term Memory Unit Different Network-Structures and Applications
Many to Many
§ Special type of unit with three gates
§ Forget gate Ich gehe gerne segel
§ Input gate n
I like sailing <eos>
§ Output gate
D D D D
A A A A
E E E Ich gehe gerne
<sos> I like sailing
I like sailing
152 153
A A A A A A A A A A
I like to go sailing
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 154 Material of KNIME AG used under CC BY 4.0 155
154 155
156 157
25
21/06/24
158 159
Recommendation Engines
160 161
Recommendation Engines and Market Basket Analysis Recommendation Engines or Market Basket Analysis
From the analysis of many From the analysis of the reactions
A-priori algorithm Recommendation Recommendation
shopping baskets ... of many people to the same item ...
Collaborative Filtering
IF +
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 162 Material of KNIME AG used under CC BY 4.0 163
162 163
26
21/06/24
{A, B, F, H}
Search for {A, B, C}
IF + THEN frequent item sets
{B, C, H}
{D, E , F}
{D, E}
{A, B}
Antecedent Consequent {A, C}
{H, F}
…
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 164 Material of KNIME AG used under CC BY 4.0 165
164 165
The rules with support, confidence and lift above a threshold à most reliable ones
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 166 Material of KNIME AG used under CC BY 4.0 167
166 167
Discover all frequent and strong association rules § Let‘s consider milk, diaper, and beer: 𝑚𝑖𝑙𝑘, 𝑑𝑖𝑎𝑝𝑒𝑟 ⇒ 𝑏𝑒𝑒𝑟
XÞY à “if X then Y”
with sufficient support and confidence § How often are they found together across all shopping baskets?
§ How often are they found together across all shopping baskets containing the
antecedents?
support
Two phases: TID Transactions
𝑠 𝑚𝑖𝑙𝑘, 𝑑𝑖𝑎𝑝𝑒𝑟, 𝑏𝑒𝑒𝑟
1. find all frequent itemsets (FI) ß Most of the complexity 1 Bread, Milk 𝑃 𝑚𝑖𝑙𝑘, 𝑑𝑖𝑎𝑝𝑒𝑟, 𝑏𝑒𝑒𝑟 2
§ Select itemsets with a minimum support = = = 0.4
2 Bread, Diaper, Beer, Eggs 𝑇 5
𝐹𝐼 = 𝑋, 𝑌 , 𝑋, 𝑌 ⊂ 𝐼|𝑠 𝑋, 𝑌 ≥ 𝑆Q#$
3 Milk, Diaper, Beer, Coke
2. build strong association rules 𝑃 𝑚𝑖𝑙𝑘, 𝑑𝑖𝑎𝑝𝑒𝑟, 𝑏𝑒𝑒𝑟 2
§ Select rules with a minimum confidence: User parameters 4 Bread, Milk, Diaper, Beer 𝑐= = = 0.67
𝑅𝑢𝑙𝑒𝑠: 𝑋 ⇒ 𝑌, 𝑋, 𝑌 ⊂ 𝐹𝐼, ”𝑐 𝑋 ⇒ 𝑌 ≥ 𝐶Q#$ 𝑃 𝑚𝑖𝑙𝑘, 𝑑𝑖𝑎𝑝𝑒𝑟 3
5 Bread, Milk, Diaper, Coke
confidence
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 168 Material of KNIME AG used under CC BY 4.0 169
168 169
27
21/06/24
170 171
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 172 Material of KNIME AG used under CC BY 4.0 173
172 173
Exercises:
§ Neural Network
§ Goal: Train an MLP to solve our classification problem (rank: high/low)
§ 01_Simple_Neural_Network_exercise
175 176
28
21/06/24
177 178
C4 C5 C6
These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 180
179 180
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 181 Material of KNIME AG used under CC BY 4.0 182
181 182
29
21/06/24
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 183 Material of KNIME AG used under CC BY 4.0 184
183 184
Partitioning
Goal:
A (disjoint) partitioning into k clusters with minimal costs
185 186
9 9 9
8 8
7 7 7
6
Cluster 6
Calculation of 6
2
3
2
3
4. Go back to Step 2, repeat until the updated centroids stop moving significantly
1 1 1
0 0 0
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10
9 9
8 8
7 7
6 6
5 5
4 4
Repeat unitl 3
2
Calculation of 3
centers stop
0 0
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10
moving
Changes in
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
cluster assignment
Material of KNIME AG used under CC BY 4.0 187 Material of KNIME AG used under CC BY 4.0 188
187 188
30
21/06/24
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 189 Material of KNIME AG used under CC BY 4.0 190
189 190
§ Lp-Metric (Minkowski-Distance) R D
𝑑𝑖𝑠𝑡(𝑥, 𝑦) = i 𝑥$ − 𝑦$ \
$%)
§ Euclidean Distance (𝑝 = 2)
D
Clustering: Distance Functions 𝑑𝑖𝑠𝑡(𝑥, 𝑦) = i 𝑥$ − 𝑦$ (
$%)
§ Manhattan-Distance (𝑝 = 1)
D
𝑑𝑖𝑠𝑡 𝑥, 𝑦 = i 𝑥$ − 𝑦$
$%)
§ Maximum-Distance (𝑝 = ∞)
𝑑𝑖𝑠𝑡 𝑥, 𝑦 = max 𝑥$ − 𝑦$
)—$—D
These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 192
191 192
193 194
31
21/06/24
Between-Cluster Variation
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 195 Material of KNIME AG used under CC BY 4.0 196
195 196
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 197 Material of KNIME AG used under CC BY 4.0 198
197 198
Silhouette-Coefficient Silhouette-Coefficient
…not so good… …bad clustering.
Cluster 1
Cluster 1 Cluster 2
𝑎(𝑥)
𝑏(𝑥) Cluster 2
𝑎(𝑥)
199 200
32
21/06/24
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 201 Material of KNIME AG used under CC BY 4.0 202
201 202
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 203 Material of KNIME AG used under CC BY 4.0 204
203 204
Dendrogram:
Clustering: Linkage
§ A tree representing hierarchy of clusters,
Hierarchical Clustering with the following properties:
Distance
205 206
33
21/06/24
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 207 Material of KNIME AG used under CC BY 4.0 209
207 209
𝐷𝑖𝑠𝑡(𝐶z, 𝐶ž) = min {𝑑𝑖𝑠𝑡(𝑝, 𝑞)} 𝐷𝑖𝑠𝑡(𝐶z, 𝐶ž) = m𝑎𝑥 {𝑑𝑖𝑠𝑡(𝑝, 𝑞)}
-∈®n ,¯∈®o -∈®n ,¯∈®o
Distance of the closest two points, one from each cluster Distance of the farthest two points, one from each cluster
§ Merge Step: Union of two subsets of data points § Merge Step: Union of two subsets of data points
Cj Cj
Ci Ci
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 210 Material of KNIME AG used under CC BY 4.0 211
210 211
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 212 Material of KNIME AG used under CC BY 4.0 213
212 213
34
21/06/24
§ Single Linkage:
§ Prefers well-separated clusters
§ Complete Linkage:
§ Prefers small, compact clusters
§ Average Linkage:
§ Prefers small, well-separated clusters… Clustering: Density
DBSCAN
214 215
DBSCAN - a density-based clustering algorithm - defines five types of points in a dataset. § t = Core point
Core Point s n § s = Boarder point
§ Core Points are points that have at least a minimum number of neighbors (MinPts) vs. Border Point
within a specified distance (𝜀). § n = Noise point
vs. Noise
§ Noise Points are neither core points nor border points. t
§ Border Points are points that are within 𝜀 of a core point, but have less than MinPts
neighbors.
§ Directly Density Reachable Points are within 𝜀 of a core point.
§ Density Reachable Points are reachable with a chain of Directly Density Reachable Directly Density § z is directly density
points. Reachable s reachable from t
vs. Density z § s is not directly density
Reachable t reachable from t, but
Clusters are built by joining core and density-reachable points to one another.
density reachable via z
Note: But t is not density reachable from s, because s is not a Core point
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 216 Material of KNIME AG used under CC BY 4.0 217
216 217
DBSCAN [Density Based Spatial Clustering of Applications with Noise] DBSCAN [Density Based Spatial Clustering of Applications with Noise]
§ For each point, DBSCAN determines the e-environment and checks whether it contains § For each point, DBSCAN determines the e-environment and checks whether it contains
more than MinPts data points è core point more than MinPts data points è core point
§ Iteratively increases the cluster by adding density-reachable points § Iteratively increases the cluster by adding density-reachable points
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 218 Material of KNIME AG used under CC BY 4.0 219
218 219
35
21/06/24
Summary: DBSCAN DBSCAN [Density Based Spatial Clustering of Applications with Noise]
Clustering: § DBSCAN uses (spatial) index structures for determining the e-environment:
§ A density-based clustering 𝐶 of a dataset D w.r.t. 𝜀 and MinPts is the set of all density- à computational complexity 𝑂(𝑛 log 𝑛) instead of 𝑂(𝑛2)
based clusters 𝐶$ w.r.t. 𝜀 and MinPts in D. § Arbitrary shape clusters found by DBSCAN
§ The set 𝑁𝑜𝑖𝑠𝑒𝐶𝐿 („noise“) is defined as the set of all objects in D which do not belong to § Parameters: 𝜀 and 𝑀𝑖𝑛𝑃𝑡𝑠
any of the clusters.
Property:
§ Let 𝐶$ be a density-based cluster and 𝑝Î𝐶$ be a core object.
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 220 Material of KNIME AG used under CC BY 4.0 221
220 221
Motivation
§ Real world data is “dirty“
à Contains missing values, noises, outliers, inconsistencies
§ Comes from different information sources
à Different attribute names, values expressed differently, related tuples
§ Different value ranges and hierarchies
Data Preparation à One attribute range may overpower another
§ Huge amount of data
à Makes analyis difficult and time consuming
222 223
Data Preparation
§ Data Cleaning & Standardization (domain dependent)
§ Aggregations (often domain dependent)
§ Normalization
§ Dimensionality Reduction
§ Outlier Detection
§ Missing Value Imputation Data Preparation: Normalization
§ Feature Selection
§ Feature Engineering
§ Sampling
§ Integration of multiple Data Sources
224 225
36
21/06/24
𝑥 − 𝑚𝑒𝑎𝑛(𝑥)
Goal of normalization: 𝑦=
𝑠𝑡𝑑𝑑𝑒𝑣(𝑥)
§ Transformation of attributes to make record ranges comparable
§ normalization by decimal scaling
E
𝑦 = )& ¢ where j is the smallest integer for max(𝑦) < 1
226 227
PMML
§ Predictive Model Mark-up Language (PMML) standard XML-based interchange format
for predictive models.
§ Interchange. PMML provides a way to describe and exchange predictive models
produced by machine learning algorithms
§ Standard. In theory, a PMML model exported from KNIME can be read by PMML
compatible functions in other tools Data Preparation: Missing Value
§ It does not work that well for the modern / ensemble algorithms, such as random forest Imputation
or deep learning. In this case, other formats have been experimented.
228 229
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 230 Material of KNIME AG used under CC BY 4.0 231
230 231
37
21/06/24
232 233
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 234 Material of KNIME AG used under CC BY 4.0 235
234 235
§ Knowledge-based
§ Statistics-based
§ Distance from the median
§ Position in the distribution tails
§ Distance to the closest cluster center
§ Error produced by an autoencoder
§ Number of random splits to isolate a data
point
from other data
https://www.knime.com/blog/four-techniques-for-outlier-detection
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 236 Material of KNIME AG used under CC BY 4.0 237
236 237
38
21/06/24
238 239
240 241
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 242 Material of KNIME AG used under CC BY 4.0 243
242 243
39
21/06/24
§ The first principal component 𝑃𝐶) follows the direction § Thus, to describe the data we could use only the top 𝑚 < 𝑛 (i.e., 𝑃𝐶), 𝑃𝐶(, ⋯ 𝑃𝐶i)
(eigenvector) of the largest possible variance (largest eigenvalue Image from W ikipedia components with little - if any - loss of information
of the covariance matrix) in the data. x2
§ Each succeeding component 𝑃𝐶u follows the direction of the next PC2 § Caveats:
largest possible variance under the constraint that it is orthogonal Dimensionality Reduction
to (i.e., uncorrelated with) the preceding components PC1 § Results of PCA are quite difficult to interpret
𝑃𝐶), 𝑃𝐶(, ⋯𝑃𝐶u^) .
§ Normalization required
If you’re still curious, there’s LOTS of different ways to think about PCA: § Only effective on numeric columns
https://stats.stackexchange.com/questions/2691/making-sense-of-principal-
component-analysis-eigenvectors-eigenvalues
x1
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 244 Material of KNIME AG used under CC BY 4.0 245
244 245
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 246 Material of KNIME AG used under CC BY 4.0 247
246 247
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 248 Material of KNIME AG used under CC BY 4.0 249
248 249
40
21/06/24
250 251
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 252 Material of KNIME AG used under CC BY 4.0 253
252 253
https://thenewstack.io/3-new-techniques-for-data-dimensionality-reduction-in-machine-learning/
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 254 Material of KNIME AG used under CC BY 4.0 255
254 255
41
21/06/24
Data Preparation:
Feature Engineering
256 257
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 258 Material of KNIME AG used under CC BY 4.0 259
258 259
Exercises
§ Clustering
§ Goal: Cluster location data from California
§ 01_Clustering_exercise
§ Data Preparation
§ 02_Missing_Value_Handling_exercise
§ 03_Outlier_Detection_exercise
Session 4 exercises § 04_Dimensionality_Reduction_exercise
§ 05_Feature_Selection_exercise
260 261
42
21/06/24
262 263
Get certified at: All cheat sheets are available for download at:
knime.com/certification-program knime.com/cheat-sheets
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 264 Material of KNIME AG used under CC BY 4.0 265
264 265
All KNIME Press books are available for download at: Find the challenges at:
knime.com/knimepress knime.com/just-knime-it
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 266 Material of KNIME AG used under CC BY 4.0 267
266 267
43
21/06/24
Low Code for Data Science community journal on Medium Write a review on G2:
medium.com/low-code-for-advanced-data-science www.g2.com/products/knime-analytics-platform/take_survey
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 268 Material of KNIME AG used under CC BY 4.0 269
268 269
Follow us on
Blog: KNIME Self-Paced Courses:
social media:
knime.com/blog knime.com/knime-self-paced-
courses
Forum: Email:
forum.knime.com education@knime.com
These slides are a derivative of KNIME Course These slides are a derivative of KNIME Course
Material of KNIME AG used under CC BY 4.0 259 Material of KNIME AG used under CC BY 4.0
270 271
44