0% found this document useful (0 votes)
6 views5 pages

aml1 (1)

Download as pdf or txt
Download as pdf or txt
Download as pdf or txt
You are on page 1/ 5

Reading Material for AML

1. Bias in a Machine Learning Model


Definition: Bias refers to the error introduced when a model makes overly
simplistic assumptions about the data. It prevents the model from capturing
the true relationships within the data.
Impact: High bias often leads to underfitting, where the model performs
poorly on both training and test data.
Example: Using a straight line to fit a complex curve.
Key Point: Bias measures how far off the model’s predictions are from the
actual data.

2. High Variance in a Model


Definition: Variance measures the sensitivity of a model to changes in the
training data. High variance often leads to overfitting, where the model
performs well on training data but poorly on new, unseen data.
Impact: Models with high variance are too complex and memorize training
data instead of generalizing.
Key Point: High variance indicates overfitting.

3. High Bias and Low Variance


Scenario: When a model is too simple (high bias) and has low variance, it
fails to learn the underlying patterns in the data.
Outcome: This situation is referred to as underfitting.
Key Point: Balance between bias and variance is necessary for good model
performance.

4. Bias-Variance Tradeoff
Goal: Achieve a balance between bias and variance to minimize total error.
Key Concept: A good model has neither too much bias (underfitting) nor
too much variance (overfitting).

1
5. Reducing Variance
Methods:
• Simplifying the model (e.g., reducing complexity).
• Collecting more training data.
• Using techniques like regularization.
Key Point: Increasing the dataset size can help reduce variance.

6. Simple Linear Regression Formula


Formula: y = mx + c
• y: Target variable.
• x: Predictor variable.
• m: Slope of the line.
• c: Intercept.
Purpose: Predicts the relationship between one independent variable and a
dependent variable.

7. R-Squared Value in Linear Regression


Definition: The R-squared value represents the proportion of variance in
the dependent variable that is explained by the independent variables.
Range: Between 0 and 1. A higher R-squared value indicates a better fit.
Key Point: Measures how well the model fits the data.

8. Assumptions in Linear Regression


Linear regression requires:
• A linear relationship between variables.
• Independence of residuals.
• Homoscedasticity (constant variance of residuals).
Not Required: Predictors don’t need to follow a normal distribution.

2
9. Multicollinearity in Linear Regression
Definition: Multicollinearity occurs when independent variables are highly
correlated with each other.
Impact: It can make the coefficient estimates unstable and unreliable.
Key Point: Address multicollinearity using techniques like ridge regression.

10. Evaluating Regression Accuracy


Metric: Mean Squared Error (MSE).

• Measures the average squared difference between actual and predicted


values.

• Lower MSE indicates better model performance.

11. Ridge Regression


Penalty Term: Adds λ w2 to the loss function, where w are the model’s
P
coefficients.
Use Case: Addresses multicollinearity by shrinking coefficients.
Regularization Type: L2 regularization.

12. Lasso Regression


P
Penalty Term: Adds λ |w| to the loss function.
Key Feature: Performs feature selection by shrinking some coefficients to
zero.
Use Case: Preferred when feature selection is required.

13. Elastic Net


Definition: Combines L1 (lasso) and L2 (ridge) regularization, balancing
feature selection and regularization.
Advantage: Effective when features are highly correlated.

3
14. Principal Component Analysis (PCA)
Purpose: Reduces the number of features (dimensionality reduction) while
retaining the most variance in the data.
Steps:
• Scale the features (essential).

• Calculate the principal components.

• Retain components based on explained variance.

15. Naive Bayes


Based On: Bayes’ theorem.
Assumption: Features are conditionally independent.
Variants:
• Gaussian Naive Bayes: Handles continuous data with normal dis-
tribution.

• Multinomial Naive Bayes: Used for text classification.

16. Decision Tree


Leaf Node: A terminal node where no further splits occur; represents a
prediction.
Gini Impurity: Measures the homogeneity of a dataset. Lower values
indicate purer nodes.
Pruning: Reduces tree depth to prevent overfitting and reduce variance.

17. Support Vector Machines (SVM)


Parameters:
• C: Controls the tradeoff between bias and variance.

• Gamma: Determines the influence of individual data points.


Kernels:
• Linear Kernel: Suitable for linearly separable data.

4
• RBF Kernel: Effective for non-linear data by mapping to higher di-
mensions.

18. Regularization
Purpose: Reduces overfitting by adding a penalty term to the model’s loss
function.
Types:

• L1 (Lasso): Shrinks some coefficients to zero (feature selection).

• L2 (Ridge): Shrinks coefficients without setting them to zero.

You might also like