- CFA Exams
- 2023 Level II
- Topic 1. Quantitative Methods
- Learning Module 6. Machine Learning
- Subject 2. Overview of Evaluating ML Algorithm Performance

### Why should I choose AnalystNotes?

AnalystNotes specializes in helping candidates pass. Period.

##### Subject 2. Overview of Evaluating ML Algorithm Performance PDF Download

**Overfitting**occurs when a model tries to fit the training data so closely that it does not generalize well to new data.

**Generalization**refers to the model's ability to adapt properly to new, previously unseen data, drawn from the same distribution as the one used to create the model.

There are three types of samples:

**training sample**: The subset of the data set used to train a model.**validation set**: A subset of the data set - disjunct from the training set - that you use to validate the model and adjust hyperparameters.**test set**: The subset of the data set that you use to test your model after the model has gone through initial vetting by the validation set.

- It is when you create a model which is predicting the noise in the data rather than the real signal.
- Overfitting generally occurs when a model is excessively complex, such as having too many parameters relative to the number of observations.

#### Errors and Overfitting

In machine learning, the bias-variance tradeoff (or dilemma) is the problem of simultaneously minimizing two sources of error that prevent supervised learning algorithms from generalizing beyond their training set:

- The
**bias**is error from erroneous assumptions in the learning algorithm. High bias can cause an algorithm to miss the relevant relations between features and target outputs (underfitting). - The
**variance**is error from sensitivity to small fluctuations in the training set. High variance can cause overfitting: modeling the random noise in the training data, rather than the intended outputs.

The bias-variance decomposition is a way of analyzing a learning algorithm's expected generalization error with respect to a particular problem as a sum of three terms, the bias, variance, and a quantity called the

**base error**(irreducible), resulting from noise in the problem itself.

_{out}= bias error + variance error + base error

#### Preventing Overfitting

There are two common principles:- Complexity reduction: simple algorithms, limit the number of features, etc to keep things simple.
- Avoid sampling bias: K-fold-cross-validation technique can be used to mitigate the holdout sample problem (excessive reduction of the training set size).

**Learning Outcome Statements**

CFA® 2023 Level II Curriculum, Volume 1, Module 6

###
**User Contributed Comments**
0

You need to log in first to add your comment.

I was very pleased with your notes and question bank. I especially like the mock exams because it helped to pull everything together.