๐—ฅ๐—ฒ๐—ด๐˜‚๐—น๐—ฎ๐—ฟ๐—ถ๐˜‡๐—ฎ๐˜๐—ถ๐—ผ๐—ป in ๐——๐—ฒ๐—ฒ๐—ฝ ๐—Ÿ๐—ฒ๐—ฎ๐—ฟ๐—ป๐—ถ๐—ป๐—ด: Intuition behind.

Facebook
Twitter
LinkedIn

Table of Contents

๐Ÿ’ก๐—ฅ๐—ฒ๐—ด๐˜‚๐—น๐—ฎ๐—ฟ๐—ถ๐˜‡๐—ฎ๐˜๐—ถ๐—ผ๐—ป in ๐——๐—ฒ๐—ฒ๐—ฝ ๐—Ÿ๐—ฒ๐—ฎ๐—ฟ๐—ป๐—ถ๐—ป๐—ด: Intuition behind.

Regularization 101 definition :ย Model does well on training data and not so well on unseen data. Overfitting ๐Ÿ™‚

But is there more to that, Letโ€™s figure out.

Remember that one guy in school who memorized everything what is mentioned in books or uttered from teacherโ€™s mouth.ย But didnโ€™t perform well when questions were twisted a bit.

What happened?

He only memorised the lessons but didnโ€™t understand the concept behind it to apply on previously unseen questions.

Thatโ€™s ๐—ผ๐˜ƒ๐—ฒ๐—ฟ๐—ณ๐—ถ๐˜๐˜๐—ถ๐—ป๐—ด and to correct that we need regularization.

Regularization acts as that good teacher, guiding the student to focus on core concepts rather than memorizing irrelevant details.

Regularization essentially solves 3 problems.

1๏ธโƒฃ Overfitting: Prevents the model from fitting the noise or irrelevant details in the training data.

2๏ธโƒฃ Model Complexity: Reduces the complexity of the model by constraining its capacity, ensuring it doesnโ€™t overlearn.

3๏ธโƒฃ Bias-Variance Tradeoff: Strikes a balance between underfitting (too simple) and overfitting (too complex).

So, how do we do regularization ?

Quite a few ways, actually.

Letโ€™s see the most important ones. And letโ€™s try to understand it without getting any maths involved. Shall we?

1๏ธโƒฃL1 and L2 Regularizationย – Way to discourage large weights.ย A penalty term ensures that large weights are dampened.ย Penalty addedย to abosolute weight (L1)ย squared weights (L2)

2๏ธโƒฃDropout – Randomly “drops out” (sets to zero) a fraction of neurons during training. This forces the network to not overly rely on specific neurons and promotes generalization.

3๏ธโƒฃ Data Augmentation – Why not give different variants of question to that friend so that he becomes really good at grasping concepts.

4๏ธโƒฃ Early stopping – Before it starts memorising, stop your training.

5๏ธโƒฃ Batch Norm – Normalise ( mean centre to 0 and variance 1) . Ensure, all neurons gets a fair chance in next layer.

6๏ธโƒฃ Elastic Net – Combination of L1 and L2.

Facebook
Twitter
LinkedIn

Similar Posts

Contact Us

We would be delighted to help !

Contact Us

Call Now Button