Tech Corner

Ultimate Cheatsheet to tune your Hyper-Parameters in ML Algorithms

Engati Team
last edited on
November 20, 2023
6-7 mins

Table of contents

Automate your business at $5/day with Engati

Switch to Engati: Smarter choice for WhatsApp Campaigns 🚀
Machine learning algorithm

It is always said that hyper-parameter Tuning is an iterative process and there is no shortcut to arrive at the best hyper-parameter in machine learning that your model requires. Having said that, there are certain factors which give you direction in which you should start tuning your hyper-parameters in machine learning and neural network algorithms.

Machine Learning & Neural Network Algorithms


Variance Bias Tradeoff

Generally, the accuracy metric that is monitored for ML and DL algorithms are Root Mean Square Error (RMSE) or Mean Absolute Error (MAE). Hyper-parameter tuning is done to reduce the magnitude of these errors. So, how do we understand that the error is due to variance or bias? The image below describes on how you can infer whether your model has a high bias or a high variance -

variance bias error
Variance-Bias Error

Low variance (high bias) algorithms tend to be less complex with simple or rigid underlying structure.

  • They train models that are consistent but inaccurate on average.
  • These include linear or parametric algorithms such as regression and naive Bayes.

On the other hand, low bias (high variance) algorithms tend to be more complex with a flexible underlying structure.

  • They train models that are accurate on average but inconsistent.
  • These include non-linear or non-parametric algorithms such as decision trees and nearest neighbours.

variance bias trade off
Variance Bias Trade Off

So, to summarise the Variance-Bias tradeoff, we are often in one of the following two situations when tuning our hyper-parameters i.e High Variance or High Bias.

machine learning and artificial intelligence

How to overcome High Bias error?

  • Make the model more complex. (This can be done in different ways which I will explain in detail later in the blog)
  • Train longer (learning rate)

How to overcome High Variance error?

  • Train the model with more data.
  • Make sure that the samples in your training set are not much different than your samples in the test seto Regularise the model.

Boosting and Bagging

Boosting is based on weak learners (high bias, low variance). In terms of decision trees, weak learners are shallow trees, sometimes even as small as decision stumps (trees with two leaves). Boosting reduces error, mainly by reducing bias (and also to some extent variance, by aggregating the output from many models).

On the other hand, Random Forest uses fully grown decision trees (low bias, high variance). It tackles the error reduction task in the opposite way: by reducing variance. The trees are made uncorrelated to maximise the decrease in variance, but the algorithm cannot reduce bias (which is slightly higher than the bias of an individual tree in the forest). Hence, the need for large, unpruned trees, so that the bias is initially as low as possible.



Machine Learning

Depth of the tree for tree based algorithms

One straight-forward way is to limit the maximum allowable tree depth. The common way for tree based algorithms to overfit is when they get too deep. Thus, you can use the maximum depth parameter as the regularisation parameter — making it smaller will reduce the overfitting and introduce bias, increasing it will do the opposite.

Larger the depth, more complex the model; higher chances of overfitting. There is no standard value for max_depth. Larger data sets require deep trees to learn the rules from data.

The depth of the tree should be tuned using cross validation.

  • It controls regularization (or prevents overfitting). The optimal value of gamma depends on the data set and other parameter values.
  • Higher the value, higher the regularization. Regularization means penalizing large coefficients which don’t improve the model’s performance. Default in XGB= 0 means no regularization.
  • Tune trick: Start with 0 and check CV error rate. If you see train error >>> test error, bring gamma into action. Higher the gamma, lower the difference in train and test CV. If you have no clue what value to use, use gamma=5 and see the performance. Remember that gamma brings improvement when you want to use shallow (low max_depth) trees.

  • Default in XGB=0
  • It controls L2 regularization (equivalent to Ridge regression) on weights. It is used to avoid overfitting.

  • Default in XGB=1
  • It controls L1 regularization (equivalent to Lasso regression) on weights. In addition to shrinkage, enabling alpha also results in feature selection. Hence, it’s more useful on high dimensional data sets.

Cross Validation

Cross Validation is a technique which involves reserving a particular sample of a data set on which you do not train the model. Later, you test the model on this sample before finalizing the model.

One widely used cross validation technique is k- fold cross validation. Here are the quick steps:

  • Randomly split your entire dataset into k”folds”.
  • For each k folds in your dataset, build your model on k — 1 folds of the data set. Then, test the model to check the effectiveness for kth fold.
  • Record the error you see on each of the predictions.
  • Repeat this until each of the k folds has served as the test set.

The average of your k recorded errors is called the cross-validation errorand will serve as your performance metric for the model.

Below is the visualization of how a k-fold validation works for k=10.

Always remember, lower value of K is more biased and hence, undesirable. On the other hand, a higher value of K is less biased, but can suffer from large variability. It is good to know that, smaller value of k always takes us towards validation set approach, where as higher value of k leads to LOOCV approach.

Hence, it is often suggested to use k=10.

k-fold cross validation that shows a visual representation
k-fold cross validation: A visual representation

Neural Networks

Both MLPRegressor and MLPClassifier use parameter alpha for regularization (L2 regularization) term which helps in avoiding overfitting by penalizing weights with large magnitudes.

Alpha is a parameter for regularization term, aka penalty term, that combats overfitting by constraining the size of the weights. Increasing alpha may fix high variance (a sign of overfitting) by encouraging smaller weights, resulting in a decision boundary plot that appears with lesser curvatures. Similarly, decreasing alpha may fix high bias (a sign of under-fitting) by encouraging larger weights, potentially resulting in a more complicated decision boundary.

a comparison of different values for regularisation parameter 'alpha' on synthetic datasets
A comparison of different values for regularization parameter ‘alpha’ on synthetic datasets

When you are tuning a neural network, based on whether your initial model results indicate a high variance or a high bias, the alpha value can be increased or decreased accordingly.


Model Complexity and Machine Learning

Bias is the difference between your model’s expected predictions and the true values.

That might sound strange because shouldn’t you “expect” your predictions to be close to the true values? Well, it’s not always that easy because some algorithms are simply too rigid to learn complex signals from the dataset.

Imagine fitting a linear regression to a dataset that has a non-linear pattern:

low complexity model in machine learning
Low Complexity Model

No matter how many more observations you collect, a linear regression won’t be able to model the curves in that data! This is known as under-fitting.

Variance refers to your algorithm’s sensitivity to specific sets of training data.

High variance algorithms will produce drastically different models depending on the training set.

For example, imagine an algorithm that fits a completely unconstrained, super-flexible model to the same dataset from above:

high complexity model in machine learning
High Complexity Model

As you can see, this unconstrained model has basically memorized the training set, including all of the noise. This is known as over-fitting.

There are some ways you can hypertune the complexity of the algorithm that you are working on:

K-Nearest Neighbors

Increasing “k” will decrease variance and increase bias.

Decreasing “k” will increase variance and decrease bias.


Increasing the degree of the polynomial would make it more complex.

Decreasing the degree of the polynomial would decrease the complexity of the model.


sklearn.neural_network.MLPRegressor - scikit-learn 0.19.1 documentation
class sklearn.neural_network. MLPRegressor( hidden_layer_sizes=(100, ), activation='relu', solver='adam', alpha=0.0001…

Read our blog on AI & Conservation of Energy

Blog Cover Photo by Franck V. on Unsplash

Engati Team

At the forefront for digital customer experience, Engati helps you reimagine the customer journey through engagement-first solutions, spanning automation and live chat.

Close Icon
Request a Demo!
Get started on Engati with the help of a personalised demo.
This is some text inside of a div block.
This is some text inside of a div block.
This is some text inside of a div block.
This is some text inside of a div block.
*only for sharing demo link on WhatsApp
Thanks for the information.
We will be shortly getting in touch with you.
Oops! something went wrong!
For any query reach out to us on
Close Icon
Congratulations! Your demo is recorded.

Select an option on how Engati can help you.

I am looking for a conversational AI engagement solution for the web and other channels.

I would like for a conversational AI engagement solution for WhatsApp as the primary channel

I am an e-commerce store with Shopify. I am looking for a conversational AI engagement solution for my business

I am looking to partner with Engati to build conversational AI solutions for other businesses

Close Icon
You're a step away from building your Al chatbot

How many customers do you expect to engage in a month?

Less Than 2000


More than 5000

Close Icon
Thanks for the information.

We will be shortly getting in touch with you.

Close Icon

Contact Us

Please fill in your details and we will contact you shortly.

This is some text inside of a div block.
This is some text inside of a div block.
This is some text inside of a div block.
This is some text inside of a div block.
This is some text inside of a div block.
Thanks for the information.
We will be shortly getting in touch with you.
Oops! Looks like there is a problem.
Never mind, drop us a mail at