# MLJAR's Blog

• ### How to visualize a single Decision Tree from the Random Forest in Scikit-Learn (Python)?

June 29, 2020 by Piotr Płoński

The Random Forest is an esemble of Decision Trees. A single Decision Tree can be easily visualized in several different ways. In this post I will show you, how to visualize a Decision Tree from the Random Forest.

• ### Random Forest Feature Importance Computed in 3 Ways with Python

June 29, 2020 by Piotr Płoński

The feature importance (variable importance) describes which features are relevant. It can help with better understanding of the solved problem and sometimes lead to model improvements by employing the feature selection. In this post, I will present 3 ways (with code examples) how to compute feature importance for the Random Forest algorithm from `scikit-learn` package (in Python).

• ### How to save and load Random Forest from Scikit-Learn in Python?

June 24, 2020 by Piotr Płoński

In this post I will show you how to save and load Random Forest model trained with scikit-learn in Python. The method presented here can be applied to any algorithm from sckit-learn (this is amazing about scikit-learn!).

• ### How to reduce memory used by Random Forest from Scikit-Learn in Python?

June 24, 2020 by Piotr Płoński

The Random Forest algorithm from scikit-learn package can sometimes consume too much memory:

• ### Visualize a Decision Tree in 4 Ways with Scikit-Learn and Python

June 22, 2020 by Piotr Płoński

A Decision Tree is a supervised algorithm used in machine learning. It is using a binary tree graph (each node has two children) to assign for each data sample a target value. The target values are presented in the tree leaves. To reach to the leaf, the sample is propagated through nodes, starting at the root node. In each node a decision is made, to which descendant node it should go. A decision is made based on the selected sample’s feature. Decision Tree learning is a process of finding the optimal rules in each internal tree node according to the selected metric.

• ### Compare MLJAR with Google AutoML Tables

May 17, 2019 by Piotr Płoński

Recently, Google has released AutoML service for structured datasets. It is called AutoML Tables and is currently available in Beta. I’ve decided to compare my open-source solution with Google AutoML Tables.

• ### AutoML software and services

May 14, 2019 by Piotr Płoński

Automated Machine Learning is the end-to-end process of applying machine learning in an automatic way.

• ### Random Forest vs Neural Network (classification, tabular data)

May 10, 2019 by Piotr Płoński

Which is better: Random Forest or Neural Network? This is a common question, with a very easy answer: it depends :) I will try to show you when it is good to use Random Forest and when to use Neural Network.

• ### Random Forest vs AutoML (with python code)

May 07, 2019 by Piotr Płoński

Random Forest versus AutoML you say. Hmmm…, it’s obvious that the performance of AutoML will be better. You will check many models and then ensemble them. This is true, but I would like to show you other advantages of AutoML, that will help you deal with dirty, real-life data.

• ### Does Random Forest overfit?

April 05, 2019 by Piotr Płoński

When I first saw this question I was a little surprised. The first thought is, of course, they do! Any complex machine learning algorithm can overfit. I’ve trained hundreds of Random Forest (RF) models and many times observed they overfit. The second thought, wait, why people are asking such a question? Let’s dig more and do some research. After quick googling, I’ve found the following paragraph on Leo Breiman (the creator of the Random Forest algorithm) website: