Beautiful decision tree visualizations with dtreeviz
Improve the old way of plotting the decision trees and never go back!
By Eryk Lewinson, Senior Data Scientist
Decision trees are a very important class of machine learning models and they are also building blocks of many more advanced algorithms, such as Random Forest or the famous XGBoost. The trees are also a good starting point for a baseline model, which we subsequently try to improve upon with more complex algorithms.
One of the biggest advantages of the decision trees is their interpretability — after fitting the model, it is effectively a set of rules that can be used to predict the target variable. That is also why it is easy to plot the rules and show them to stakeholders, so they can easily understand the model’s underlying logic. Of course, provided that the tree is not too deep.
Visualizing the decision trees can be really simple using a combination of
matplotlib. However, there is a nice library called
dtreeviz, which brings much more to the table and creates visualizations that are not only prettier but also convey more information about the decision process.
In this article, I will first show the “old way” of plotting the decision trees and then introduce the improved approach using
As always, we need to start by importing the required libraries.
Then, we load the Iris data set from
scikit-learn. We will also go over a regression example, but we will load the Boston housing data set for this later on.
The “old way”
The next step involves creating the training/test sets and fitting the decision tree classifier to the Iris data set. In this article, we focus purely on visualizing the decision trees. Thus, we do not pay any attention to fitting the model or finding a good set of hyperparameters (there are a lot of articles on these topics). The only thing that we will “tune” is the maximum depth of the tree — we constraint it to 3, so the trees can still fit in the image and remain readable.
Now that we have a fitted decision tree model and we can proceed to visualize the tree. We start with the easiest approach — using the
plot_tree function from
Image by author
OK, not bad for a one-liner. But it is not very readable, for example, there are no feature names (only their column indices) or class labels. We can easily improve that by running the following snippet.
Image by author
Much better! Now, we can quite easily interpret the decision tree. It is also possible to use the
graphviz library for visualizing the decision trees, however, the outcome is very similar, with the same set of elements as the graph above. That is why we will skip it here, but you can find the implementation in the Notebook on GitHub.
dtreeviz in action
Having seen the old way of plotting the decision trees, let’s jump right into the
The code snippet is pretty much self-explanatory, so we can move on to the outcome. First of all, let’s take a moment to acknowledge how big of an improvement it is, especially given that the function call is very similar.
Image by author
Let’s go over the plot step by step. At each node, we can see a stacked histogram of the feature that is used for splitting the observations, colored by class. This way, we can see how the classes are segregated by each split. The small triangle with the value at the x-axis is the splitting point. In the first histogram, we can clearly see that all observations of the setosa class have petal length smaller than 2.45 cm.
The right branches of the tree indicate selecting the values larger or equal to the splitting value, while the left one lesser than the splitting value. The leaf nodes are represented as pie charts, which show what fraction of the observations within the leaf belongs to which class. This way, we easily see which class is the majority one, so also the model’s prediction.
One thing we do not see at this plot is the value of the Gini coefficient at each node. In my opinion, the histogram provides more intuition about the split and the value of the coefficient might not be that relevant in case of a presentation for the stakeholders either way.
Note: We can also create a similar visualization for the test set, we just need to replace the
y_data arguments while calling the function.
If you are not a fan of the histograms and want to simplify the plot, you can specify
fancy=False to receive the following simplified plot.
Image by author
Another handy feature of
dtreeviz which improves the model’s interpretability is path highlighting of a particular observation on the plot. This way, we clearly see which features contributed to the class prediction.
Using the snippet below, we highlight the path of the first observation of the test set.
The plot is very similar to the previous one, however, the orange highlight clearly shows which path the observation followed. Additionally, we can see the orange triangle at each of the histograms. It represents the indicated observation’s value of the given feature. At the very end, we see the values of all the features of this observation, with the ones used for decision highlighted in orange. In this case, only two features were used for predicting that the observation belongs to the versicolor class.
Image by author
Tip: We can also change the orientation of the plots from top-to-bottom to left-to-right by setting
orientation=”LR”. We do not show it in this article, as the charts will not be scaled that nicely for devices with a narrower screen.
Lastly, we can print the decisions used for this observation’s prediction in plain English. To do so, we run the following command.
This way, we can clearly see the conditions that this observation fulfills.
We have already covered a classification example, which showed most of the interesting functionalities of the library. But for completeness’ sake, we also go over an example of a regression problem, to show how the plots differ. We use another popular data set — the Boston housing one. It’s a problem in which we use a set of different areas to predict the median housing price within certain areas of Boston.
The code already feels similar. The only change is that we added
show_node_labels = True. It can be especially handy for larger decision trees. So while discussing the plot with a group, it is very easy to indicate which split we are discussing by the node’s number.
Image by author
Let’s dive into the differences between classification trees and regression ones. This time, instead of looking at histograms, we inspect scatterplots of the feature used for the split vs. the target. On those scatterplots, we see some dashed lines. Their interpretation is as follows:
- horizontal lines are the target’s mean for the left and right buckets in decision nodes.
- vertical lines are the split point. It is the very same information as represented by the black triangle, however, it makes comparing the horizontal lines easier -> easy separation between the sides.
In the leaf nodes, the dashed line shows the mean of the target within the leaf, which is also the model’s prediction.
We already showed that we can highlight the decision path for a certain observation. We can take it a step further and only plot the nodes used for that prediction. We do so by specifying
show_just_path=True. The following plot shows only the selected nodes from the tree above.
Image by author
In this article, I showed how to use the
dtreeviz library for creating elegant and insightful visualizations of decision trees. Having played around with it for a bit, I will definitely keep on using it as the go-to tool for visualizing decision trees. I do believe that the plots created using this library are much easier to understand for people who do not work with ML on a daily basis and can help in conveying the model’s logic to the stakeholders.
It is also worth mentioning that
dtreeviz supports some visualizations for XGBoost and Spark MLlib trees.
If you liked this article, you might also be interested in one of the following:
Improve the train-test split with the hashing function
The best way to make sure the training and test sets are never mixed while updating the data set
Lazy Predict: fit and evaluate all the models from scikit-learn with a single line of code
The easiest way to see which models work best for your dataset!
Explaining Feature Importance by example of a Random Forest
Learn the most popular methods of determining feature importance in Python
Bio: Eryk Lewinson is currently working as a senior data scientist in the Netherlands' biggest e-commerce platform. In his professional career, he gained experience in the practical application of data science methods while working for two Big 4 companies and a Dutch FinTech scale-up. In his work, he focuses on using machine learning to provide business value to his company.
Original. Reposted with permission.
- All Machine Learning Algorithms You Should Know in 2021
- 11 Essential Code Blocks for Complete EDA (Exploratory Data Analysis)
- Telling a Great Data Story: A Visualization Decision Tree
Top Stories Past 30 Days