We can not continue treating our models as black boxes anymore. Remember, nobody trusts computers for making a very important decision (yet!). That's why the interpretation of Machine Learning models has become a major research topic. SHAP is a very robust approach for providing interpretability to any machine learning model. For multi-classification problems, however, documentation and examples are not very clear. [8min read]