In 1912, the RMS Titanic hit an iceberg in the North Atlantic Ocean about 400 miles south of Newfoundland, Canada and sank. Unfortunately, there were not enough lifeboats onboard to accommodate all passengers and 67% of the passengers died. In this article, we walk through the use of SHAP values to explain, in a detailed manner, why an AI model decides to predict whether a given passenger will or will not survive.
Inside the Black Box: Developing Explainable AI Models Using SHAP
Explainable AI refers to the ability to interpret model outcomes in a way that is easily understood by human beings. We explore why this matters, and discuss in detail tools that help shine light inside the AI "black box" -- we wish to not just understand feature importance at the population level, but to actually quantify feature importance on a per-outcome basis.