Your Features Are Important? It Doesn’t Mean They Are Good
<p>The concept of “feature importance” is widely used in machine learning as the most basic type of model explainability. For example, it is used in Recursive Feature Elimination (RFE), to iteratively drop the least important feature of the model.</p>
<p>However, there is a misconception about it.</p>
<blockquote>
<p><strong>The fact that a feature is important doesn’t imply that it is beneficial for the model!</strong></p>
</blockquote>
<p>Indeed, when we say that a feature is important, this simply means that the feature brings a high contribution to the predictions made by the model. But we should consider that <strong>such contribution may be wrong</strong>.</p>
<p>Take a simple example: a data scientist accidentally forgets the Customer ID between its model’s features. The model uses Customer ID as a highly predictive feature. As a consequence, this feature will have a high feature importance even if it is actually worsening the model, because it cannot work well on unseen data.</p>
<p>To make things clearer, we will need to make a distinction between two concepts:</p>
<ul>
<li><strong>Prediction Contribution</strong>: what part of the predictions is due to the feature; this is equivalent to feature importance.</li>
<li><strong>Error Contribution</strong>: what part of the prediction errors is due to the presence of the feature in the model.</li>
</ul>
<p>In this article, we will see how to calculate these quantities and how to use them to get valuable insights about a predictive model (and to improve it).</p>
<p><a href="https://towardsdatascience.com/your-features-are-important-it-doesnt-mean-they-are-good-ff468ae2e3d4">Website</a></p>