Does "feature importance" depend on the model type?
I was working on a small classification problem (breast cancer data set from sklearn), and trying to decide which features were most important to predict the labels. I understand that there are several ways to define important feature here (permutation importance, importance in trees ...), but I did the following: 1) rank the features by coefficient value in a logistic regression ; 2) rank the features by feature importance from a random forest. These don't quite tell the same story, and I'm thinking that a feature that might be unimportant in a linear model could be very discriminative in a non-linear model that can understand it.
Is that true in general? Or should important features (those that contribute most to a classification score) be the same across all types of models?
Topic predictor-importance feature-engineering feature-selection
Category Data Science