You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I found a strange phenomenon. For the same model, the same training sample and test sample, other operations are identical. Theoretically, the values obtained by using the XAI method (like Saliency) to evaluate the interpretability of the model should be the same. However, I retrained a new model, and the interpretability values obtained are completely different from those obtained from the previous model. Does anyone know why this happens? The interpretability value is completely unstable, and the results cannot be reproduced. Unless I completely save this model after training it, and then reload this parameter, the results will be the same. Does anyone know why
The text was updated successfully, but these errors were encountered:
This is a common phenomenon when considering perturbation based XAI methods. In order to estimate the marginal contribution of your input features (pixels in the case of saliency maps), XAI algorithms often perform some sort of sampling in the input space, which can be non-deterministic (especially if a random sampling is done). This is also why explanations obtained using this kind of method needs to be manipulated carefully.
What they are really doing is creating a local simple surrogate (linear classifier, small decision tree) to explain one sample, and in order to make sense, this makes the assumption that the decision boundary will be approximated by a simple model.
I found a strange phenomenon. For the same model, the same training sample and test sample, other operations are identical. Theoretically, the values obtained by using the XAI method (like Saliency) to evaluate the interpretability of the model should be the same. However, I retrained a new model, and the interpretability values obtained are completely different from those obtained from the previous model. Does anyone know why this happens? The interpretability value is completely unstable, and the results cannot be reproduced. Unless I completely save this model after training it, and then reload this parameter, the results will be the same. Does anyone know why
The text was updated successfully, but these errors were encountered: