Submitted by Longjumping_Essay498 t3_1003d7w in MachineLearning
Can we use attention weights from causal models, as explanations or causal attributes for next word predictions?
Submitted by Longjumping_Essay498 t3_1003d7w in MachineLearning
Can we use attention weights from causal models, as explanations or causal attributes for next word predictions?
currentscurrents t1_j2f996k wrote
Attention maps can be a type of explanation.
It tells you what the model was looking at when it generated a word or identified an image, but it doesn't tell you why it looked at those bits or why it made the decision it did. You can get some useful information by looking at it, but not everything you need to explain the model.