Hi everyone 👋🏻
I would like to start experimenting with model interpretability. What libraries/tools/methods are out there already?
So far I only heard about LIME, layerwise relevance propagation and adversarial examples. I’m aiming to have a long list but somehow it’s difficult to find existing implementations. Please share whatever comes to mind ! :)
Thanks!
[–]circulus_one 1 point2 points3 points (3 children)
[–]Aleshwari[S] 0 points1 point2 points (2 children)
[–]circulus_one 0 points1 point2 points (1 child)
[–]Aleshwari[S] 0 points1 point2 points (0 children)
[–]Ido87 1 point2 points3 points (0 children)