Publications

What could go wrong? Discovering and describing failure modes in computer vision

VTCD: Understanding Video Transformers via Universal Concept Discovery

Visual Concept Connectome (VCC): Open World Concept Discovery and their Interlayer Connections in Deep Models

Top-GAP: Integrating Size Priors in CNNs for more Interpretability, Robustness, and Bias Mitigation

Q-SENN: Quantized Self-Explaining Neural Networks

Pruning By Explaining Revisited: Optimizing Attribution Methods to Prune CNNs and Transformers

ProactiV: Studying Deep Learning Model Behavior under Input Transformations

PDiscoFormer: Relaxing Part Discovery Constraints with Vision Transformers

Network Inversion of Binarised Neural Nets

Localization-Guided Supervision for Robust Medical Image Classification by Vision Transformers

Interpretability benchmark for evaluating spatial misalignment of prototypical parts explanations

Integrating Local and Global Interpretability for Deep Concept-Based Reasoning Models

Image-guided topic modeling for interpretable privacy classification

Global Counterfactual Directions

From Flexibility to Manipulation: The Slippery Slope of XAI Evaluation

Feature Contribution in Monocular Depth Estimation

Explanation Alignment: Quantifying the Correctness of Model Reasoning At Scale

Concept-Based Explanations in Computer Vision: Where Are We and Where Could We Go?

Can Biases in ImageNet Models Explain Generalization?

Attribute Based Interpretable Evaluation Metrics for Generative Models

As large as it gets: Learning infinitely large Filters via Neural Implicit Functions in the Fourier Domain

Analyzing Vision Transformers for Image Classification in Class Embedding Space

An Investigation on The Position Encoding in Vision-Based Dynamics Prediction

"Help Me Help the AI": Understanding How Explainability Can Support Human-AI Interaction