TAB: Transformer Attention Bottlenecks enable User Intervention and Debugging in Vision-Language Models
Published in IEEE/CVF International Conference on Computer Vision (ICCV), 2025
Recommended citation: Rahmanzadehgervi P, Nguyen HH, Liu R, Mai L, Nguyen AT. TAB: Transformer Attention Bottlenecks enable User Intervention and Debugging in Vision-Language Models. In Proceedings of the IEEE/CVF International Conference on Computer Vision 2025 (pp. 22551-22562). https://arxiv.org/pdf/2412.18675
