Contributed Talk Sessions | Poster Sessions | All Posters | Search Papers
Poster Session B: Wednesday, August 13, 1:00 – 4:00 pm, de Brug & E‑Hall
Do Vision Transformers See Like Humans? Evaluating their Perceptual Alignment
Pablo Hernández-Cámara1, Jose Manuel Jaén-Lorites2, Jorge Vila-Tomás1, Valero Laparra3, Jesus Malo4; 1Universidad de Valencia, 2Universidad Politécnica de Valencia, 3Universitat de València, 4Universitat de Valencia
Presenter: Jose Manuel Jaén-Lorites
Vision Transformers (ViTs) achieve remarkable performance in image recognition tasks, yet their alignment with human perception remains largely unexplored. This study systematically analyzes how model size, dataset size, data augmentation and regularization impact ViT perceptual alignment with human judgments on the TID2013 dataset. Our findings confirm that larger models exhibit lower perceptual alignment, consistent with previous works. Increasing dataset diversity has a minimal impact, but exposing models to the same images more times reduces alignment. Stronger data augmentation and regularization further decrease alignment, especially in models exposed to repeated training cycles. These results highlight a trade-off between model complexity, training strategies, and alignment with human perception, raising important considerations for applications requiring human-like visual understanding.
Topic Area: Visual Processing & Computational Vision
Extended Abstract: Full Text PDF