Evaluating ventral visual stream contributions to human visual robustness with deep learning approaches
Shao, Zhenan
This item is only available for download by members of the University of Illinois community. Students, faculty, and staff at the U of I may log in with your NetID and password to view the item. If you are trying to access an Illinois-restricted dissertation or thesis, you can request a copy through your library's Inter-Library Loan office or purchase a copy directly from ProQuest.
Permalink
https://hdl.handle.net/2142/132655
Description
Title
Evaluating ventral visual stream contributions to human visual robustness with deep learning approaches
Author(s)
Shao, Zhenan
Issue Date
2025-11-25
Director of Research (if dissertation) or Advisor (if thesis)
Beck, Diane M
Doctoral Committee Chair(s)
Beck, Diane M
Committee Member(s)
Federmeier, Kara
Koyejo, Sanmi
Uddenburg, Stefan
Willits, Jon
Department of Study
Psychology
Discipline
Psychology
Degree Granting Institution
University of Illinois Urbana-Champaign
Degree Name
Ph.D.
Degree Level
Dissertation
Keyword(s)
Human visual robustness
Deep neural networks
Ventral visual stream
Representation learning
Adversarial robustness
Object recognition
Abstract
The human visual system exhibits remarkable robustness, allowing resilient object recognition despite noisy and ambiguous sensory inputs. Such perceptual invariance has been proposed to stem from hierarchical computations along the ventral visual stream (VVS), a series of brain regions that progressively transform visual representation into increasingly abstract representations. In contrast, deep convolutional neural networks (DCNNs), despite being regarded as the closest approximator of the biological visual system, remain vulnerable to variations and adversarial perturbations that humans easily overcome. This discrepancy raises questions about the underlying mechanisms that confer robustness in human vision and whether they can be adequately captured by DCNNs.
This thesis leverages deep convolutional neural networks (DCNNs) to investigate the role of VVS in achieving visual robustness, with a particular focus on how neural representational spaces evolve along the hierarchy. We start with aligning DCNNs to successive regions of the VVS and demonstrate hierarchical improvements in robustness that emerge as progressively higher-order representations are used, suggesting that robustness arises from the transformation of representations across the visual hierarchy rather than from any single brain region. We next examine what makes VVS representations uniquely suited to supporting robustness from two complementary perspectives. First, we assess whether the critical aspect of the representational spaces across the VVS is on the granularity of neural manifold characteristics and their separability, in accordance with a popular manifold disentangling framework. Second, inspired by recent efforts to characterize differences in spatial frequency preferences between humans and models, we test whether VVS representations implicitly bias DCNNs towards the frequency channels most relied on by human vision, thereby guiding models towards the spectral features critical for robust visual processing. The findings from these experiments are expected to contribute to not only the understanding of robust visual inference in the human brain, but also the development of more robust artificial vision systems that better emulate human perceptual capabilities.
Use this login method if you
don't
have an
@illinois.edu
email address.
(Oops, I do have one)
IDEALS migrated to a new platform on June 23, 2022. If you created
your account prior to this date, you will have to reset your password
using the forgot-password link below.