Language:
English
繁體中文
Help
回圖書館首頁
手機版館藏查詢
Login
Back
Switch To:
Labeled
|
MARC Mode
|
ISBD
Comparing Human and Machine Visual P...
~
Veerabadran, Vijay.
Linked to FindBook
Google Book
Amazon
博客來
Comparing Human and Machine Visual Perception.
Record Type:
Electronic resources : Monograph/item
Title/Author:
Comparing Human and Machine Visual Perception./
Author:
Veerabadran, Vijay.
Published:
Ann Arbor : ProQuest Dissertations & Theses, : 2024,
Description:
125 p.
Notes:
Source: Dissertations Abstracts International, Volume: 85-10, Section: B.
Contained By:
Dissertations Abstracts International85-10B.
Subject:
Neurosciences. -
Online resource:
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=30992102
ISBN:
9798382223575
Comparing Human and Machine Visual Perception.
Veerabadran, Vijay.
Comparing Human and Machine Visual Perception.
- Ann Arbor : ProQuest Dissertations & Theses, 2024 - 125 p.
Source: Dissertations Abstracts International, Volume: 85-10, Section: B.
Thesis (Ph.D.)--University of California, San Diego, 2024.
In this dissertation, we focus on examining differences in perception between humans and computer vision models and contribute novel research methods to increase their alignment. In recent studies comparing how humans and deep neural networks used in computer vision perceive visual stimuli, we find extensive evidence on how these highly performant models' visual perception often poorly aligns with human perception. For example, these models have been shown to classify objects in a scene solely based on a small fraction of border pixels in an image (Carter et al., 2021), preferentially attend to information outside the human frequency sensitivity spectrum (Subramanian et al., 2023), and (in)famously classify images by local texture rather than by global form (Geirhos et al., 2019b). These deviations of machine vision are often due to their overreliance on short-range features and our first set of contributions directly address this by adding lateral connections---critical for long-range spatial feature processing in biological vision---into deep neural networks. First, in Chapters 2 and 3, we introduce the bio-inspired DivNormEI and V1Net models respectively which implement feedforward and recurrent lateral connections in deep neural networks (DNNs). We show that these models develop bio-realistic orientation tuning and directly lead to robust object recognition/segmentation. We also show that recurrent lateral connections give rise to parameter-efficient contour integration (a task well-known to test long-range feature integration capacity). In Chapter 4, we introduce LocRNN, a high-performing recurrent circuit evolved from V1Net and propose combining it with Adaptive Computation Time (ACT) to learn a dynamic instance-conditional number of RNN timesteps. ACT enables LocRNN to generalize in a zero-shot manner to novel test-time difficulty levels of challenging visual path integration tasks. These chapters together highlight the effectiveness of our proposed bio-inspired design in creating human-like robustness to out-of-distribution settings. Complimentary to bio-inspired design, we also propose a new way to compare human and machine perception; advancing this area helps us better identify factors of deviation between these systems and guides us in building future neural networks with stronger alignment. In an elaborate psychophysics study described in Chapter 5, we explored how humans and deep neural networks alike, can be tricked by barely noticeable adversarial changes to images. We discuss the degree of alignment between the two visual systems and identify factors which influence this alignment. Our actionable predictions we discuss in this Chapter inspires the design of future neural network models with a goal of strengthening their alignment to human perception. We conclude this dissertation by eliciting important future directions of expansion of the research described here to build the next generation of computer vision models increasingly aligned with human vision.
ISBN: 9798382223575Subjects--Topical Terms:
588700
Neurosciences.
Subjects--Index Terms:
Adversarial machine learning
Comparing Human and Machine Visual Perception.
LDR
:04224nmm a2200409 4500
001
2398158
005
20240812064353.5
006
m o d
007
cr#unu||||||||
008
251215s2024 ||||||||||||||||| ||eng d
020
$a
9798382223575
035
$a
(MiAaPQ)AAI30992102
035
$a
AAI30992102
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Veerabadran, Vijay.
$3
3768067
245
1 0
$a
Comparing Human and Machine Visual Perception.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2024
300
$a
125 p.
500
$a
Source: Dissertations Abstracts International, Volume: 85-10, Section: B.
500
$a
Advisor: de Sa, Virginia R.
502
$a
Thesis (Ph.D.)--University of California, San Diego, 2024.
520
$a
In this dissertation, we focus on examining differences in perception between humans and computer vision models and contribute novel research methods to increase their alignment. In recent studies comparing how humans and deep neural networks used in computer vision perceive visual stimuli, we find extensive evidence on how these highly performant models' visual perception often poorly aligns with human perception. For example, these models have been shown to classify objects in a scene solely based on a small fraction of border pixels in an image (Carter et al., 2021), preferentially attend to information outside the human frequency sensitivity spectrum (Subramanian et al., 2023), and (in)famously classify images by local texture rather than by global form (Geirhos et al., 2019b). These deviations of machine vision are often due to their overreliance on short-range features and our first set of contributions directly address this by adding lateral connections---critical for long-range spatial feature processing in biological vision---into deep neural networks. First, in Chapters 2 and 3, we introduce the bio-inspired DivNormEI and V1Net models respectively which implement feedforward and recurrent lateral connections in deep neural networks (DNNs). We show that these models develop bio-realistic orientation tuning and directly lead to robust object recognition/segmentation. We also show that recurrent lateral connections give rise to parameter-efficient contour integration (a task well-known to test long-range feature integration capacity). In Chapter 4, we introduce LocRNN, a high-performing recurrent circuit evolved from V1Net and propose combining it with Adaptive Computation Time (ACT) to learn a dynamic instance-conditional number of RNN timesteps. ACT enables LocRNN to generalize in a zero-shot manner to novel test-time difficulty levels of challenging visual path integration tasks. These chapters together highlight the effectiveness of our proposed bio-inspired design in creating human-like robustness to out-of-distribution settings. Complimentary to bio-inspired design, we also propose a new way to compare human and machine perception; advancing this area helps us better identify factors of deviation between these systems and guides us in building future neural networks with stronger alignment. In an elaborate psychophysics study described in Chapter 5, we explored how humans and deep neural networks alike, can be tricked by barely noticeable adversarial changes to images. We discuss the degree of alignment between the two visual systems and identify factors which influence this alignment. Our actionable predictions we discuss in this Chapter inspires the design of future neural network models with a goal of strengthening their alignment to human perception. We conclude this dissertation by eliciting important future directions of expansion of the research described here to build the next generation of computer vision models increasingly aligned with human vision.
590
$a
School code: 0033.
650
4
$a
Neurosciences.
$3
588700
650
4
$a
Computer science.
$3
523869
650
4
$a
Statistics.
$3
517247
650
4
$a
Experimental psychology.
$3
2144733
653
$a
Adversarial machine learning
653
$a
Recurrent Neural Networks
653
$a
Computer vision
653
$a
Deep learning
653
$a
Object recognition
690
$a
0800
690
$a
0317
690
$a
0984
690
$a
0623
690
$a
0463
710
2
$a
University of California, San Diego.
$b
Cognitive Science.
$3
1037944
773
0
$t
Dissertations Abstracts International
$g
85-10B.
790
$a
0033
791
$a
Ph.D.
792
$a
2024
793
$a
English
856
4 0
$u
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=30992102
based on 0 review(s)
Location:
ALL
電子資源
Year:
Volume Number:
Items
1 records • Pages 1 •
1
Inventory Number
Location Name
Item Class
Material type
Call number
Usage Class
Loan Status
No. of reservations
Opac note
Attachments
W9506478
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
On shelf
0
1 records • Pages 1 •
1
Multimedia
Reviews
Add a review
and share your thoughts with other readers
Export
pickup library
Processing
...
Change password
Login