語系:
繁體中文
English
說明(常見問題)
回圖書館首頁
手機版館藏查詢
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Comparing Human and Machine Visual P...
~
Veerabadran, Vijay.
FindBook
Google Book
Amazon
博客來
Comparing Human and Machine Visual Perception.
紀錄類型:
書目-電子資源 : Monograph/item
正題名/作者:
Comparing Human and Machine Visual Perception./
作者:
Veerabadran, Vijay.
出版者:
Ann Arbor : ProQuest Dissertations & Theses, : 2024,
面頁冊數:
125 p.
附註:
Source: Dissertations Abstracts International, Volume: 85-10, Section: B.
Contained By:
Dissertations Abstracts International85-10B.
標題:
Neurosciences. -
電子資源:
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=30992102
ISBN:
9798382223575
Comparing Human and Machine Visual Perception.
Veerabadran, Vijay.
Comparing Human and Machine Visual Perception.
- Ann Arbor : ProQuest Dissertations & Theses, 2024 - 125 p.
Source: Dissertations Abstracts International, Volume: 85-10, Section: B.
Thesis (Ph.D.)--University of California, San Diego, 2024.
In this dissertation, we focus on examining differences in perception between humans and computer vision models and contribute novel research methods to increase their alignment. In recent studies comparing how humans and deep neural networks used in computer vision perceive visual stimuli, we find extensive evidence on how these highly performant models' visual perception often poorly aligns with human perception. For example, these models have been shown to classify objects in a scene solely based on a small fraction of border pixels in an image (Carter et al., 2021), preferentially attend to information outside the human frequency sensitivity spectrum (Subramanian et al., 2023), and (in)famously classify images by local texture rather than by global form (Geirhos et al., 2019b). These deviations of machine vision are often due to their overreliance on short-range features and our first set of contributions directly address this by adding lateral connections---critical for long-range spatial feature processing in biological vision---into deep neural networks. First, in Chapters 2 and 3, we introduce the bio-inspired DivNormEI and V1Net models respectively which implement feedforward and recurrent lateral connections in deep neural networks (DNNs). We show that these models develop bio-realistic orientation tuning and directly lead to robust object recognition/segmentation. We also show that recurrent lateral connections give rise to parameter-efficient contour integration (a task well-known to test long-range feature integration capacity). In Chapter 4, we introduce LocRNN, a high-performing recurrent circuit evolved from V1Net and propose combining it with Adaptive Computation Time (ACT) to learn a dynamic instance-conditional number of RNN timesteps. ACT enables LocRNN to generalize in a zero-shot manner to novel test-time difficulty levels of challenging visual path integration tasks. These chapters together highlight the effectiveness of our proposed bio-inspired design in creating human-like robustness to out-of-distribution settings. Complimentary to bio-inspired design, we also propose a new way to compare human and machine perception; advancing this area helps us better identify factors of deviation between these systems and guides us in building future neural networks with stronger alignment. In an elaborate psychophysics study described in Chapter 5, we explored how humans and deep neural networks alike, can be tricked by barely noticeable adversarial changes to images. We discuss the degree of alignment between the two visual systems and identify factors which influence this alignment. Our actionable predictions we discuss in this Chapter inspires the design of future neural network models with a goal of strengthening their alignment to human perception. We conclude this dissertation by eliciting important future directions of expansion of the research described here to build the next generation of computer vision models increasingly aligned with human vision.
ISBN: 9798382223575Subjects--Topical Terms:
588700
Neurosciences.
Subjects--Index Terms:
Adversarial machine learning
Comparing Human and Machine Visual Perception.
LDR
:04224nmm a2200409 4500
001
2398158
005
20240812064353.5
006
m o d
007
cr#unu||||||||
008
251215s2024 ||||||||||||||||| ||eng d
020
$a
9798382223575
035
$a
(MiAaPQ)AAI30992102
035
$a
AAI30992102
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Veerabadran, Vijay.
$3
3768067
245
1 0
$a
Comparing Human and Machine Visual Perception.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2024
300
$a
125 p.
500
$a
Source: Dissertations Abstracts International, Volume: 85-10, Section: B.
500
$a
Advisor: de Sa, Virginia R.
502
$a
Thesis (Ph.D.)--University of California, San Diego, 2024.
520
$a
In this dissertation, we focus on examining differences in perception between humans and computer vision models and contribute novel research methods to increase their alignment. In recent studies comparing how humans and deep neural networks used in computer vision perceive visual stimuli, we find extensive evidence on how these highly performant models' visual perception often poorly aligns with human perception. For example, these models have been shown to classify objects in a scene solely based on a small fraction of border pixels in an image (Carter et al., 2021), preferentially attend to information outside the human frequency sensitivity spectrum (Subramanian et al., 2023), and (in)famously classify images by local texture rather than by global form (Geirhos et al., 2019b). These deviations of machine vision are often due to their overreliance on short-range features and our first set of contributions directly address this by adding lateral connections---critical for long-range spatial feature processing in biological vision---into deep neural networks. First, in Chapters 2 and 3, we introduce the bio-inspired DivNormEI and V1Net models respectively which implement feedforward and recurrent lateral connections in deep neural networks (DNNs). We show that these models develop bio-realistic orientation tuning and directly lead to robust object recognition/segmentation. We also show that recurrent lateral connections give rise to parameter-efficient contour integration (a task well-known to test long-range feature integration capacity). In Chapter 4, we introduce LocRNN, a high-performing recurrent circuit evolved from V1Net and propose combining it with Adaptive Computation Time (ACT) to learn a dynamic instance-conditional number of RNN timesteps. ACT enables LocRNN to generalize in a zero-shot manner to novel test-time difficulty levels of challenging visual path integration tasks. These chapters together highlight the effectiveness of our proposed bio-inspired design in creating human-like robustness to out-of-distribution settings. Complimentary to bio-inspired design, we also propose a new way to compare human and machine perception; advancing this area helps us better identify factors of deviation between these systems and guides us in building future neural networks with stronger alignment. In an elaborate psychophysics study described in Chapter 5, we explored how humans and deep neural networks alike, can be tricked by barely noticeable adversarial changes to images. We discuss the degree of alignment between the two visual systems and identify factors which influence this alignment. Our actionable predictions we discuss in this Chapter inspires the design of future neural network models with a goal of strengthening their alignment to human perception. We conclude this dissertation by eliciting important future directions of expansion of the research described here to build the next generation of computer vision models increasingly aligned with human vision.
590
$a
School code: 0033.
650
4
$a
Neurosciences.
$3
588700
650
4
$a
Computer science.
$3
523869
650
4
$a
Statistics.
$3
517247
650
4
$a
Experimental psychology.
$3
2144733
653
$a
Adversarial machine learning
653
$a
Recurrent Neural Networks
653
$a
Computer vision
653
$a
Deep learning
653
$a
Object recognition
690
$a
0800
690
$a
0317
690
$a
0984
690
$a
0623
690
$a
0463
710
2
$a
University of California, San Diego.
$b
Cognitive Science.
$3
1037944
773
0
$t
Dissertations Abstracts International
$g
85-10B.
790
$a
0033
791
$a
Ph.D.
792
$a
2024
793
$a
English
856
4 0
$u
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=30992102
筆 0 讀者評論
館藏地:
全部
電子資源
出版年:
卷號:
館藏
1 筆 • 頁數 1 •
1
條碼號
典藏地名稱
館藏流通類別
資料類型
索書號
使用類型
借閱狀態
預約狀態
備註欄
附件
W9506478
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
在架
0
1 筆 • 頁數 1 •
1
多媒體
評論
新增評論
分享你的心得
Export
取書館
處理中
...
變更密碼
登入