語系:
繁體中文
English
說明(常見問題)
回圖書館首頁
手機版館藏查詢
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Model-Agnostic Trajectory Abstractio...
~
Takagi, Yoshiki.
FindBook
Google Book
Amazon
博客來
Model-Agnostic Trajectory Abstraction and Visualization Method for Explainability in Reinforcement Learning.
紀錄類型:
書目-電子資源 : Monograph/item
正題名/作者:
Model-Agnostic Trajectory Abstraction and Visualization Method for Explainability in Reinforcement Learning./
作者:
Takagi, Yoshiki.
出版者:
Ann Arbor : ProQuest Dissertations & Theses, : 2024,
面頁冊數:
105 p.
附註:
Source: Dissertations Abstracts International, Volume: 85-12, Section: B.
Contained By:
Dissertations Abstracts International85-12B.
標題:
Computer science. -
電子資源:
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=30993031
ISBN:
9798383111680
Model-Agnostic Trajectory Abstraction and Visualization Method for Explainability in Reinforcement Learning.
Takagi, Yoshiki.
Model-Agnostic Trajectory Abstraction and Visualization Method for Explainability in Reinforcement Learning.
- Ann Arbor : ProQuest Dissertations & Theses, 2024 - 105 p.
Source: Dissertations Abstracts International, Volume: 85-12, Section: B.
Thesis (Ph.D.)--University of Hawai'i at Manoa, 2024.
Reinforcement learning (RL) has evolved rapidly in the past decade and is now capable of achieving human capabilities, such as self-driving cars. Moreover, in the last few years, the performance of deep RL, which applies deep neural networks to RL, has surpassed that of skilled human players in areas of video games, chess, and Go. However, as deep RL models become more complex, understanding and interpreting these models poses significant challenges.Explainable AI (XAI) research has shown the potential to close the gap between humans and a deep RL agent by providing explanations that help users to understand how the agent works. XAI approaches have been tailored for both RL experts and non-experts. For RL experts, visualizations of internal agent parameters reveal the learning mechanisms of deep RL agents, offering precise insights into agent behavior. However, this approach is less accessible to users who do not have RL expertise (non-RL experts). The communication gap between RL experts and non-experts thus remains a critical issue. For example, in discussions about the decision boundaries of autonomous Unmanned Aerial Vehicles (UAVs) between RL practitioners and pilots, the following issues arise:Pilots, who are non-RL experts, have domain knowledge, but they cannot use XAI interfaces designed for RL experts in the assessment of the RL model;{A0}In order to obtain feedback from pilots, RL experts need to explain the behavior of the RL model while minimizing the use of RL terminology;Pilots may use domain specific terminology during the assessment and the RL expert needs to interpret the pilot's statements and apply them to the model;Therefore, the central questions are: How can both RL experts and non-RL experts understand the behavior of an agent? In other words, how can humans naturally build a mental model of an agent? A promising approach is the 'familiarization effect' from cognitive psychology, where exposure to an agent's behavior in various scenarios helps users intuitively understand the agent, which is later applied to Human Robot Interaction. For instance, one research group observed that watching a robot's trajectory in videos enables users to predict the robot's future trajectory. Another study pointed out that short video clips of an agents' game-play can effectively build mental models of the agents' performance. However, this strategy may be less effective with multiple agents or in complex, extended tasks due to human limitations in short-term visual memory.{A0}Therefore, this dissertation addresses this problem by proposing a trajectory visualization that gives a high-level view of agents' behaviors through an abstraction of agents' behavior. This research will open up new directions, such as that domain experts who are not familiar with RL can get more involved in the development of RL which can lead to identifying important agent's behavior patterns that cannot be recognized by RL experts alone, and that the possibility of allowing general users to assess the capabilities and limitations of agents in the task of monitoring self-driving agents as a driver.
ISBN: 9798383111680Subjects--Topical Terms:
523869
Computer science.
Subjects--Index Terms:
Explainable AI
Model-Agnostic Trajectory Abstraction and Visualization Method for Explainability in Reinforcement Learning.
LDR
:04394nmm a2200409 4500
001
2401884
005
20241022111556.5
006
m o d
007
cr#unu||||||||
008
251215s2024 ||||||||||||||||| ||eng d
020
$a
9798383111680
035
$a
(MiAaPQ)AAI30993031
035
$a
AAI30993031
035
$a
2401884
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Takagi, Yoshiki.
$3
3772104
245
1 0
$a
Model-Agnostic Trajectory Abstraction and Visualization Method for Explainability in Reinforcement Learning.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2024
300
$a
105 p.
500
$a
Source: Dissertations Abstracts International, Volume: 85-12, Section: B.
500
$a
Advisor: Leigh, Jason.
502
$a
Thesis (Ph.D.)--University of Hawai'i at Manoa, 2024.
520
$a
Reinforcement learning (RL) has evolved rapidly in the past decade and is now capable of achieving human capabilities, such as self-driving cars. Moreover, in the last few years, the performance of deep RL, which applies deep neural networks to RL, has surpassed that of skilled human players in areas of video games, chess, and Go. However, as deep RL models become more complex, understanding and interpreting these models poses significant challenges.Explainable AI (XAI) research has shown the potential to close the gap between humans and a deep RL agent by providing explanations that help users to understand how the agent works. XAI approaches have been tailored for both RL experts and non-experts. For RL experts, visualizations of internal agent parameters reveal the learning mechanisms of deep RL agents, offering precise insights into agent behavior. However, this approach is less accessible to users who do not have RL expertise (non-RL experts). The communication gap between RL experts and non-experts thus remains a critical issue. For example, in discussions about the decision boundaries of autonomous Unmanned Aerial Vehicles (UAVs) between RL practitioners and pilots, the following issues arise:Pilots, who are non-RL experts, have domain knowledge, but they cannot use XAI interfaces designed for RL experts in the assessment of the RL model;{A0}In order to obtain feedback from pilots, RL experts need to explain the behavior of the RL model while minimizing the use of RL terminology;Pilots may use domain specific terminology during the assessment and the RL expert needs to interpret the pilot's statements and apply them to the model;Therefore, the central questions are: How can both RL experts and non-RL experts understand the behavior of an agent? In other words, how can humans naturally build a mental model of an agent? A promising approach is the 'familiarization effect' from cognitive psychology, where exposure to an agent's behavior in various scenarios helps users intuitively understand the agent, which is later applied to Human Robot Interaction. For instance, one research group observed that watching a robot's trajectory in videos enables users to predict the robot's future trajectory. Another study pointed out that short video clips of an agents' game-play can effectively build mental models of the agents' performance. However, this strategy may be less effective with multiple agents or in complex, extended tasks due to human limitations in short-term visual memory.{A0}Therefore, this dissertation addresses this problem by proposing a trajectory visualization that gives a high-level view of agents' behaviors through an abstraction of agents' behavior. This research will open up new directions, such as that domain experts who are not familiar with RL can get more involved in the development of RL which can lead to identifying important agent's behavior patterns that cannot be recognized by RL experts alone, and that the possibility of allowing general users to assess the capabilities and limitations of agents in the task of monitoring self-driving agents as a driver.
590
$a
School code: 0085.
650
4
$a
Computer science.
$3
523869
650
4
$a
Computer engineering.
$3
621879
650
4
$a
Automotive engineering.
$3
2181195
653
$a
Explainable AI
653
$a
Human-computer interaction
653
$a
Reinforcement learning
653
$a
Storyline visualization
653
$a
Trajectory abstraction
690
$a
0984
690
$a
0464
690
$a
0800
690
$a
0540
710
2
$a
University of Hawai'i at Manoa.
$b
Computer Science.
$3
3683211
773
0
$t
Dissertations Abstracts International
$g
85-12B.
790
$a
0085
791
$a
Ph.D.
792
$a
2024
793
$a
English
856
4 0
$u
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=30993031
筆 0 讀者評論
館藏地:
全部
電子資源
出版年:
卷號:
館藏
1 筆 • 頁數 1 •
1
條碼號
典藏地名稱
館藏流通類別
資料類型
索書號
使用類型
借閱狀態
預約狀態
備註欄
附件
W9510204
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
在架
0
1 筆 • 頁數 1 •
1
多媒體
評論
新增評論
分享你的心得
Export
取書館
處理中
...
變更密碼
登入