語系:
繁體中文
English
說明(常見問題)
回圖書館首頁
手機版館藏查詢
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Privacy Protection in Conversations.
~
Xu, Qiongkai.
FindBook
Google Book
Amazon
博客來
Privacy Protection in Conversations.
紀錄類型:
書目-電子資源 : Monograph/item
正題名/作者:
Privacy Protection in Conversations./
作者:
Xu, Qiongkai.
出版者:
Ann Arbor : ProQuest Dissertations & Theses, : 2022,
面頁冊數:
99 p.
附註:
Source: Dissertations Abstracts International, Volume: 84-06, Section: B.
Contained By:
Dissertations Abstracts International84-06B.
標題:
Language. -
電子資源:
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=30167689
ISBN:
9798358422407
Privacy Protection in Conversations.
Xu, Qiongkai.
Privacy Protection in Conversations.
- Ann Arbor : ProQuest Dissertations & Theses, 2022 - 99 p.
Source: Dissertations Abstracts International, Volume: 84-06, Section: B.
Thesis (Ph.D.)--The Australian National University (Australia), 2022.
This item must not be sold to any third party vendors.
Leakage of personal information in online conversations raises serious privacy concerns. For example, malicious users might collect sensitive personal information from vulnerable users via deliberately designed conversations. This thesis tackles the problem of privacy leakage in textual conversations and proposes to mitigate the risks of privacy disclosure by detecting and rewriting the risky utterances. Previous research on privacy protection in text has a focus on manipulating the implicit semantic representations in a continuous high dimensional space, which are mostly used for eliminating trails of personal information to machine learning models. Our research has a focus on the explicit expressions of conversations, namely sequences of words or tokens, which are generally used between human interlocutors or human-computer interactions. The new setting for privacy protection in text could be applied to the conversations by individual human users, such as vulnerable people, and artificial conversational bots, such as digital personal assistants.This thesis consists of two parts, answering two research questions: How to detect the utterances with the risk of privacy leakage? and How to modify or rewrite the utterances into the ones with less private information?In the first part of this thesis, we aim to detect the utterances with privacy leakage risk and report the sensitive utterances to authorized users for approval. One of the essential challenge of the detection task is that we cannot acquire a large-scale aligned corpus for supervised training of natural language inference for private information. A compact dataset is collect to merely validate the privacy leakage detection models. We investigate weakly supervised methods to learn utterance-level inference from coarse set-level alignment signals. Then, we propose novel alignment models for utterance inference. Our approaches manage to outperform competitive baseline alignment methods. Additionally, we develop a privacy-leakage detection system integrated in Facebook Messenger to demonstrate the utility of our proposed task in real-world usage scenarios.In the second part of this thesis, we investigate two pieces of work to rewrite the privacy-leakage sentences automatically into less sensitive ones. The first work discusses obscuring personal information in form of classifiable attributes. We propose to reduce the bias of sensitive attributes, such as gender, political slant and race, using an obscured text rewriting models. The rewriting models are guided by corresponding classifiers for the personal attributes to obscure. Adversarial training and fairness risk measurement are proposed to enhance the fairness of the generators, alleviating privacy leakage of the target attributes. The second work protects personal information in the form of open-domain textual descriptions. We further explore three feasible rewriting strategies, deleting, obscuring, and steering, for privacy-aware text rewriting. We investigate the possibility of fine-tuning a pre-trained language model for privacy-aware text rewriting. Based on our dataset, we further observe the relation of rewriting strategies to their semantic spaces in a knowledge graph. Then, a simple but effective decoding method is developed to incorporate these semantic spaces into the rewriting models.As a whole, this thesis presents a comprehensive study and the first solutions in varying settings for protecting privacy in conversations. We demonstrate that both privacy leakage detection and privacy-aware text rewriting are plausible using machine learning methods. Our contributions also include novel ideas for text alignment for natural language inference, training technologies for attribute obfuscating, and open-domain knowledge guidance to text rewriting. This thesis opens up inquiries into protecting sensitive user information in conversations from the perspective of explicit text representation.
ISBN: 9798358422407Subjects--Topical Terms:
643551
Language.
Privacy Protection in Conversations.
LDR
:05096nmm a2200373 4500
001
2394438
005
20240422070857.5
006
m o d
007
cr#unu||||||||
008
251215s2022 ||||||||||||||||| ||eng d
020
$a
9798358422407
035
$a
(MiAaPQ)AAI30167689
035
$a
(MiAaPQ)AustNatlU1885266428
035
$a
AAI30167689
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Xu, Qiongkai.
$3
3763912
245
1 0
$a
Privacy Protection in Conversations.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2022
300
$a
99 p.
500
$a
Source: Dissertations Abstracts International, Volume: 84-06, Section: B.
500
$a
Advisor: Qu, Lizhen.
502
$a
Thesis (Ph.D.)--The Australian National University (Australia), 2022.
506
$a
This item must not be sold to any third party vendors.
520
$a
Leakage of personal information in online conversations raises serious privacy concerns. For example, malicious users might collect sensitive personal information from vulnerable users via deliberately designed conversations. This thesis tackles the problem of privacy leakage in textual conversations and proposes to mitigate the risks of privacy disclosure by detecting and rewriting the risky utterances. Previous research on privacy protection in text has a focus on manipulating the implicit semantic representations in a continuous high dimensional space, which are mostly used for eliminating trails of personal information to machine learning models. Our research has a focus on the explicit expressions of conversations, namely sequences of words or tokens, which are generally used between human interlocutors or human-computer interactions. The new setting for privacy protection in text could be applied to the conversations by individual human users, such as vulnerable people, and artificial conversational bots, such as digital personal assistants.This thesis consists of two parts, answering two research questions: How to detect the utterances with the risk of privacy leakage? and How to modify or rewrite the utterances into the ones with less private information?In the first part of this thesis, we aim to detect the utterances with privacy leakage risk and report the sensitive utterances to authorized users for approval. One of the essential challenge of the detection task is that we cannot acquire a large-scale aligned corpus for supervised training of natural language inference for private information. A compact dataset is collect to merely validate the privacy leakage detection models. We investigate weakly supervised methods to learn utterance-level inference from coarse set-level alignment signals. Then, we propose novel alignment models for utterance inference. Our approaches manage to outperform competitive baseline alignment methods. Additionally, we develop a privacy-leakage detection system integrated in Facebook Messenger to demonstrate the utility of our proposed task in real-world usage scenarios.In the second part of this thesis, we investigate two pieces of work to rewrite the privacy-leakage sentences automatically into less sensitive ones. The first work discusses obscuring personal information in form of classifiable attributes. We propose to reduce the bias of sensitive attributes, such as gender, political slant and race, using an obscured text rewriting models. The rewriting models are guided by corresponding classifiers for the personal attributes to obscure. Adversarial training and fairness risk measurement are proposed to enhance the fairness of the generators, alleviating privacy leakage of the target attributes. The second work protects personal information in the form of open-domain textual descriptions. We further explore three feasible rewriting strategies, deleting, obscuring, and steering, for privacy-aware text rewriting. We investigate the possibility of fine-tuning a pre-trained language model for privacy-aware text rewriting. Based on our dataset, we further observe the relation of rewriting strategies to their semantic spaces in a knowledge graph. Then, a simple but effective decoding method is developed to incorporate these semantic spaces into the rewriting models.As a whole, this thesis presents a comprehensive study and the first solutions in varying settings for protecting privacy in conversations. We demonstrate that both privacy leakage detection and privacy-aware text rewriting are plausible using machine learning methods. Our contributions also include novel ideas for text alignment for natural language inference, training technologies for attribute obfuscating, and open-domain knowledge guidance to text rewriting. This thesis opens up inquiries into protecting sensitive user information in conversations from the perspective of explicit text representation.
590
$a
School code: 0433.
650
4
$a
Language.
$3
643551
650
4
$a
Text categorization.
$3
3689475
650
4
$a
Software.
$2
gtt.
$3
619355
650
4
$a
Personal information.
$3
3562412
650
4
$a
Gender.
$3
2001319
650
4
$a
Privacy.
$3
528582
650
4
$a
Semantics.
$3
520060
650
4
$a
Computer science.
$3
523869
650
4
$a
Criminology.
$3
533274
650
4
$a
Logic.
$3
529544
650
4
$a
Cybercrime.
$3
2195714
690
$a
0679
690
$a
0800
690
$a
0984
690
$a
0627
690
$a
0395
710
2
$a
The Australian National University (Australia).
$3
1952885
773
0
$t
Dissertations Abstracts International
$g
84-06B.
790
$a
0433
791
$a
Ph.D.
792
$a
2022
793
$a
English
856
4 0
$u
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=30167689
筆 0 讀者評論
館藏地:
全部
電子資源
出版年:
卷號:
館藏
1 筆 • 頁數 1 •
1
條碼號
典藏地名稱
館藏流通類別
資料類型
索書號
使用類型
借閱狀態
預約狀態
備註欄
附件
W9502758
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
在架
0
1 筆 • 頁數 1 •
1
多媒體
評論
新增評論
分享你的心得
Export
取書館
處理中
...
變更密碼
登入