44
0

Investigating User Perspectives on Differentially Private Text Privatization

Abstract

Recent literature has seen a considerable uptick in Differentially Private Natural Language Processing\textit{Differentially Private Natural Language Processing} (DP NLP). This includes DP text privatization, where potentially sensitive input texts are transformed under DP to achieve privatized output texts that ideally mask sensitive information and\textit{and} maintain original semantics. Despite continued work to address the open challenges in DP text privatization, there remains a scarcity of work addressing user perceptions of this technology, a crucial aspect which serves as the final barrier to practical adoption. In this work, we conduct a survey study with 721 laypersons around the globe, investigating how the factors of scenario\textit{scenario}, data sensitivity\textit{data sensitivity}, mechanism type\textit{mechanism type}, and reason for data collection\textit{reason for data collection} impact user preferences for text privatization. We learn that while all these factors play a role in influencing privacy decisions, users are highly sensitive to the utility and coherence of the private output texts. Our findings highlight the socio-technical factors that must be considered in the study of DP NLP, opening the door to further user-based investigations going forward.

View on arXiv
@article{meisenbacher2025_2503.09338,
  title={ Investigating User Perspectives on Differentially Private Text Privatization },
  author={ Stephen Meisenbacher and Alexandra Klymenko and Alexander Karpp and Florian Matthes },
  journal={arXiv preprint arXiv:2503.09338},
  year={ 2025 }
}
Comments on this paper