ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2307.06708
16
3

To share or not to share: What risks would laypeople accept to give sensitive data to differentially-private NLP systems?

13 July 2023
Christopher F. Weiss
Frauke Kreuter
Ivan Habernal
ArXivPDFHTML
Abstract

Although the NLP community has adopted central differential privacy as a go-to framework for privacy-preserving model training or data sharing, the choice and interpretation of the key parameter, privacy budget ε\varepsilonε that governs the strength of privacy protection, remains largely arbitrary. We argue that determining the ε\varepsilonε value should not be solely in the hands of researchers or system developers, but must also take into account the actual people who share their potentially sensitive data. In other words: Would you share your instant messages for ε\varepsilonε of 10? We address this research gap by designing, implementing, and conducting a behavioral experiment (311 lay participants) to study the behavior of people in uncertain decision-making situations with respect to privacy-threatening situations. Framing the risk perception in terms of two realistic NLP scenarios and using a vignette behavioral study help us determine what ε\varepsilonε thresholds would lead lay people to be willing to share sensitive textual data - to our knowledge, the first study of its kind.

View on arXiv
Comments on this paper