ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1909.02560
114
1
v1v2v3v4v5 (latest)

Adversarial Examples with Difficult Common Words for Paraphrase Identification

5 September 2019
Zhouxing Shi
Minlie Huang
Ting Yao
Jingfang Xu
ArXiv (abs)PDFHTML
Abstract

Deep models are commonly vulnerable to adversarial examples. In this paper, we propose the first algorithm for effectively generating both positive and negative adversarial examples for paraphrase identification. We first sample an original sentence pair from the dataset and then adversarially replace some word pairs with difficult common words. We take multiple steps and use beam search to find a modification that makes the target model fail, and thereby obtain an adversarial example. The word replacement is also constrained by heuristic rules and a language model, to preserve the label and language quality during modification. Experiments show that the performance of the target models has a severe drop on our adversarially modified examples.Meanwhile, human annotators are much less affected, and the generated sentences retain a good language quality. We also show that adversarial training with generated adversarial examples can improve model robustness, while previous work provides little improvement on our adversarial examples.

View on arXiv
Comments on this paper