Adv-OLM: Generating Textual Adversaries via OLM

Vijit Malik, Ashwani Bhat, Ashutosh Modi

Document analysis including Text Categorization and Topic Models Short paper Paper

Zoom-1B: Apr 21, Zoom-1B: Apr 21 (08:00-09:00 UTC) [Join Zoom Meeting]
Gather-1A: Apr 21, Gather-1A: Apr 21 (13:00-15:00 UTC) [Join Gather Meeting]

You can open the pre-recorded video in separate windows.

Abstract: Deep learning models are susceptible to adversarial examples that have imperceptible perturbations in the original input, resulting in adversarial attacks against these models. Analysis of these attacks on the state of the art transformers in NLP can help improve the robustness of these models against such adversarial inputs. In this paper, we present Adv-OLM, a black-box attack method that adapts the idea of Occlusion and Language Models (OLM) to the current state of the art attack methods. OLM is used to rank words of a sentence, which are later substituted using word replacement strategies. We experimentally show that our approach outperforms other attack methods for several text classification tasks.
NOTE: Video may display a random order of authors. Correct author list is at the top of this page.

Connected Papers in EACL2021

Similar Papers

Frequency-Guided Word Substitutions for Detecting Textual Adversarial Examples
Maximilian Mozes, Pontus Stenetorp, Bennett Kleinberg, Lewis Griffin,
Evaluating Neural Model Robustness for Machine Comprehension
Winston Wu, Dustin Arendt, Svitlana Volkova,
Diverse Adversaries for Mitigating Bias in Training
Xudong Han, Timothy Baldwin, Trevor Cohn,