Charles Explorer logo
🇬🇧

Using Adversarial Examples in Natural Language Processing

Publication at Faculty of Mathematics and Physics |
2018

Abstract

Machine learning models have been providing promising results in many fields including natural language processing. These models are, nevertheless, prone to adversarial examples.

These are artificially constructed examples which evince two main features: they resemble the real training data but they deceive already trained model. This paper investigates the effect of using adversarial examples during the training of recurrent neural networks whose text input is in the form of a sequence of word/character embeddings.

The effects are studied on a compilation of eight NLP datasets whose interface was unified for quick experimenting. Based on the experiments and the dataset characteristics, we conclude that using the adversarial examples for NLP tasks that are modeled by recurrent neural networks provides a regularization effect and enables the training of models with greater number of parameters without overfitting.

In addition, we discuss which combinations of datasets and model settings might benefit f