stat946w18/Synthetic and natural noise both break neural machine translation

From statwiki
Revision as of 23:08, 28 February 2018 by Z267xu (talk | contribs)
Jump to: navigation, search

Introduction

  • Humans have surprisingly robust language processing systems which can easily overcome typos, e.g.

Aoccdrnig to a rscheearch at Cmabrigde Uinervtisy, it deosn't mttaer in waht oredr the ltteers in a wrod are, the olny iprmoetnt tihng is taht the frist and lsat ltteer be at the rghit pclae.

  • A person's ability to read this text comes as no surprise to the Psychology literature
    1. Saberi \& Perrott (1999) found that this robustness extends to audio as well.
    2. Rayner et al. (2006) found that in noisier settings reading comprehension only slowed by 11 \%.
    3. McCusker et al. (1981) found that the common case of swapping letters could often go unnoticed by the reader.
    4. Mayall et al (1997) shows that we rely on word shape.
    5. Reicher, 1969; Pelli et al., (2003) found that we can switch between whole word recognition but the first and last letter positions are required to stay constant for comprehension

However, NMT(neural machine translation) systems are brittle. i.e. The Arabic word Good morning.PNG means a blessing for good morning, however Hunt.PNG means hunt or slaughter.

Facebook's MT system mistakenly confused two words that only differ by one character, a situation that is challenging for a character-based NMT system.

Figure 1 shows the performance translating German to English as a function of the percent of German words modified. Here we show two types of noise: (1) Random permutation of the word and (2) Swapping a pair of adjacent letters in the centre of words. The important thing to note is that even small amounts of noise lead to substantial drops in performance.

BLEU plot.PNG

BLEU (bilingual evaluation understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another. Quality is considered to be the correspondence between a machine's output and that of a human: "the closer a machine translation is to a professional human translation, the better it is". BLEU is between 0 and 1.

This paper explores two simple strategies for increasing model robustness:

  1. using structure-invariant representations ( character CNN representation)
  2. robust training on noisy data, a form of adversarial training.

Adversarial examples

The growing literature on adversarial examples has demonstrated how dangerous it can be to have brittle machine learning systems being used so pervasively in the real world.

The paper devises simple methods for generating adversarial examples for NMT. They do not assume any access to the NMT models' gradients, instead relying on cognitively-informed and naturally occurring language errors to generate noise.

MT system

We experiment with three different NMT systems with access to character information at different levels.

  1. Use the fully character-level model (Lee et al. 2017). This is a sequence to sequence model with attention that is trained on char2char.
  2. Use Nematus (Sennrich et al., 2017), a popular NMT toolkit. It is another sequence-to-sequence model with several architecture modifications, especially operating on sub-word units using byte-pair encoding.
  3. Use an attentional sequence-to-sequence model with a word representation based on a character convolutional neural network (charCNN). The charCNN model has two long short-term memory (LSTM) layers in the encoder and the decoder.

DATA

MY DATA

We use the TED talks parallel corpus prepared for IWSLT 2016 (Cettolo et al., 2012) for testing all of the NMT systems. Table1x.PNG

NATURAL AND ARTIFICIAL NOISE

NATURAL NOISE

To three different languages French, German and Czech, they have their own frequent natural errors.

The author harvest naturally occurring errors (typos, misspellings, etc.) corresponding to these three languages from available corpora of edits to build a look-up table of possible lexical replacements.

Synthetic Noise

In addition to naturally collected sources of error, we also experiment with four types of synthetic noise: Swap, Middle Random, Fully Random, and Key Typo.

  1. The first and simplest source of noise is swapping two letters (do not alter the first or last letters).
  2. \code{Middle Random}: Randomize the order of all the letters in a word except for the first and last.
  3. \code{Fully Random} Completely randomized words.
  4. \code{Keyboard Typo}Randomly replace one letter in each word with an adjacent key

Table3x.PNG

Table 3 shows BLEU scores of models trained on clean (Vanilla) texts and tested on clean and noisy texts. All models suffer a significant drop in BLEU when evaluated on noisy texts. This is true for both natural noise and all kinds of synthetic noise. The more noise in the text, the worse the translation quality, with random scrambling producing the lowest BLEU scores.

Dealing with noise

STRUCTURE INVARIANT REPRESENTATIONS

The three NMT models are all sensitive to word structure. The \code{char2char} and charCNN models both have convolutional layers on character sequences, designed to capture character n-grams. The model in \code{Nematus} is based on sub-word units obtained with BPE. It thus relies on character order.

The simplest to improve such model is to take the average character embeddings as a word representation. This model, referred to as \code{meanChar}, first generates a word representation by averaging character embeddings, and then proceeds with a word-level encoder similar to the charCNN model.

Table5x.PNG

\code{meanChar} is good with the other three scrambling errors (Swap, Middle Random and Fully Random), but bad with Keyboard error and Natural errors.

BLACK-BOX ADVERSARIAL TRAINING

charCNN Performance Table6x.PNG

Analysis

LEARNING MULTIPLE KINDS OF NOISE IN charCNN

They analyze the weights learned by charCNN models trained on two kinds of input: completely scrambled words (Rand) without other kinds of noise, and a mix of Rand+Key+Nat kinds of noise.

For each model, they compute the variance across the filter dimension for each one of the 1000 filters and for each one out of 25 character embedding dimensions. The we average the variances across the 1000 filters.

Table7x.PNG

Conclusion

In this work, they have shown that character-based NMT models are extremely brittle and tend to break when presented with both natural and synthetic kinds of noise. After models comparison, they found that a character-based CNN can learn to address multiple types of errors that are seen in training.