Adversarial Attack on Neural Machine Translation System

dc.contributor.authorAbijith, K P
dc.date.accessioned2025-01-29T07:03:33Z
dc.date.available2025-01-29T07:03:33Z
dc.date.issued2019-06
dc.descriptionDissertation under the supervision of Dr. Utpal Garainen_US
dc.description.abstractNowadays Deep Neural Network based solutions are deployed to solve numerous tasks. Thus, it has become absolutely important to study the robustness of these systems. Machine Translation is one of the popular applications of Deep Neural Networks. This thesis studies the robustness of Neural Machine Translation systems by generating adversarial examples with the objective to fool the model. Whenever there is a change in the source, i.e. when a word in the input sentence is replaced by an unrelated word, the translation system is supposed to re ect the changes while doing translation. These unwanted invariance learned by the model is undesirable. With intention to exploit this undesirable property learned by a Neural Machine Translation system we design an attack called: Invariance-based targeted attack. This attack introduces multiple changes(replacement of words) to the original input sentence, keeping the translation unchanged. In-order to facilitate the explanation of the design of the attack we introduce two methods: (i) Min-Grad method: To identify the position where a replacement of the word makes the least change in the translation, and (ii) Soft-Attn method: To search for a new word to replace, given a list of choices. The initial part of the report explain the preliminary explorations we did in-order to get some insights on how to do the problem formulation. These experiments are run on LSTM based models with single replacement policy. Using the learning from the rst part we extend the experiments to Transformer and BLSTM based models, which are considered as the state-of-the-art systems for machine translation.en_US
dc.identifier.citation40p.en_US
dc.identifier.urihttp://hdl.handle.net/10263/7492
dc.language.isoenen_US
dc.publisherIndian Statistical Institute, Kolkataen_US
dc.relation.ispartofseriesMTech(CS) Dissertation;17-30
dc.subjectNeural Machine Translationen_US
dc.subjectMachine Translation System Dissertationen_US
dc.subjectNeural Networksen_US
dc.subjectLSTM modelsen_US
dc.titleAdversarial Attack on Neural Machine Translation Systemen_US
dc.typeOtheren_US

Files

Original bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
2019 - Abijith K P - CS1730.pdf
Size:
870.25 KB
Format:
Adobe Portable Document Format
Description:
Dissertation

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: