Startsida
Hjälp
Sök i LIBRIS databas

     

 

Sökning: onr:kxvc9tgbh27sf09w > Understanding Neura...

Understanding Neural Machine Translation [Elektronisk resurs] An investigation into linguistic phenomena and attention mechanisms

Tang, Gongbo (författare)
Nivre, Joakim (preses)
Sennrich, Rico (preses)
Koehn, Philipp (opponent)
Uppsala universitet Humanistisk-samhällsvetenskapliga vetenskapsområdet (utgivare)
Publicerad: Uppsala : Acta Universitatis Upsaliensis, 2020
Engelska 137
Serie: Studia Linguistica Upsaliensia, 1652-1366 1652-1366 ; 26
Läs hela texten
Läs hela texten
  • E-bokAvhandling(Diss. Uppsala : Uppsala universitet, 2020)
Sammanfattning Ämnesord
Stäng  
  • In this thesis, I explore neural machine translation (NMT) models via targeted investigation of various linguistic phenomena and thorough exploration of the internal structure of NMT models, in particular the attention mechanism. With respect to linguistic phenomena, I explore the ability of NMT models to translate ambiguous words, to learn long-range dependencies, to learn morphology, and to translate negation—linguistic phenomena that have been challenging for the older paradigm of statistical machine translation. I find that morphological inflection and negation are better modeled in encoder hidden states, while the senses of ambiguous words are better learned in decoder hidden states. Hidden states from lower layers are better at capturing aspects of form, such as morphological inflections and negation cues, while hidden states from higher layers are better at capturing semantic and relational aspects, such as word senses, negation events, and negation scope. I conclude that NMT models learn linguistic knowledge in a bottom-up manner. In the final part of the thesis, I interpret attention mechanisms in encoder-free models and character-level models. I show that attending to word embeddings directly does not make attention mechanisms more alignment-like but instead demonstrates that the attention mechanism is adaptable and more important for NMT than encoders. In character-level models, all characters attract equal attention except the final separators. Overall, the ability of NMT models to deal with the studied linguistic phenomena gets stronger with the evolution of architectures. NMT models perform well in translating frequent ambiguous words and learning long-range dependencies, but still suffer from morphological errors and the under-translation of negation. Attention mechanisms are crucial and adaptable, and there is no uniform behavior in different settings. 

Ämnesord

Natural Sciences  (hsv)
Computer and Information Sciences  (hsv)
Language Technology (Computational Linguistics)  (hsv)
Naturvetenskap  (hsv)
Data- och informationsvetenskap  (hsv)
Språkteknologi (språkvetenskaplig databehandling)  (hsv)
Datorlingvistik  (uu)
Computational Linguistics  (uu)

Genre

government publication  (marcgt)

Indexterm och SAB-rubrik

Neural machine translation
Linguistic phenomena
Ambiguity
Long-range dependency
Morphology
Negation
Attention mechanisms
Interpretation
Inställningar Hjälp

Uppgift om bibliotek saknas i LIBRIS

Kontakta ditt bibliotek, eller sök utanför LIBRIS. Se högermenyn.

Om LIBRIS
Sekretess
Hjälp
Fel i posten?
Kontakt
Teknik och format
Sök utifrån
Sökrutor
Plug-ins
Bookmarklet
Anpassa
Textstorlek
Kontrast
Vyer
LIBRIS söktjänster
SwePub
Uppsök

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

Copyright © LIBRIS - Nationella bibliotekssystem

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy