From the course: Introduction to Attention-Based Neural Networks

Unlock this course with a free trial

Join today to access over 24,600 courses taught by industry experts.

Sequence to sequence models for language translation

Sequence to sequence models for language translation

From the course: Introduction to Attention-Based Neural Networks

Sequence to sequence models for language translation

- Now that we have a big picture idea of how language generation models work, let's understand how language translation models work. These are sequence to sequence models. Let's assume we have a simple language translation model to translate from English to German. So if you feed an I ate an apple you'll get the equivalent sentence in German. Please forgive me if I don't try to pronounce the German sentence. I'll just fail miserably. Now, language translation models are sequence to sequence models, and the model structure uses two different networks. The first network is an encoder, and the second network is a decoder. Because the encoder takes in the sentence in the original language as an input, The encoder is an RNN. The decoder produces a sentence in the target language as an output. The decoder is also an RNN. The encoder works on the sequential input in the source language. Decoder produces sequential output…

Contents