 The transformer consists of two key components, an encoder and a decoder. The encoder takes the English words simultaneously, and it generates embeddings for every word simultaneously. These embeddings are vectors that encapsulate the meaning of the word. Similar words have closer numbers in their vectors. The decoder takes these embeddings from the encoder and the previously generated words of the translated French sentence, and then it uses them to generate the next French word. And we keep generating the French translation one word at a time until the end of sentence is reached.