attention-vs-transformer
Transformer models are a more powerful type of attention model that can achieve state-of-the-art performance on a variety of NLP tasks. Attention model has encoder and decoder. ransformer model uses Encoder and decoder, both made up of self-attention layers Transformer model use specific type of attention "elf attention" that makes it powerful. in self supervised learning. |