Sunday, June 23, 2019

How to code The Transformer in PyTorch

https://blog.floydhub.com/the-transformer-in-pytorch/

Could The Transformer be another nail in the coffin for RNNs?
Doing away with clunky for-loops, the transformer instead finds a way to allow whole sentences to simultaneously enter the network in batches. With this technique, NLP reclaims the advantage of Python’s highly efficient linear algebra libraries. This time-saving can then be spent deploying more layers into the model.
So far, it seems the result from transformers is faster convergence and better results. What’s not to love?

No comments:

Post a Comment