https://blog.floydhub.com/the-transformer-in-pytorch/
Could The Transformer be another nail in the coffin for RNNs?
Doing
away with clunky for-loops, the transformer instead finds a way to
allow whole sentences to simultaneously enter the network in batches.
With this technique, NLP reclaims the advantage of Python’s highly
efficient linear algebra libraries. This time-saving can then be spent
deploying more layers into the model.
So far, it seems the result from transformers is faster convergence and better results. What’s not to love?
No comments:
Post a Comment