NIPS 2017 paper.
We dig into the details of the Transformer, from the "attention is all you need" paper. Ashish and Jakob give us some motivation for replacing RNNs and CNNs with a more parallelizable self-attention mec…
Home
Feed
Search
Library
Download