The Universal Transformer is a generalization of the Transformer architecture. Universal Transformers combine the parallelizability and global receptive field of feed-forward sequence models like the Transformer with the recurrent inductive bias of RNNs. They also utilise a dynamic per-position halting mechanism.
Source: Universal TransformersPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Sentence | 4 | 11.11% |
Language Modelling | 3 | 8.33% |
Text Generation | 2 | 5.56% |
Instance Segmentation | 1 | 2.78% |
Semantic Segmentation | 1 | 2.78% |
Reinforcement Learning (RL) | 1 | 2.78% |
Semantic Similarity | 1 | 2.78% |
Semantic Textual Similarity | 1 | 2.78% |
Video Inpainting | 1 | 2.78% |