GPT is a Transformer-based architecture and training procedure for natural language processing tasks. Training follows a two-stage procedure. First, a language modeling objective is used on the unlabeled data to learn the initial parameters of a neural network model. Subsequently, these parameters are adapted to a target task using the corresponding supervised objective.
Source: Improving Language Understanding by Generative Pre-TrainingPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Language Modelling | 85 | 10.87% |
Large Language Model | 51 | 6.52% |
Question Answering | 34 | 4.35% |
Prompt Engineering | 26 | 3.32% |
Text Generation | 24 | 3.07% |
Retrieval | 21 | 2.69% |
Sentence | 20 | 2.56% |
Decision Making | 19 | 2.43% |
In-Context Learning | 18 | 2.30% |