Latest articles in transformer package

Generative Pre-training (GPT) for  Natural Language Understanding

Generative Pre-training (GPT) for Natural Language Understanding

GPT-1 uses a 12-layer decoder-only transformer framework with masked self-attention for training the language model.

Popular transformer package

More articles in transformer package