Generative Pre-trained Transformer. OpenAI's series of large language models. GPT-3 showed the power of scale, GPT-4 demonstrated multimodal capabilities, and GPT-5 pushed reasoning further. The architecture is a decoder-only transformer trained to predict the next token in a sequence.
The AI company behind ChatGPT, GPT-4, DALL-E, and Whisper.
The neural network architecture behind virtually all modern AI language models.
An AI model with billions of parameters trained on massive text datasets.
A mathematical function applied to a neuron's output that introduces non-linearity into the network.
An optimization algorithm that combines the best parts of two other methods — AdaGrad and RMSProp.
Artificial General Intelligence.
Browse our complete glossary or subscribe to our newsletter for the latest AI news and insights.