An AI model that understands and generates human language. At its core, it predicts the probability of the next word in a sequence. Modern large language models (LLMs) scale this simple idea to billions of parameters and trillions of training tokens, producing remarkably capable systems.
An AI model with billions of parameters trained on massive text datasets.
Natural Language Processing.
The neural network architecture behind virtually all modern AI language models.
A mathematical function applied to a neuron's output that introduces non-linearity into the network.
An optimization algorithm that combines the best parts of two other methods — AdaGrad and RMSProp.
Artificial General Intelligence.
Browse our complete glossary or subscribe to our newsletter for the latest AI news and insights.